Feat: kokoro onnx (WIP)
This commit is contained in:
27
Makefile
27
Makefile
@@ -1,4 +1,4 @@
|
||||
.PHONY: setconfig run lint lintall install-linters setup-whisper build-whisper download-whisper-model docker-up docker-down docker-logs noextra-run installdelve checkdelve fetch-onnx install-onnx-deps
|
||||
.PHONY: setconfig run lint lintall install-linters setup-whisper build-whisper download-whisper-model docker-up docker-down docker-logs noextra-run installdelve checkdelve fetch-onnx install-onnx-deps fetch-kokoro-voices install-espeak
|
||||
|
||||
run: setconfig
|
||||
go build -tags extra -o gf-lt && ./gf-lt
|
||||
@@ -33,6 +33,9 @@ lintall: lint
|
||||
fetch-onnx:
|
||||
mkdir -p onnx/embedgemma && curl -o onnx/embedgemma/config.json -L https://huggingface.co/onnx-community/embeddinggemma-300m-ONNX/resolve/main/config.json && curl -o onnx/embedgemma/tokenizer.json -L https://huggingface.co/onnx-community/embeddinggemma-300m-ONNX/resolve/main/tokenizer.json && curl -o onnx/embedgemma/model_q4.onnx -L https://huggingface.co/onnx-community/embeddinggemma-300m-ONNX/resolve/main/onnx/model_q4.onnx && curl -o onnx/embedgemma/model_q4.onnx_data -L https://huggingface.co/onnx-community/embeddinggemma-300m-ONNX/resolve/main/onnx/model_q4.onnx_data?download=true
|
||||
|
||||
fetch-kokoro-onnx:
|
||||
mkdir -p onnx/kokoro && curl -o onnx/kokoro/config.json -L https://huggingface.co/onnx-community/Kokoro-82M-v1.0-ONNX/resolve/main/config.json && curl -o onnx/kokoro/tokenizer.json -L https://huggingface.co/onnx-community/Kokoro-82M-v1.0-ONNX/resolve/main/tokenizer.json && curl -o onnx/kokoro/model_quantized.onnx -L https://huggingface.co/onnx-community/Kokoro-82M-v1.0-ONNX/resolve/main/onnx/model_quantized.onnx && curl -o onnx/kokoro/voices.bin -L https://github.com/thewh1teagle/kokoro-onnx/releases/download/model-files-v1.0/voices-v1.0.bin
|
||||
|
||||
install-onnx-deps: ## Install ONNX Runtime with CUDA support (or CPU fallback)
|
||||
@echo "=== ONNX Runtime Installer ===" && \
|
||||
echo "" && \
|
||||
@@ -194,3 +197,25 @@ docker-logs-whisper: ## View logs from Whisper STT service only
|
||||
docker-logs-kokoro: ## View logs from Kokoro TTS service only
|
||||
@echo "Displaying logs from Kokoro TTS service..."
|
||||
docker-compose -f batteries/docker-compose.yml logs -f kokoro-tts
|
||||
|
||||
# Kokoro ONNX TTS Setup
|
||||
install-espeak: ## Install espeak-ng for phoneme tokenization
|
||||
@echo "=== Installing espeak-ng ===" && \
|
||||
if command -v espeak-ng >/dev/null 2>&1; then \
|
||||
echo "espeak-ng is already installed:" && \
|
||||
espeak-ng --version && \
|
||||
exit 0; \
|
||||
fi && \
|
||||
echo "Installing espeak-ng..." && \
|
||||
sudo apt-get update && \
|
||||
sudo apt-get install -y espeak-ng espeak && \
|
||||
echo "espeak-ng installed successfully!" && \
|
||||
espeak-ng --version
|
||||
|
||||
fetch-kokoro-voices: ## Download Kokoro voice files (PyTorch format)
|
||||
@echo "=== Downloading Kokoro voices ===" && \
|
||||
mkdir -p onnx/kokoro/voices && \
|
||||
echo "Downloading af_bella voice..." && \
|
||||
curl -L -o onnx/kokoro/voices/af_bella.pt https://raw.githubusercontent.com/hexgrad/kokoro/main/kokoro/voices/af_heart.pt && \
|
||||
echo "Voice file downloaded to onnx/kokoro/voices/" && \
|
||||
ls -lh onnx/kokoro/voices/
|
||||
|
||||
Reference in New Issue
Block a user