|
--- |
|
license: cc-by-sa-4.0 |
|
language: |
|
- en |
|
- hi |
|
base_model: |
|
- 11mlabs/indri-0.1-124m-tts |
|
pipeline_tag: text-to-speech |
|
--- |
|
|
|
# Indri GGUF Inference |
|
|
|
Refer to the original model and more details [here](https://huggingface.co/11mlabs/indri-0.1-124m-tts). |
|
|
|
This guide will help in running Indri models on CPU in GGUF format. |
|
|
|
## Step 1: Build llama.cpp |
|
|
|
To run the inference locally, you need to build `llama.cpp` project. The updated guide to do so can be found [here](https://github.com/ggerganov/llama.cpp/blob/master/docs/build.md). |
|
|
|
```bash |
|
git clone https://github.com/ggerganov/llama.cpp |
|
cd llama.cpp |
|
|
|
cmake -B build |
|
cmake --build build --config Release |
|
``` |
|
|
|
## Step 2: Download the model |
|
|
|
Download the GGUF format models from HuggingFace and place them inside `llama.cpp/models/`. |
|
The models can be found [here](https://huggingface.co/11mlabs/indri-0.1-124m-tts-GGUF/tree/main). |
|
|
|
Once the model is placed inside the directory, run the `llama-cpp` server from inside the `llama.cpp` directory |
|
|
|
```bash |
|
# For F16 model, update for different quantization accordingly |
|
./build/bin/llama-server -m /indri-0.1-124M-tts-F16.gguf --samplers 'top_k:temperature' --top_k 15 |
|
``` |
|
|
|
Refer [here](https://github.com/ggerganov/llama.cpp/tree/master/examples/main) if you are facing issues in running the llama-server locally. |
|
|
|
## Step 3: Run the inference script |
|
|
|
Clone the GitHub repository: |
|
|
|
```bash |
|
git clone https://github.com/cmeraki/indri.git |
|
cd indri |
|
|
|
python -m src.tts_gguf --text 'hi my name is Indri' --speaker '[spkr_63]' --out out.wav |
|
``` |
|
|
|
Speakers are available [here](https://github.com/cmeraki/indri?tab=readme-ov-file#available-speakers). |
|
|
|
You can also run an inference server |
|
|
|
```bash |
|
pip install -r requirements.txt |
|
|
|
# Install ffmpeg (for Mac/Windows, refer here: https://www.ffmpeg.org/download.html) |
|
sudo apt update -y |
|
sudo apt upgrade -y |
|
sudo apt install ffmpeg -y |
|
|
|
python -m server_ggpuf |
|
``` |
|
|
|
Redirect to `http://localhost:8000/docs` to see the API documentation and test the service. |
|
|
|
We are working on making this process more straightforward. Stay tuned for updates! |
|
|