You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Local XTTS you also might need cuDNN for using nvidia GPU https://developer.nvidia.com/cudnn and make sure C:\Program Files\NVIDIA\CUDNN\v9.5\bin\12.6
83
83
is in system PATH or whatever version you downloaded
84
84
85
-
### Optional - Download Checkpoints - ONLY IF YOU ARE USING THE LOCAL TTS
85
+
### Optional - Download Checkpoints - ONLY IF YOU ARE USING THE LOCAL XTTS
86
86
87
87
If you are only using speech with Openai or Elevenlabs then you don't need this. To use the local TTS download the checkpoints for the models used in this project ( the docker image has the local xtts and checkpoints in it already ). You can download them from the GitHub releases page and extract the zip and put into the project folder.
88
88
@@ -211,7 +211,9 @@ Running from wsl
211
211
docker run -d --gpus all -e "PULSE_SERVER=/mnt/wslg/PulseServer" -v \\wsl$\Ubuntu\mnt\wslg:/mnt/wslg/ --env-file .env --name voice-chat-ai -p 8000:8000 voice-chat-ai:latest
212
212
```
213
213
214
-
## Docker build without local xtts and no cuda - Recommended - 6 GB image
214
+
## Build without local xtts and no cuda - 5 GB image - Recommended
215
+
216
+
This is for when you only want to use Openai or Elevenlabs for speech. No need to download checkpoints, cuda and cudnn not required. Uses cpu for faster whisper. So no gpu needed.
0 commit comments