Skip to content

Commit b154fa7

Browse files
committed
update readme
1 parent 1ed0d0a commit b154fa7

File tree

1 file changed

+41
-36
lines changed

1 file changed

+41
-36
lines changed

README.md

+41-36
Original file line numberDiff line numberDiff line change
@@ -115,42 +115,47 @@ unzip XTTS-v2.zip -d .
115115

116116
1. Rename the .env.sample to `.env` in the root directory of the project and configure it with the necessary environment variables: - The app is controlled based on the variables you add.
117117

118-
```env
119-
# Conditional API Usage: Depending on the value of MODEL_PROVIDER, that's what will be used when ran
120-
# use either ollama or openai, can mix and match, use local olllama with openai speech or use openai model with local xtts, ect..
121-
122-
# openai or ollama
123-
MODEL_PROVIDER=ollama
124-
125-
# Enter charactor name to use - samantha, wizard, pirate, valleygirl, newscaster1920s,
126-
CHARACTER_NAME=pirate
127-
128-
# Text-to-Speech Provider - (xtts local uses the custom charactor .wav) or (openai text to speech uses openai tts voice)
129-
# xtts or openai
130-
TTS_PROVIDER=xtts
131-
132-
# The voice speed for xtts only ( 1.0 - 1.5 , default 1.1)
133-
XTTS_SPEED=1.1
134-
135-
# OpenAI TTS Voice - When TTS Provider is set to openai above it will use the chosen voice
136-
# Examples here https://platform.openai.com/docs/guides/text-to-speech
137-
# Choose the desired voice options are - alloy, echo, fable, onyx, nova, and shimmer
138-
OPENAI_TTS_VOICE=onyx
139-
140-
# SET THESE BELOW AND NO NEED TO CHANGE OFTEN #
141-
142-
# Endpoints
143-
OPENAI_BASE_URL=https://api.openai.com/v1/chat/completions
144-
OPENAI_TTS_URL=https://api.openai.com/v1/audio/speech
145-
OLLAMA_BASE_URL=http://localhost:11434
146-
147-
# OpenAI API Key for models and speech
148-
OPENAI_API_KEY=sk-11111111
149-
150-
# Models to use - llama3 works good for local
151-
OPENAI_MODEL=gpt-4o
152-
OLLAMA_MODEL=llama3
153-
```
118+
```env
119+
# Conditional API Usage: Depending on the value of MODEL_PROVIDER, that's what will be used when run.
120+
# You can mix and match; use local Ollama with OpenAI speech or use OpenAI model with local XTTS, etc.
121+
122+
# Model Provider: openai or ollama
123+
MODEL_PROVIDER=ollama
124+
125+
# Character to use - Options: samantha, wizard, pirate, valleygirl, newscaster1920s, alien_scientist, cyberpunk, detective
126+
CHARACTER_NAME=wizard
127+
128+
# Text-to-Speech Provider - Options: xtts (local uses the custom character .wav) or openai (uses OpenAI TTS voice)
129+
TTS_PROVIDER=xtts
130+
131+
# OpenAI TTS Voice - When TTS_PROVIDER is set to openai above, it will use the chosen voice.
132+
# If MODEL_PROVIDER is ollama, then it will use the .wav in the character folder.
133+
# Voice options: alloy, echo, fable, onyx, nova, shimmer
134+
OPENAI_TTS_VOICE=onyx
135+
136+
# Endpoints (set these below and no need to change often)
137+
OPENAI_BASE_URL=https://api.openai.com/v1/chat/completions
138+
OPENAI_TTS_URL=https://api.openai.com/v1/audio/speech
139+
OLLAMA_BASE_URL=http://localhost:11434
140+
141+
# OpenAI API Key for models and speech (replace with your actual API key)
142+
OPENAI_API_KEY=sk-proj-1111111111
143+
144+
# Models to use - llama3 works well for local usage.
145+
# OPTIONAL: For screen analysis, if MODEL_PROVIDER is ollama, llava will be used by default.
146+
# Ensure you have llava downloaded with Ollama. If OpenAI is used, gpt-4o works well.
147+
OPENAI_MODEL=gpt-4o
148+
OLLAMA_MODEL=llama3
149+
150+
# The voice speed for XTTS only (1.0 - 1.5, default is 1.1)
151+
XTTS_SPEED=1.2
152+
153+
# NOTES:
154+
# List of trigger phrases to have the model view your desktop (desktop, browser, images, etc.).
155+
# It will describe what it sees, and you can ask questions about it:
156+
# "what's on my screen", "take a screenshot", "show me my screen", "analyze my screen",
157+
# "what do you see on my screen", "screen capture", "screenshot"
158+
```
154159

155160
## Usage
156161

0 commit comments

Comments
 (0)