mirror of
https://github.com/remsky/Kokoro-FastAPI.git
synced 2025-04-13 09:39:17 +00:00
Update the readme
This commit is contained in:
parent
0db7b73678
commit
a1e01dbdbe
1 changed files with 24 additions and 6 deletions
24
README.md
24
README.md
|
@ -39,16 +39,34 @@ The service can be accessed through either the API endpoints or the Gradio web i
|
||||||
# cd docker/cpu # Run this or the above
|
# cd docker/cpu # Run this or the above
|
||||||
docker compose up --build
|
docker compose up --build
|
||||||
```
|
```
|
||||||
- OR running the API alone using Docker (model + voice packs baked in):
|
Once started:
|
||||||
```bash
|
- The API will be available at http://localhost:8880
|
||||||
|
- The UI can be accessed at http://localhost:7860
|
||||||
|
|
||||||
|
- OR run the API alone using Docker (model + voice packs baked in):
|
||||||
|
```bash
|
||||||
docker run -p 8880:8880 ghcr.io/remsky/kokoro-fastapi-cpu:latest # CPU
|
docker run -p 8880:8880 ghcr.io/remsky/kokoro-fastapi-cpu:latest # CPU
|
||||||
docker run --gpus all -p 8880:8880 ghcr.io/remsky/kokoro-fastapi-gpu:latest # Nvidia GPU
|
docker run --gpus all -p 8880:8880 ghcr.io/remsky/kokoro-fastapi-gpu:latest # Nvidia GPU
|
||||||
|
|
||||||
# Minified versions are available with `:latest-slim` tag, though it is a first test and may not be functional
|
# Minified versions are available with `:latest-slim` tag, though it is a first test and may not be functional
|
||||||
```
|
```
|
||||||
|
|
||||||
|
3. Running the UI Docker Service:
|
||||||
|
|
||||||
2. Run locally as an OpenAI-Compatible Speech Endpoint
|
- If you only want to run the Gradio web interface separately and connect it to an existing API service:
|
||||||
|
```bash
|
||||||
|
docker run -p 7860:7860 \
|
||||||
|
-e API_HOST=<api-hostname-or-ip> \
|
||||||
|
-e API_PORT=8880 \
|
||||||
|
ghcr.io/remsky/kokoro-fastapi-ui:v0.1.0
|
||||||
|
```
|
||||||
|
|
||||||
|
- Replace `<api-hostname-or-ip>` with:
|
||||||
|
- `kokoro-tts` if the UI container is running in the same Docker Compose setup.
|
||||||
|
- `localhost` if the API is running on your local machine.
|
||||||
|
|
||||||
|
|
||||||
|
4. Run locally as an OpenAI-Compatible Speech Endpoint
|
||||||
```python
|
```python
|
||||||
from openai import OpenAI
|
from openai import OpenAI
|
||||||
client = OpenAI(
|
client = OpenAI(
|
||||||
|
|
Loading…
Add table
Reference in a new issue