File size: 3,145 Bytes
4b9d55e
 
82175fe
 
 
 
 
 
4b9d55e
 
82175fe
ba81a8e
 
82175fe
4b9d55e
 
 
 
82175fe
ba81a8e
 
 
82175fe
4b9d55e
 
 
 
82175fe
ba81a8e
 
82175fe
4b9d55e
 
82175fe
 
 
 
 
 
d1eca8b
 
 
 
 
 
 
 
 
 
 
 
4b9d55e
 
 
 
 
669a858
 
 
 
ba81a8e
669a858
 
ba81a8e
4b9d55e
 
 
 
 
669a858
 
 
 
ba81a8e
669a858
 
ba81a8e
4b9d55e
 
 
 
 
669a858
 
 
 
ba81a8e
669a858
ba81a8e
4b9d55e
 
d1eca8b
 
 
ba81a8e
d1eca8b
 
 
ba81a8e
d1eca8b
 
ba81a8e
d1eca8b
 
669a858
4b9d55e
 
ba81a8e
 
4b9d55e
 
 
ba81a8e
4b9d55e
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
## Docker Compose (Recommended)

!!! note

    I'm using newer Docker Compsose features. If you are using an older version of Docker Compose, you may need need to update.

Download the necessary Docker Compose files

=== "CUDA"

    ```bash
    curl --silent --remote-name https://raw.githubusercontent.com/speaches-ai/speaches/master/compose.yaml
    curl --silent --remote-name https://raw.githubusercontent.com/speaches-ai/speaches/master/compose.cuda.yaml
    export COMPOSE_FILE=compose.cuda.yaml
    ```

=== "CUDA (with CDI feature enabled)"

    ```bash
    curl --silent --remote-name https://raw.githubusercontent.com/speaches-ai/speaches/master/compose.yaml
    curl --silent --remote-name https://raw.githubusercontent.com/speaches-ai/speaches/master/compose.cuda.yaml
    curl --silent --remote-name https://raw.githubusercontent.com/speaches-ai/speaches/master/compose.cuda-cdi.yaml
    export COMPOSE_FILE=compose.cuda-cdi.yaml
    ```

=== "CPU"

    ```bash
    curl --silent --remote-name https://raw.githubusercontent.com/speaches-ai/speaches/master/compose.yaml
    curl --silent --remote-name https://raw.githubusercontent.com/speaches-ai/speaches/master/compose.cpu.yaml
    export COMPOSE_FILE=compose.cpu.yaml
    ```

Start the service

```bash
docker compose up --detach
```

??? note "Build from source"

    ```bash
    # NOTE: you need to install and enable [buildx](https://github.com/docker/buildx) for multi-platform builds

    # Build image with CUDA support
    docker compose --file compose.cuda.yaml build

    # Build image without CUDA support
    docker compose --file compose.cpu.yaml build
    ```

## Docker

=== "CUDA"

    ```bash
    docker run \
      --rm \
      --detach \
      --publish 8000:8000 \
      --name speaches \
      --volume hf-hub-cache:/home/ubuntu/.cache/huggingface/hub \
      --gpus=all \
      ghcr.io/speaches-ai/speaches:latest-cuda
    ```

=== "CUDA (with CDI feature enabled)"

    ```bash
    docker run \
      --rm \
      --detach \
      --publish 8000:8000 \
      --name speaches \
      --volume hf-hub-cache:/home/ubuntu/.cache/huggingface/hub \
      --device=nvidia.com/gpu=all \
      ghcr.io/speaches-ai/speaches:latest-cuda
    ```

=== "CPU"

    ```bash
    docker run \
      --rm \
      --detach \
      --publish 8000:8000 \
      --name speaches \
      --volume hf-hub-cache:/home/ubuntu/.cache/huggingface/hub \
      ghcr.io/speaches-ai/speaches:latest-cpu
    ```

??? note "Build from source"

    ```bash
    docker build --tag speaches .

    # NOTE: you need to install and enable [buildx](https://github.com/docker/buildx) for multi-platform builds
    # Build image for both amd64 and arm64
    docker buildx build --tag speaches --platform linux/amd64,linux/arm64 .

    # Build image without CUDA support
    docker build --tag speaches --build-arg BASE_IMAGE=ubuntu:24.04 .
    ```

## Python (requires Python 3.12+ and `uv` package manager)

```bash
git clone https://github.com/speaches-ai/speaches.git
cd speaches
uv venv
sourve .venv/bin/activate
uv sync --all-extras
uvicorn --factory --host 0.0.0.0 speaches.main:create_app
```