diff --git a/README.md b/README.md index 8bce1e9..0a64dfb 100755 --- a/README.md +++ b/README.md @@ -9,8 +9,6 @@ Speech-to-Text in golang. This is an early development version. ## Running -(Note: Docker images are not created yet - this is some forward planning!) - You can either run the whisper service as a CLI command or in a docker container. There are docker images for arm64 and amd64 (Intel). The arm64 image is built for Jetson GPU support specifically, but it will also run on Raspberry Pi's. @@ -27,11 +25,11 @@ The following command will run the server on port 8080 for an NVIDIA GPU: docker run \ --name whisper-server --rm \ --runtime nvidia --gpus all \ # When using a NVIDIA GPU - -v whisper:/models -p 8080:8080 -e WHISPER_DATA=/models \ - ghcr.io/mutablelogic/go-whisper + -v whisper:/data -p 8080:80 \ + ghcr.io/mutablelogic/go-whisper:latest ``` -If you include a `-debug` flag at the end, you'll get more verbose output. The API is then +The API is then available at `http://localhost:8080/v1` and it generally conforms to the [OpenAI API](https://platform.openai.com/docs/api-reference/audio) spec. @@ -71,19 +69,19 @@ There's more information on the API [here](doc/API.md). ## Building -If you are building a docker image, you just need Docker installed: +If you are building a docker image, you just need make and docker installed: * `DOCKER_REGISTRY=docker.io/user make docker` - builds a docker container with the server binary, tagged to a specific registry -If you want to build the server yourself for your specific combination of hardware, -you can use the `Makefile` in the root directory and have the following dependencies +If you want to build the server yourself for your specific combination of hardware (for example, +on MacOS), you can use the `Makefile` in the root directory and have the following dependencies met: * Go 1.22 * C++ compiler * FFmpeg 6.1 libraries (see [here](doc/build.md) for more information) -* For CUDA, you'll need the CUDA toolkit including the `nvcc` compiler +* For CUDA, you'll need the CUDA toolkit installed including the `nvcc` compiler The following `Makefile` targets can be used: diff --git a/etc/Dockerfile b/etc/Dockerfile index 699c980..4fb8fbe 100644 --- a/etc/Dockerfile +++ b/etc/Dockerfile @@ -1,4 +1,4 @@ -ARG BASE_TAG=1.0.1 +ARG BASE_TAG=1.0.2 ARG BASE_DEV_CONTAINER=ghcr.io/mutablelogic/cuda-dev:${BASE_TAG} ARG BASE_RUN_CONTAINER=ghcr.io/mutablelogic/cuda-rt:${BASE_TAG} ARG CUDA_DOCKER_ARCH=all @@ -50,4 +50,4 @@ ENTRYPOINT [ "/entrypoint.sh" ] STOPSIGNAL SIGQUIT EXPOSE 80 VOLUME [ "/data" ] -CMD [ "/usr/local/bin/whisper", "--dir", "/data", "--listen", ":80", "--endpoint", "/api/v1", "server" ] +CMD [ "/usr/local/bin/whisper", "server", "--dir", "/data", "--listen", ":80", "--endpoint", "/v1" ]