Skip to content

Commit 7a6fad6

Browse files
update readme
1 parent 3b56d76 commit 7a6fad6

File tree

2 files changed

+2
-2
lines changed

2 files changed

+2
-2
lines changed

README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -155,7 +155,7 @@ model=meta-llama/Llama-2-7b-chat-hf
155155
volume=$PWD/data # share a volume with the Docker container to avoid downloading weights every run
156156
token=<your cli READ token>
157157

158-
docker run --gpus all --shm-size 1g -e HUGGING_FACE_HUB_TOKEN=$token -p 8080:80 -v $volume:/data ghcr.io/huggingface/text-generation-inference:1.0.3 --model-id $model
158+
docker run --gpus all --shm-size 1g -e HUGGING_FACE_HUB_TOKEN=$token -p 8080:80 -v $volume:/data ghcr.io/huggingface/text-generation-inference:1.1.0 --model-id $model
159159
```
160160

161161
### A note on Shared Memory (shm)

docs/source/quicktour.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -85,7 +85,7 @@ curl 127.0.0.1:8080/generate \
8585
To see all possible deploy flags and options, you can use the `--help` flag. It's possible to configure the number of shards, quantization, generation parameters, and more.
8686

8787
```bash
88-
docker run ghcr.io/huggingface/text-generation-inference:1.0.3 --help
88+
docker run ghcr.io/huggingface/text-generation-inference:1.1.0 --help
8989
```
9090

9191
</Tip>

0 commit comments

Comments
 (0)