Update README.md
Browse files
README.md
CHANGED
@@ -207,19 +207,19 @@ print(scores.tolist())
|
|
207 |
|
208 |
### Text Embeddings Inference (TEI) Usage
|
209 |
|
210 |
-
You can either deploy TEI on
|
211 |
|
212 |
```bash
|
213 |
docker run -p 8080:80 -v hf_cache:/data --pull always ghcr.io/huggingface/text-embeddings-inference:cpu-1.7.2 --model-id Qwen/Qwen3-Embedding-0.6B
|
214 |
```
|
215 |
|
216 |
-
Or on
|
217 |
|
218 |
```bash
|
219 |
docker run --gpus all -p 8080:80 -v hf_cache:/data --pull always ghcr.io/huggingface/text-embeddings-inference:1.7.2 --model-id Qwen/Qwen3-Embedding-0.6B --dtype float16
|
220 |
```
|
221 |
|
222 |
-
And then generate embeddings sending a HTTP POST request as:
|
223 |
|
224 |
```bash
|
225 |
curl http://localhost:8080/embed \
|
|
|
207 |
|
208 |
### Text Embeddings Inference (TEI) Usage
|
209 |
|
210 |
+
You can either run / deploy TEI on NVIDIA GPUs as:
|
211 |
|
212 |
```bash
|
213 |
docker run -p 8080:80 -v hf_cache:/data --pull always ghcr.io/huggingface/text-embeddings-inference:cpu-1.7.2 --model-id Qwen/Qwen3-Embedding-0.6B
|
214 |
```
|
215 |
|
216 |
+
Or on CPU devices as:
|
217 |
|
218 |
```bash
|
219 |
docker run --gpus all -p 8080:80 -v hf_cache:/data --pull always ghcr.io/huggingface/text-embeddings-inference:1.7.2 --model-id Qwen/Qwen3-Embedding-0.6B --dtype float16
|
220 |
```
|
221 |
|
222 |
+
And then, generate the embeddings sending a HTTP POST request as:
|
223 |
|
224 |
```bash
|
225 |
curl http://localhost:8080/embed \
|