You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
1. Make sure your `MODELDIR` exists on the node where your workload is schedueled so you can cache the downloaded model for next time use. Otherwise, set `global.modelUseHostPath` to 'null' if you don't want to cache the model.
34
+
35
+
## Verify
36
+
37
+
To verify the installation, run the command `kubectl get pod` to make sure all pods are running.
38
+
39
+
Then run the command `kubectl port-forward svc/chatqna 8888:8888` to expose the service for access.
40
+
41
+
Open another terminal and run the following command to verify the service if working:
42
+
43
+
```console
44
+
curl http://localhost:8888/v1/chatqna \
45
+
-H "Content-Type: application/json" \
46
+
-d '{"messages": "What is the revenue of Nike in 2023?"}'
| global.HUGGINGFACEHUB_API_TOKEN | string |`""`| Your own Hugging Face API token |
28
-
| global.modelUseHostPath | string |`"/mnt/opea-models"`| Cached models directory, tgi will not download if the model is cached here. The host path "volume" will be mounted to container as /data directory |
29
-
| tgi.LLM_MODEL_ID | string |`"Intel/neural-chat-7b-v3-3"`| Models id from https://huggingface.co/, or predownloaded model directory |
| global.HUGGINGFACEHUB_API_TOKEN | string |`""`| Your own Hugging Face API token |
56
+
| global.modelUseHostPath | string |`"/mnt/opea-models"`| Cached models directory, tgi will not download if the model is cached here. The host path "modelUseHostPath" will be mounted to container as /data directory. Set this to empty/null will force it to download models every time.|
57
+
| tgi.LLM_MODEL_ID | string |`"Intel/neural-chat-7b-v3-3"`| Models id from https://huggingface.co/, or predownloaded model directory |
0 commit comments