-
Notifications
You must be signed in to change notification settings - Fork 73
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
tgi: revert xeon version to 2.2.0 #328
Conversation
Due to bug opea-project/GenAIExamples#636, revert tgi xeon version to 2.2.0. Signed-off-by: Lianhao Lu <lianhao.lu@intel.com>
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
lgtm
The reason why gmc xeon test fails is because the model used by gmc chatqna CI test |
Signed-off-by: Lianhao Lu <lianhao.lu@intel.com>
note: TGI 2.2.0 cannot work with bigscience/bloom-560m |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
thanks for fixing this!!!
Whereas 2.0.x & 2.1.x versions regressed TGI Prometheus support so one does not get any token metrics out of them (although they worked fine with TGI 1.x). That was fixed only in 2.2.0: huggingface/text-generation-inference#2184 |
* tgi: revert xeon version to 2.2.0 Due to bug opea-project/GenAIExamples#636, revert tgi xeon version to 2.2.0. * Retriever: Fix missing HF_TOKEN issue of v0.9 retriever-redis image Signed-off-by: Lianhao Lu <lianhao.lu@intel.com> (cherry picked from commit 076e81e)
Description
Due to bug opea-project/GenAIExamples#636, revert tgi xeon version to 2.2.0.
Issues
opea-project/GenAIExamples#636
Type of change
List the type of change like below. Please delete options that are not relevant.
Dependencies
n/a.
Tests
Describe the tests that you ran to verify your changes.