[BugFix]add int8 cache dtype when using attention quantization#128
Open
Angazenn wants to merge 6 commits intovllm-project:mainfrom Angazenn:bug_fix
+212
Commits
Commits on Feb 21, 2025
- committedangazenn
- committedangazenn
- committedangazenn
- committedangazenn
- committedangazenn
Commits on Feb 22, 2025
- committedangazenn