Skip to content

[BugFix]add int8 cache dtype when using attention quantization#128

Open
Angazenn wants to merge 6 commits intovllm-project:mainfrom Angazenn:bug_fix

Commits

Commits on Feb 21, 2025

Commits on Feb 22, 2025

  • clean code

    author
    angazenn
    committed