Skip to content

[BugFix]add int8 cache dtype when using attention quantization #143

[BugFix]add int8 cache dtype when using attention quantization

[BugFix]add int8 cache dtype when using attention quantization #143

Annotations

2 errors

mypy (3.9)

cancelled Feb 21, 2025 in 2m 42s