Skip to content

[BugFix]add int8 cache dtype when using attention quantization #158

[BugFix]add int8 cache dtype when using attention quantization

[BugFix]add int8 cache dtype when using attention quantization #158