Skip to content

Add support for int8 quantization backend#37

Draft
silveroxides wants to merge 3 commits into
Comfy-Org:mainfrom
silveroxides:feature/int8-tensorwise
Draft

Add support for int8 quantization backend#37
silveroxides wants to merge 3 commits into
Comfy-Org:mainfrom
silveroxides:feature/int8-tensorwise

Conversation

@silveroxides
Copy link
Copy Markdown

This draft adds support for int8 quantized models with full backend support and with optimized matmul kernels using triton.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant