Skip to content

Add independent K scale support in FP8 prefill attention#40

Open
xueyangcs wants to merge 1 commit intoTencent:mainfrom
xueyangcs:feature/add_k_scale_for_prefill_attn
Open

Add independent K scale support in FP8 prefill attention#40
xueyangcs wants to merge 1 commit intoTencent:mainfrom
xueyangcs:feature/add_k_scale_for_prefill_attn

Conversation

@xueyangcs
Copy link
Copy Markdown
Contributor

Summary

Split the combined qkscale parameter into separate qscale and kscale in the FP8 prefill attention kernel, allowing Q and K to use different quantization granularities.

Breaking Change:

qkscale: [num_batch, num_head_q, max_seqlens_q_pad], float32 is replaced by:
qscale: [num_batch, num_head_q, max_seqlens_q_pad], float32 and
kscale: [1], float32

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant