[Doc] Change total -> total_q
This commit is contained in:
parent
9610114ce8
commit
e8a0b4acdd
@ -279,7 +279,7 @@ def flash_attn_unpadded_kvpacked_func(q, kv, cu_seqlens_q, cu_seqlens_k, max_seq
|
||||
(they might not have the right scaling).
|
||||
deterministic: bool. Whether or not to ensure deterministic execution.
|
||||
Return:
|
||||
out: (total, nheads, headdim).
|
||||
out: (total_q, nheads, headdim).
|
||||
softmax_lse [optional, if return_attn_probs=True]: (batch_size, nheads, seqlen). The
|
||||
logsumexp of each row of the matrix QK^T * scaling (e.g., log of the softmax
|
||||
normalization factor).
|
||||
@ -315,7 +315,7 @@ def flash_attn_unpadded_func(q, k, v, cu_seqlens_q, cu_seqlens_k, max_seqlen_q,
|
||||
(they might not have the right scaling).
|
||||
deterministic: bool. Whether or not to ensure deterministic execution.
|
||||
Return:
|
||||
out: (total, nheads, headdim).
|
||||
out: (total_q, nheads, headdim).
|
||||
softmax_lse [optional, if return_attn_probs=True]: (batch_size, nheads, seqlen). The
|
||||
logsumexp of each row of the matrix QK^T * scaling (e.g., log of the softmax
|
||||
normalization factor).
|
||||
|
||||
Loading…
Reference in New Issue
Block a user