Correct alignment in the seq_len diagram. (#5592)
Co-authored-by: Liqian Chen <liqian.chen@deeplang.ai>
This commit is contained in:
parent
9333fb8eb9
commit
9e74d9d003
@ -83,7 +83,7 @@ class FlashAttentionMetadata(AttentionMetadata):
|
|||||||
# |---------------- N iteration ---------------------|
|
# |---------------- N iteration ---------------------|
|
||||||
# |- tokenA -|......................|-- newTokens ---|
|
# |- tokenA -|......................|-- newTokens ---|
|
||||||
# |---------- context_len ----------|
|
# |---------- context_len ----------|
|
||||||
# |-------------------- seq_len ----------------------|
|
# |-------------------- seq_len ---------------------|
|
||||||
# |-- query_len ---|
|
# |-- query_len ---|
|
||||||
|
|
||||||
# Maximum query length in the batch. None for decoding.
|
# Maximum query length in the batch. None for decoding.
|
||||||
|
|||||||
Loading…
Reference in New Issue
Block a user