We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Forward output nan, GatedLinearAttention
o, recurrent_state = fused_chunk_gla( q=q, k=k, v=v, g=gk, initial_state=recurrent_state, output_final_state=use_cache, head_first=False ) if torch.isnan(o).any(): 1==1
head = 1 seq_len=43884 dim=64
There is no nan in the forward output
The text was updated successfully, but these errors were encountered:
@980202006 Hello, could you provide the input tensors resulting in NaNs?
Sorry, something went wrong.
BTW, it's recomended to utilize chunk mode instead of fused_chunk
chunk
fused_chunk
[GLA] Fix potential exp overflows in fused_chunk (#122)
458c018
@980202006 Could you check 458c018 again?
yzhangcs
No branches or pull requests
Describe the Bug
Forward output nan, GatedLinearAttention
Steps to Reproduce the Bug
head = 1
seq_len=43884
dim=64
Expected Behavior
There is no nan in the forward output
Environment Information
The text was updated successfully, but these errors were encountered: