Releases: lucidrains/ring-attention-pytorch
Releases · lucidrains/ring-attention-pytorch
0.5.20
16 May 00:15
Compare
Sorry, something went wrong.
No results found
fix maximum update in triton flash attn
0.5.19
25 Oct 15:22
Compare
Sorry, something went wrong.
No results found
0.5.18
23 Oct 19:56
Compare
Sorry, something went wrong.
No results found
address https://github.com/lucidrains/ring-attention-pytorch/issues/21
0.5.17
15 Aug 18:30
Compare
Sorry, something went wrong.
No results found
0.5.12
14 Aug 22:09
Compare
Sorry, something went wrong.
No results found
0.5.10
14 Aug 19:56
Compare
Sorry, something went wrong.
No results found
keep numerator and denominator in float32 for tree attn decoding
0.5.9
14 Aug 17:46
Compare
Sorry, something went wrong.
No results found
some asserts and cleanup for tree attn decoding
0.5.8
14 Aug 17:34
Compare
Sorry, something went wrong.
No results found
make tree attention decoding work with triton flash attention forward
0.5.6
14 Aug 17:14
Compare
Sorry, something went wrong.
No results found
account for scenario where keys and values are already sharded in tre…
0.5.5
13 Aug 17:29
Compare
Sorry, something went wrong.
No results found