Skip to content

Releases: lucidrains/ring-attention-pytorch

0.5.20

16 May 00:15

Choose a tag to compare

fix maximum update in triton flash attn

0.5.19

25 Oct 15:22

Choose a tag to compare

What's Changed

Full Changelog: 0.5.18...0.5.19

0.5.18

23 Oct 19:56

Choose a tag to compare

address https://github.com/lucidrains/ring-attention-pytorch/issues/21

0.5.17

15 Aug 18:30

Choose a tag to compare

Full Changelog: 0.5.10...0.5.17

0.5.12

14 Aug 22:09

Choose a tag to compare

Full Changelog: 0.5.10...0.5.12

0.5.10

14 Aug 19:56

Choose a tag to compare

keep numerator and denominator in float32 for tree attn decoding

0.5.9

14 Aug 17:46

Choose a tag to compare

some asserts and cleanup for tree attn decoding

0.5.8

14 Aug 17:34

Choose a tag to compare

make tree attention decoding work with triton flash attention forward

0.5.6

14 Aug 17:14

Choose a tag to compare

account for scenario where keys and values are already sharded in tre…

0.5.5

13 Aug 17:29

Choose a tag to compare

Full Changelog: 0.5.4...0.5.5