Skip to content

Releases: lucidrains/block-recurrent-transformer-pytorch

0.2.1

04 Apr 03:47
Compare
Choose a tag to compare
take care of removing padding from the blocked keys and values of the…

0.2.0

04 Apr 03:28
Compare
Choose a tag to compare
update the state block by block, rather than one segment at a time, t…

0.1.4

31 Mar 14:22
Compare
Choose a tag to compare
actually use the recurrent states instead of always initial state, th…

0.1.2

26 Mar 01:56
Compare
Choose a tag to compare
cache causal mask and rotary positional embeddings

0.1.1

24 Mar 22:53
Compare
Choose a tag to compare
bug with xpos for rotary embeddings

0.1.0

24 Mar 22:16
Compare
Choose a tag to compare
it turns out flash attention in pytorch 2.0 is not handling causal co…

0.0.19

24 Mar 20:55
Compare
Choose a tag to compare
switch to rotary positional embedding with xpos, to prepare for flash…

0.0.18

26 Feb 17:20
Compare
Choose a tag to compare
allow for returning memories and states during training, for potentia…

0.0.17

22 Feb 16:40
Compare
Choose a tag to compare
0.0.17

0.0.16

22 Feb 16:21
Compare
Choose a tag to compare
add single headed kv