s_tworkowski pfp
s_tworkowski
@2039900203990082
Introducing LongLLaMA ๐Ÿฆ™, an unlimited-context version of OpenLLaMA fine-tuned at 8k & capable of extrapolating to 256k tokens! We train it using our new Focused Transformer ๐ŸŽฏ technique (FoT). No degradation on short context, drop-in compatibility & Apache 2.0 license ๐Ÿ”ฅ๐Ÿ”ฅ ๐Ÿงต https://t.co/QiNl5xNYvl
0 reply
0 recast
0 reaction