File size: 279 Bytes
1164ff2
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
---
language:
- en
---


This is a pure sub-quadtratic linear attention 70B parameter model, linearized from the Meta Llama 3.1 70B model starting point.

Details on this model and how to train your own are provided at: https://github.com/HazyResearch/lolcats/tree/lolcats-scaled