I’m proud to share our new paper in
@natcomputsci.nature.com :
www.nature.com/articles/s43....
We show that attention in LLMs can be accelerated with analog in-memory computing using Gain Cell circuits. Simuating a 1.5B-parameter model we achieve up to 70 000× lower energy and 100× speedup vs GPUs.