AI Lab Tracker
Labs
Timeline
Engram: Conditional Memory via Scalable Lookup
paper
2026-01-12
DeepSeek
A new axis of sparsity for large language models. Builds upon finding from Geva et al "Transformer Feed-Forward Layers Are Key-Value Memories" (2021).
Paper (arXiv)
Video from Prompt Engineering
Paper
arXiv:
2601.07372
architecture
sparsity
research