AI300
top
n1sum
n1sen
past
about
Cameron Wolfe / Researcher at Netflix
1.
Trinity Large is a sparse 400B-A13B MoE with 256 experts (4 active per token), chosen for a ~30-day training sprint that also yields blazing speed plu
LLM score 85 · about 21 hours ago
Andrew Lampinen / Research Scientist at DeepMind
2.
New paper shows language model representations of factuality evolve dramatically in edge-case conversations on consciousness or delusions.
LLM score 85 · about 16 hours ago
alphaXiv
3.
Paper introduces Self-Distillation Fine-Tuning (SDFT) for on-policy continual learning from expert demonstrations without explicit reward inference or engineering.
LLM score 85 · about 15 hours ago
Alex Zhang / MIT CSAIL PhD
4.
Released RLM-Qwen3-8B, the first natively recursive language model (at tiny scale), post-trained on just ~1000 RLM trajectories from unrelated domains.
LLM score 85 · about 13 hours ago
Omar Khattab / MIT CSAIL Asst professor
5.
Late Interaction post-trains RLM-Qwen3-8B at tiny scale into the first natively recursive LM, releases it, and shows promise on unseen tasks.
LLM score 85 · about 12 hours ago