- Thomas Wolf / Hugging Face Cofounder
- Cameron Wolfe / Researcher at Netflix722.This is a really nice framing of efficiency constraints in RL training.›LLM score 92 · about 2 months ago
- Ted Sanders / OpenAI Researcher723.it's wild how much LLMs have transformed the profession of software engineering over the past year.›LLM score 85 · about 2 months ago
- Kevin Murphy / DeepMind Research Scientist724.I'm pleased to share our paper on learning temporally abstract world models and policies (options).›LLM score 95 · about 2 months ago
- Sholto Douglas / Researcher at Anthropic725.Default case right now is a software only singularity, we need to scale robots and automated labs dramatically in 28/29, or the physical world will fall far behind the digital one - and the US won’t be competitive unless we put in the investment now (fab, solar panel, actuator›LLM score 92 · about 2 months ago
- Richard Song / DeepMind Research Scientist726.“Am I disappointed when my students don’t go off and be mathematicians?›LLM score 85 · about 2 months ago
- Sholto Douglas / Researcher at Anthropic727.Long weekends are good for GDP because people discover how good the models are https://t.co/yucnlrRos4LLM score 85 · about 2 months ago
- Sherwin Wu / OpenAI API, Head of Engineering728.This article really resonated with me, especially in relation to GPT-5.3-Codex.›LLM score 85 · about 2 months ago
- Geoffrey Hinton729.This is a great report that provides a thoughtful, detailed and very well researched description of the risks of AI.›LLM score 92 · about 2 months ago
- Andrew White / Edison Scientific Cofounder730.
- Andrew White / Edison Scientific Cofounder731.Opus 4.6 is out today - which we've been using for a while.›LLM score 82 · about 2 months ago
- Behnam Neyshabur / Anthropic Researcher732.I've left Anthropic to start something new. 🧵 https://t.co/6VzY1T3ivNLLM score 75 · about 2 months ago
- Andrew White / Edison Scientific Cofounder
- Clive Chan / OpenAI Hardware734.hardware-software co-design, years in the making! https://t.co/y08aoW5tbRLLM score 92 · about 2 months ago
- Chi Jin / OpenAI Researcher735.Honored to have contributed to substantial token efficiency gains powering the GPT Codex release! https://t.co/HnefSO90oiLLM score 75 · about 2 months ago
- Andrew Lampinen / Research Scientist at DeepMind
- Clive Chan / OpenAI Hardware737.why does nobody use the same benchmarks 😭 only overlapping benchmark is TerminalBench 2.0 https://t.co/FmpdXgwQMFLLM score 85 · about 2 months ago
- Emmanuel Ameisen / Anthropic Interpretability Researcher738.
- Aidan McLaughlin / OpenAI Research Scientist739.we have an internal codex usage leaderboard and karel is like 10xing everyone else on our team›LLM score 85 · about 2 months ago
- Noam Brown / OpenAI Research Scientist740.GPT-5.3-Codex's much better token efficiency *AND* faster inference is the biggest story of this release.›LLM score 85 · about 2 months ago
- Boris Cherny / Creator of Claude Code741.Out now: Teams, aka. Agent Swarms in Claude Code Team are experimental, and use a lot of tokens.›LLM score 75 · about 2 months ago
- Boris Cherny / Creator of Claude Code742.I've been using Opus 4.6 for a bit -- it is our best model yet.›LLM score 92 · about 2 months ago
- Igor Babuschkin / Cofounder of xAI743.Goodfire is one of the many amazing companies I met through Babuschkin Ventures.›LLM score 25 · about 2 months ago
- alphaXiv744."Learning to Reason in 13 Parameters" This paper propose TinyLoRA, a type of LoRA that can scale to sizes as small as 1 parameter.›LLM score 92 · about 2 months ago
- Ben Burtenshaw / Hugging Face Researcher745.PSA: you define the leaderboards on the hub.›LLM score 92 · about 2 months ago
- Kevin Murphy / DeepMind Research Scientist746.I just read the first two of these and like them a lot.›LLM score 92 · about 2 months ago
- Kevin Murphy / DeepMind Research Scientist747.hot take: when reviewing research papers, we should take into account not just the content (elegant ideas, compelling results, etc), but also the brevity - a 1 page paper with "one weird trick" should get more points than one which rambles on for 8 pages with woo woo.›LLM score 92 · about 2 months ago
- Richard Song / DeepMind Research Scientist748.Vizier was my former team 🙂 - thanks for the shoutout! Its tuning was so efficient because its Gaussian Process regressor continually updates against lots of (x,y) evaluations to produce an estimate of the global objective landscape, rather than pure unguided evolution.›LLM score 92 · about 2 months ago
- Clive Chan / OpenAI Hardware
- Noam Brown / OpenAI Research Scientist750.GPT-5.2 evals are finally out for METR and it's state-of-the-art.›LLM score 92 · about 2 months ago