HOTFoundation Models & LLMs

Scaling Laws in LLMs

Discussions focus on the effectiveness of scaling compute and data for LLMs, with some evidence of diminishing returns and efficiency gains. This sub-topic explores how scaling drives progress but raises concerns about sustainability.

Key Players: Percy Liang, Bernhard Schölkopf
Lost in the Middle: How Language Models Use Long Contexts by Percy Liang (2024, 648 citations)

6

Related Opinions

30

Related Papers

4

KOLs Discussing

Amjad Masad
Amjad MasadReplitSupportive

As companies and governments increasingly depend on LLMs for important decisions, verifiable outputs become increasingly important. Great demo!

2/21/2026 Source
Trevor Darrell
Trevor DarrellUC BerkeleyNeutral

A truely generative meta-model of activations, for steering, probing, and understanding LLMs at scale!

2/9/2026 Source
Yarin Gal
Yarin GalUniversity of OxfordNeutral

The dangers of extrapolating scaling laws

1/12/2026 Source
Sergey Levine
Sergey LevineUC BerkeleyNeutral

Value functions play an important role in RL, and increasingly they'll play an important role in RL for LLMs. This new paper led by @rohin_manvi is one step in this direction: using value functions to optimize test-time compute with adaptive computation.

12/30/2025 Source
Trevor Darrell
Trevor DarrellUC BerkeleyNeutral

Debug your model with StringSight: LLMs all the way down!

12/17/2025 Source
Trevor Darrell
Trevor DarrellUC BerkeleyNeutral

Super excited about our new work on pretrained 4-D robotic foundation models. LLMs learned with 4-D representations on egocentric datasets transfer well to real world tasks!

2/24/2025 Source