Discussion about this post

User's avatar
Andriy Batutin's avatar

Generalization gives me nervous tick. So many times people especially management means by that AI that does everything perfectly from the first try no need to worry about anything else ever again. But that rarely true. Glad to see the thoughts on evolution of this term

Venkateshan K's avatar

"We've long been beyond the traditional statistical understanding, of generalizing to new samples from the "same distribution". Today, you can come up with a completely novel phrasing, or misspelling of the query, and LLMs will still respond correctly"

That's true but I am not sure that the interpretation of that is accurate. Statistical learning theory has a narrow formulation in terms of input output pairs (X,y) in the train and test data drawn from the same distribution.

With current LLMs, during inference we have test-time compute - CoT, process reward model, etc - that breaks the symmetry between train and test contexts and hence classical generalization theory cannot be applied. But that theory still probably holds at single step level; P(y_t| y_{t-1}, ..y_1,x)

No posts

Ready for more?