Unlock: Grokking
Models can memorize training data quickly, then generalize much later after continued training. This delayed generalization, called grokking, breaks the assumption that overfitting is a terminal state and connects to weight decay, implicit regularization, and phase transitions in learning.
310 Prerequisites0 Mastered0 Working236 Gaps
Prerequisite mastery24%
Recommended probe
Peano Axioms is your weakest prerequisite with available questions. You haven't been assessed on this topic yet.
GrokkingTARGET
Not assessed5 questions
No quiz
Not assessed4 questions
Not assessed16 questions
No quiz
Not assessed4 questions
Sign in to track your mastery and see personalized gap analysis.