NEURIPS "training dynamics" Papers
9 papers found
Beyond Random: Automatic Inner-loop Optimization in Dataset Distillation
Muquan Li, Hang Gou, Dongyang Zhang et al.
NEURIPS 2025posterarXiv:2510.04838
1
citations
Bridging Critical Gaps in Convergent Learning: How Representational Alignment Evolves Across Layers, Training, and Distribution Shifts
Chaitanya Kapoor, Sudhanshu Srivastava, Meenakshi Khosla
NEURIPS 2025posterarXiv:2502.18710
1
citations
Contrastive Learning with Data Misalignment: Feature Purity, Training Dynamics and Theoretical Generalization Guarantees
Jiawei Sun, Shuai Zhang, Hongkang Li et al.
NEURIPS 2025poster
Flatness is Necessary, Neural Collapse is Not: Rethinking Generalization via Grokking
Ting Han, Linara Adilova, Henning Petzka et al.
NEURIPS 2025oralarXiv:2509.17738
3
citations
Memorization in Graph Neural Networks
Adarsh Jamadandi, Jing Xu, Adam Dziedzic et al.
NEURIPS 2025posterarXiv:2508.19352
Scaling Off-Policy Reinforcement Learning with Batch and Weight Normalization
Daniel Palenicek, Florian Vogt, Joe Watson et al.
NEURIPS 2025posterarXiv:2502.07523
8
citations
The emergence of sparse attention: impact of data distribution and benefits of repetition
Nicolas Zucchet, Francesco D'Angelo, Andrew Lampinen et al.
NEURIPS 2025oralarXiv:2505.17863
6
citations
Trained Mamba Emulates Online Gradient Descent in In-Context Linear Regression
Jiarui Jiang, Wei Huang, Miao Zhang et al.
NEURIPS 2025posterarXiv:2509.23779
1
citations
Why Diffusion Models Don’t Memorize: The Role of Implicit Dynamical Regularization in Training
Tony Bonnaire, Raphaël Urfin, Giulio Biroli et al.
NEURIPS 2025oral