NEURIPS "training dynamics" Papers

9 papers found

Beyond Random: Automatic Inner-loop Optimization in Dataset Distillation

Muquan Li, Hang Gou, Dongyang Zhang et al.

NEURIPS 2025posterarXiv:2510.04838
1
citations

Bridging Critical Gaps in Convergent Learning: How Representational Alignment Evolves Across Layers, Training, and Distribution Shifts

Chaitanya Kapoor, Sudhanshu Srivastava, Meenakshi Khosla

NEURIPS 2025posterarXiv:2502.18710
1
citations

Contrastive Learning with Data Misalignment: Feature Purity, Training Dynamics and Theoretical Generalization Guarantees

Jiawei Sun, Shuai Zhang, Hongkang Li et al.

NEURIPS 2025poster

Flatness is Necessary, Neural Collapse is Not: Rethinking Generalization via Grokking

Ting Han, Linara Adilova, Henning Petzka et al.

NEURIPS 2025oralarXiv:2509.17738
3
citations

Memorization in Graph Neural Networks

Adarsh Jamadandi, Jing Xu, Adam Dziedzic et al.

NEURIPS 2025posterarXiv:2508.19352

Scaling Off-Policy Reinforcement Learning with Batch and Weight Normalization

Daniel Palenicek, Florian Vogt, Joe Watson et al.

NEURIPS 2025posterarXiv:2502.07523
8
citations

The emergence of sparse attention: impact of data distribution and benefits of repetition

Nicolas Zucchet, Francesco D'Angelo, Andrew Lampinen et al.

NEURIPS 2025oralarXiv:2505.17863
6
citations

Trained Mamba Emulates Online Gradient Descent in In-Context Linear Regression

Jiarui Jiang, Wei Huang, Miao Zhang et al.

NEURIPS 2025posterarXiv:2509.23779
1
citations

Why Diffusion Models Don’t Memorize: The Role of Implicit Dynamical Regularization in Training

Tony Bonnaire, Raphaël Urfin, Giulio Biroli et al.

NEURIPS 2025oral