Taiji Suzuki
18
Papers
51
Total Citations
Papers (18)
Symmetric Mean-field Langevin Dynamics for Distributional Minimax Problems
ICLR 2024
13
citations
Flow matching achieves almost minimax optimal convergence
ICLR 2025arXiv
12
citations
On the Optimization and Generalization of Two-layer Transformers with Sign Gradient Descent
ICLR 2025
9
citations
State Space Models are Provably Comparable to Transformers in Dynamic Token Selection
ICLR 2025arXiv
6
citations
Koopman-based generalization bound: New aspect for full-rank weights
ICLR 2024
6
citations
Propagation of Chaos for Mean-Field Langevin Dynamics and its Application to Model Ensemble
ICML 2025
2
citations
Weighted Point Set Embedding for Multimodal Contrastive Learning Toward Optimal Similarity Metric
ICLR 2025
1
citations
Quantifying Memory Utilization with Effective State-Size
ICML 2025
1
citations
Trained Mamba Emulates Online Gradient Descent in In-Context Linear Regression
NeurIPS 2025
1
citations
Mean-field Analysis on Two-layer Neural Networks from a Kernel Perspective
ICML 2024
0
citations
Mean Field Langevin Actor-Critic: Faster Convergence and Global Optimality beyond Lazy Learning
ICML 2024
0
citations
Provably Neural Active Learning Succeeds via Prioritizing Perplexing Samples
ICML 2024
0
citations
High-Dimensional Kernel Methods under Covariate Shift: Data-Dependent Implicit Regularization
ICML 2024
0
citations
Transformers Learn Nonlinear Features In Context: Nonconvex Mean-field Dynamics on the Attention Landscape
ICML 2024
0
citations
State-Free Inference of State-Space Models: The *Transfer Function* Approach
ICML 2024
0
citations
Mechanistic Design and Scaling of Hybrid Architectures
ICML 2024
0
citations
SILVER: Single-loop variance reduction and application to federated learning
ICML 2024
0
citations
How do Transformers Perform In-Context Autoregressive Learning ?
ICML 2024
0
citations