Taiji Suzuki

18
Papers
51
Total Citations

Papers (18)

Symmetric Mean-field Langevin Dynamics for Distributional Minimax Problems

ICLR 2024
13
citations

Flow matching achieves almost minimax optimal convergence

ICLR 2025arXiv
12
citations

On the Optimization and Generalization of Two-layer Transformers with Sign Gradient Descent

ICLR 2025
9
citations

State Space Models are Provably Comparable to Transformers in Dynamic Token Selection

ICLR 2025arXiv
6
citations

Koopman-based generalization bound: New aspect for full-rank weights

ICLR 2024
6
citations

Propagation of Chaos for Mean-Field Langevin Dynamics and its Application to Model Ensemble

ICML 2025
2
citations

Weighted Point Set Embedding for Multimodal Contrastive Learning Toward Optimal Similarity Metric

ICLR 2025
1
citations

Quantifying Memory Utilization with Effective State-Size

ICML 2025
1
citations

Trained Mamba Emulates Online Gradient Descent in In-Context Linear Regression

NeurIPS 2025
1
citations

Mean-field Analysis on Two-layer Neural Networks from a Kernel Perspective

ICML 2024
0
citations

Mean Field Langevin Actor-Critic: Faster Convergence and Global Optimality beyond Lazy Learning

ICML 2024
0
citations

Provably Neural Active Learning Succeeds via Prioritizing Perplexing Samples

ICML 2024
0
citations

High-Dimensional Kernel Methods under Covariate Shift: Data-Dependent Implicit Regularization

ICML 2024
0
citations

Transformers Learn Nonlinear Features In Context: Nonconvex Mean-field Dynamics on the Attention Landscape

ICML 2024
0
citations

State-Free Inference of State-Space Models: The *Transfer Function* Approach

ICML 2024
0
citations

Mechanistic Design and Scaling of Hybrid Architectures

ICML 2024
0
citations

SILVER: Single-loop variance reduction and application to federated learning

ICML 2024
0
citations

How do Transformers Perform In-Context Autoregressive Learning ?

ICML 2024
0
citations