🧬Efficiency

Distributed Training

Training across multiple devices

75 papers209 total citations
Compare with other topics
Feb '24 Jan '2675 papers
Also includes: distributed training, distributed optimization, parallel training, data parallel

Top Papers

#1

TorchTitan: One-stop PyTorch native solution for production ready LLM pretraining

Wanchao Liang, Tianyu Liu, Less Wright et al.

ICLR 2025
52
citations
#2

Stretching Each Dollar: Diffusion Training from Scratch on a Micro-Budget

Vikash Sehwag, Xianghao Kong, Jingtao Li et al.

CVPR 2025arXiv:2407.15811
text-to-image generationdiffusion modelstransformer architecturepatch masking+4
26
citations
#3

BooW-VTON: Boosting In-the-Wild Virtual Try-On via Mask-Free Pseudo Data Training

Xuanpu Zhang, Dan Song, pengxin zhan et al.

CVPR 2025
22
citations
#4

Training on the Benchmark Is Not All You Need

Shiwen Ni, Xiangtao Kong, Chengming Li et al.

AAAI 2025
21
citations
#5

CO2: Efficient Distributed Training with Full Communication-Computation Overlap

Weigao Sun, Qin Zhen, Weixuan Sun et al.

ICLR 2024
15
citations
#6

Mobile Video Diffusion

Haitam Ben Yahia, Denis Korzhenkov, Ioannis Lelekas et al.

ICCV 2025
12
citations
#7

Accelerating Neural Field Training via Soft Mining

Shakiba Kheradmand, Daniel Rebain, Gopal Sharma et al.

CVPR 2024
12
citations
#8

Integrated Augmented and Virtual Reality Technologies for Realistic Fire Drill Training

Hosan Kang, Jinseong Yang, Beom-Seok Ko et al.

ISMAR 2025
11
citations
#9

Bridging Training and Execution via Dynamic Directed Graph-Based Communication in Cooperative Multi-Agent Systems

Zhuohui Zhang, Bin He, Bin Cheng et al.

AAAI 2025
6
citations
#10

Poplar: Efficient Scaling of Distributed DNN Training on Heterogeneous GPU Clusters

WenZheng Zhang, Yang Hu, Jing Shi et al.

AAAI 2025
5
citations
#11

A Quadratic Synchronization Rule for Distributed Deep Learning

Xinran Gu, Kaifeng Lyu, Sanjeev Arora et al.

ICLR 2024
4
citations
#12

Faster and Better 3D Splatting via Group Training

Chengbo Wang, Guozheng Ma, Yizhen Lao et al.

ICCV 2025arXiv:2412.07608
3d gaussian splattingnovel view synthesisscene reconstructiontraining efficiency+2
3
citations
#13

Decoupling Training-Free Guided Diffusion by ADMM

Youyuan Zhang, Zehua Liu, Zenan Li et al.

CVPR 2025
3
citations
#14

Everywhere Attack: Attacking Locally and Globally to Boost Targeted Transferability

Hui Zeng, Sanshuai Cui, Biwei Chen et al.

AAAI 2025
3
citations
#15

PseudoMapTrainer: Learning Online Mapping without HD Maps

Christian Löwens, Thorben Funke, Jingchao Xie et al.

ICCV 2025
2
citations
#16

On-Device Diffusion Transformer Policy for Efficient Robot Manipulation

Yiming Wu, Huan Wang, Zhenghao Chen et al.

ICCV 2025
2
citations
#17

STraj: Self-training for Bridging the Cross-Geography Gap in Trajectory Prediction

Zhanwei Zhang, Minghao Chen, Zhihong Gu et al.

AAAI 2025
2
citations
#18

Distributed Semantic Segmentation with Efficient Joint Source and Task Decoding

Danish Nazir, Timo Bartels, Jan Piewek et al.

ECCV 2024
2
citations
#19

ACCO: Accumulate While You Communicate for Communication-Overlapped Sharded LLM Training

Adel Nabli, Louis Fournier, Pierre ERBACHER et al.

NeurIPS 2025
2
citations
#20

InstaTrain: Adaptive Training via Ultra-Fast Natural Annealing within Dynamical Systems

Chuan Liu, Ruibing Song, Chunshu Wu et al.

ICLR 2025
1
citations
#21

COGNATE: Acceleration of Sparse Tensor Programs on Emerging Hardware using Transfer Learning

Chamika Sudusinghe, Gerasimos Gerogiannis, Damitha Lenadora et al.

ICML 2025
1
citations
#22

CroCoDL: Cross-device Collaborative Dataset for Localization

Hermann Blum, Alessandro Mercurio, Joshua O'Reilly et al.

CVPR 2025
1
citations
#23

EDiT: A Local-SGD-Based Efficient Distributed Training Method for Large Language Models

Jialiang Cheng, Ning Gao, Yun Yue et al.

ICLR 2025
1
citations
#24

HAT Swapping: Virtual Agents as Stand-ins for Absent Human Instructors in Virtual Training

Jingjing Zhang, Binyang Han, Ze Dong et al.

ISMAR 2025
not collected
#25

Shared, Replicated, or Separated? A Comparative Study of Virtual Workspace Configurations for Collaborative Hands-On Learning

Juliette Le Meudec, Anastasia Bezerianos, Arnaud Prouzeau

ISMAR 2025
not collected
#26

Cross-Reality for Autonomous Mobility in Autistic Individuals: Evaluating At-Home VR Training and In-Situ AR Support in a Field Study

Francesco Vona, Mattia Gianotti, Maria Luisa Lorusso et al.

ISMAR 2025
not collected
#27

Can People's Brains Synchronize during Remote AR Collaboration?

Jaehwan You, Myeongul Jung, Kwanguk Kim

ISMAR 2025
not collected
#28

Ghidorah: Towards Robust Multi-Scale Information Diffusion Prediction via Test-Time Training

Wenting Zhu, Chaozhuo Li, Litian Zhang et al.

AAAI 2025
not collected
#29

The Indoor-Training Effect: Unexpected Gains from Distribution Shifts in the Transition Function

Serena Bono, Spandan Madan, Ishaan Grover et al.

AAAI 2025
not collected
#30

Complexity of Neural Network Training and ETR: Extensions with Effectively Continuous Functions

Teemu Hankala, Miika Hannula, Juha Kontinen et al.

AAAI 2024
not collected
#31

Heterogeneous Test-Time Training for Multi-Modal Person Re-identification

Zi Wang, Huaibo Huang, Aihua Zheng et al.

AAAI 2024
not collected
#32

High-Fidelity Gradient Inversion in Distributed Learning

Zipeng Ye, Wenjian Luo, Qi Zhou et al.

AAAI 2024
not collected
#33

Communication Efficient Distributed Newton Method over Unreliable Networks

Ming Wen, Chengchang Liu, Yuedong Xu

AAAI 2024
not collected
#34

Don’t Drop Your Samples! Coherence-Aware Training Benefits Conditional Diffusion

Nicolas Dufour, Victor Besnier, Vicky Kalogeiton et al.

CVPR 2024
not collected
#35

MS-DETR: Efficient DETR Training with Mixed Supervision

Chuyang Zhao, Yifan Sun, Wenhao Wang et al.

CVPR 2024
not collected
#36

Enhancing Neural Training via a Correlated Dynamics Model

Jonathan Brokman, Roy Betser, Rotem Turjeman et al.

ICLR 2024
not collected
#37

Flag Aggregator: Scalable Distributed Training under Failures and Augmented Losses using Convex Optimization

Hamidreza Almasi, Harsh Mishra, Balajee Vamanan et al.

ICLR 2024
not collected
#38

OmniBal: Towards Fast Instruction-Tuning for Vision-Language Models via Omniverse Computation Balance

Yongqiang Yao, Jingru Tan, Feizhao Zhang et al.

ICML 2025
not collected
#39

CTBench: A Library and Benchmark for Certified Training

Yuhao Mao, Stefan Balauca, Martin Vechev

ICML 2025
not collected
#40

DexScale: Automating Data Scaling for Sim2Real Generalizable Robot Control

Guiliang Liu, Yueci Deng, Runyi Zhao et al.

ICML 2025
not collected
#41

SMART-PC: Skeletal Model Adaptation for Robust Test-Time Training in Point Clouds

Ali Bahri, Moslem Yazdanpanah, Sahar Dastani Oghani et al.

ICML 2025
not collected
#42

HALoS: Hierarchical Asynchronous Local SGD over Slow Networks for Geo-Distributed Large Language Model Training

Geon-Woo Kim, Junbo Li, Shashidhar Gandham et al.

ICML 2025
not collected
#43

IT$^3$: Idempotent Test-Time Training

Nikita Durasov, Assaf Shocher, Doruk Oner et al.

ICML 2025
not collected
#44

Beyond Low-rank Decomposition: A Shortcut Approach for Efficient On-Device Learning

Le-Trung Nguyen, Aël Quélennec, Van-Tam Nguyen et al.

ICML 2025
not collected
#45

ATA: Adaptive Task Allocation for Efficient Resource Management in Distributed Machine Learning

Artavazd Maranjyan, El Mehdi Saad, Peter Richtarik et al.

ICML 2025
not collected
#46

Distributed Conformal Prediction via Message Passing

Haifeng Wen, Hong XING, Osvaldo Simeone

ICML 2025
not collected
#47

NDOT: Neuronal Dynamics-based Online Training for Spiking Neural Networks

Haiyan Jiang, Giulia De Masi, Huan Xiong et al.

ICML 2024
spiking neural networkssurrogate gradient methodback-propagation through timeonline training+4
not collected
#48

Sparse Cocktail: Every Sparse Pattern Every Sparse Ratio All At Once

Zhangheng Li, Shiwei Liu, Tianlong Chen et al.

ICML 2024
sparse neural networkssparsity patternssparsity ratiossparse co-training+3
not collected
#49

Position: Exploring the Robustness of Pipeline-Parallelism-Based Decentralized Training

Lin Lu, Chenxi Dai, Wangcheng Tao et al.

ICML 2024
decentralized trainingpipeline parallelismrobustness analysispoisoning attacks+4
not collected
#50

Multiply Robust Estimation for Local Distribution Shifts with Multiple Domains

Steven Wilkins-Reeves, Xu Chen, Qi Ma et al.

ICML 2024
distribution shiftsmultiple domainsrobust estimationtabular data analysis+4
not collected
#51

Fault Tolerant ML: Efficient Meta-Aggregation and Synchronous Training

Tehila Dahan, Kfir Levy

ICML 2024
byzantine-robust trainingdistributed machine learningmeta-aggregator designgradient estimation techniques+2
not collected
#52

APT: Adaptive Personalized Training for Diffusion Models with Limited Data

JungWoo Chae, Jiyoon Kim, Jaewoong Choi et al.

CVPR 2025
not collected
#53

Towards a Better Theoretical Understanding of Independent Subnetwork Training

Egor Shulgin, Peter Richtarik

ICML 2024
independent subnetwork trainingdistributed computingcommunication compressionmodel parallelism+3
not collected
#54

Synchronizing Task Behavior: Aligning Multiple Tasks during Test-Time Training

Wooseong Jeong, Jegyeong Cho, Youngho Yoon et al.

ICCV 2025
not collected
#55

FedEL: Federated Elastic Learning for Heterogeneous Devices

Letian Zhang, Bo Chen, Jieming Bian et al.

NeurIPS 2025
not collected
#56

From Promise to Practice: Realizing High-performance Decentralized Training

Zesen Wang, Jiaojiao Zhang, Xuyang Wu et al.

ICLR 2025
not collected
#57

DistRL: An Asynchronous Distributed Reinforcement Learning Framework for On-Device Control Agent

Taiyi Wang, Zhihao Wu, Jianheng Liu et al.

ICLR 2025
not collected
#58

Rethinking the Starting Point: Collaborative Pre-Training for Federated Downstream Tasks

Yun-Wei Chu, Dong-Jun Han, Seyyedali Hosseinalipour et al.

AAAI 2025
not collected
#59

Multitarget Device-Free Localization via Cross-Domain Wi-Fi RSS Training Data and Attentional Prior Fusion

Na FAN, Zeyue Tian, Amartansh DUBEY et al.

AAAI 2024
not collected
#60

Harnessing Manycore Processors with Distributed Memory for Accelerated Training of Sparse and Recurrent Models

Jan Finkbeiner, Thomas Gmeinder, Mark Pupilli et al.

AAAI 2024
not collected
#61

TinyTrain: Resource-Aware Task-Adaptive Sparse Training of DNNs at the Data-Scarce Edge

Young Kwon, Rui Li, Stylianos Venieris et al.

ICML 2024
on-device trainingsparse trainingresource-aware optimizationmicrocontroller unit deployment+4
not collected
#62

DyMO: Training-Free Diffusion Model Alignment with Dynamic Multi-Objective Scheduling

Xin Xie, Dong Gong

CVPR 2025
not collected
#63

HOT: Hadamard-based Optimized Training

Seonggon Kim, Juncheol Shin, Seung-taek Woo et al.

CVPR 2025
not collected
#64

TREAD: Token Routing for Efficient Architecture-agnostic Diffusion Training

Felix Krause, Timy Phan, Ming Gui et al.

ICCV 2025
not collected
#65

CoStoDet-DDPM: Collaborative Training of Stochastic and Deterministic Models Improves Surgical Workflow Anticipation and Recognition

Kaixiang Yang, Xin Li, Qiang Li et al.

ICCV 2025
not collected
#66

Mixtures of Subspaces for Bandwidth Efficient Context Parallel Training

Sameera Ramasinghe, Thalaiyasingam Ajanthan, Hadi Mohaghegh Dolatabadi et al.

NeurIPS 2025
not collected
#67

Analog In-memory Training on General Non-ideal Resistive Elements: The Impact of Response Functions

Zhaoxian Wu, Quan Xiao, Tayfun Gokmen et al.

NeurIPS 2025
not collected
#68

MeCeFO: Enhancing LLM Training Robustness via Fault-Tolerant Optimization

Rizhen Hu, Yutong He, Ran Yan et al.

NeurIPS 2025
not collected
#69

AC-DiT: Adaptive Coordination Diffusion Transformer for Mobile Manipulation

Sixiang Chen, Jiaming Liu, Siyuan Qian et al.

NeurIPS 2025
not collected
#70

Unveiling the Power of Multiple Gossip Steps: A Stability-Based Generalization Analysis in Decentralized Training

NeurIPS 2025arXiv:2510.07980
decentralized traininggeneralization error analysismulti-gossip stepsstability analysis+4
not collected
#71

DUO: No Compromise to Accuracy Degradation

Jinda Jia, Cong Xie, Hanlin Lu et al.

NeurIPS 2025
not collected
#72

Accelerating neural network training: An analysis of the AlgoPerf competition

Priya Kasimbeg, Frank Schneider, Runa Eschenhagen et al.

ICLR 2025
not collected
#73

NExUME: Adaptive Training and Inference for DNNs under Intermittent Power Environments

Cyan Subhra Mishra, Deeksha Chaudhary, Jack Sampson et al.

ICLR 2025
not collected
#74

An Asynchronous Bundle Method for Distributed Learning Problems

Daniel Cederberg, Xuyang Wu, Stephen Boyd et al.

ICLR 2025
not collected
#75

Learning to Help in Multi-Class Settings

Yu Wu, Yansong Li, Zeyu Dong et al.

ICLR 2025
not collected