Hongyang Li

35
Papers
308
Total Citations
1
Affiliations

Affiliations

Peking University

Papers (35)

Generalized Predictive Model for Autonomous Driving

CVPR 2024arXiv
122
citations

LaneSegNet: Map Learning with Lane Segment Perception for Autonomous Driving

ICLR 2024arXiv
60
citations

Referred by Multi-Modality: A Unified Temporal Transformer for Video Object Segmentation

AAAI 2024arXiv
58
citations

Visual In-Context Prompting

CVPR 2024arXiv
52
citations

Detect Anything 3D in the Wild

ICCV 2025arXiv
12
citations

ETA: Efficiency through Thinking Ahead, A Dual Approach to Self-Driving with Large Models

ICCV 2025arXiv
4
citations

Decoupled Diffusion Sparks Adaptive Scene Generation

ICCV 2025
0
citations

Visual Point Cloud Forecasting enables Scalable Autonomous Driving

CVPR 2024arXiv
0
citations

FastMAC: Stochastic Spectral Sampling of Correspondence Graph

CVPR 2024arXiv
0
citations

Monocular 3D Object Detection: An Extrinsic Parameter Free Approach

CVPR 2021
0
citations

Towards Improving the Consistency, Efficiency, and Flexibility of Differentiable Neural Architecture Search

CVPR 2021arXiv
0
citations

Exploring intermediate representation for monocular vehicle pose estimation

CVPR 2021arXiv
0
citations

Align Representations With Base: A New Approach to Self-Supervised Learning

CVPR 2022
0
citations

BEVFormer v2: Adapting Modern Image Backbones to Bird's-Eye-View Recognition via Perspective Supervision

CVPR 2023
0
citations

Distilling Focal Knowledge From Imperfect Expert for 3D Object Detection

CVPR 2023
0
citations

Think Twice Before Driving: Towards Scalable Decoders for End-to-End Autonomous Driving

CVPR 2023arXiv
0
citations

Stare at What You See: Masked Image Modeling Without Reconstruction

CVPR 2023arXiv
0
citations

Planning-Oriented Autonomous Driving

CVPR 2023arXiv
0
citations

Lite DETR: An Interleaved Multi-Scale Encoder for Efficient DETR

CVPR 2023arXiv
0
citations

Scene as Occupancy

ICCV 2023arXiv
0
citations

Translating Images to Road Network: A Non-Autoregressive Sequence-to-Sequence Approach

ICCV 2023
0
citations

Density-invariant Features for Distant Point Cloud Registration

ICCV 2023arXiv
0
citations

DFA3D: 3D Deformable Attention For 2D-to-3D Feature Lifting

ICCV 2023arXiv
0
citations

DriveAdapter: Breaking the Coupling Barrier of Perception and Planning in End-to-End Autonomous Driving

ICCV 2023arXiv
0
citations

Detection Transformer with Stable Matching

ICCV 2023arXiv
0
citations

Point-Set Anchors for Object Detection, Instance Segmentation and Pose Estimation

ECCV 2020
0
citations

BEVFormer: Learning Bird’s-Eye-View Representation from Multi-Camera Images via Spatiotemporal Transformers

ECCV 2022
0
citations

DCL-Net: Deep Correspondence Learning Network for 6D Pose Estimation

ECCV 2022
0
citations

ST-P3: End-to-End Vision-Based Autonomous Driving via Spatial-Temporal Feature Learning

ECCV 2022
0
citations

PersFormer: 3D Lane Detection via Perspective Transformer and the OpenLane Benchmark

ECCV 2022
0
citations

ISTA-NAS: Efficient and Consistent Neural Architecture Search by Sparse Coding

NeurIPS 2020arXiv
0
citations

Sparse Steerable Convolutions: An Efficient Learning of SE(3)-Equivariant Features for Estimation and Tracking of Object Poses in 3D Space

NeurIPS 2021arXiv
0
citations

Trajectory-guided Control Prediction for End-to-end Autonomous Driving: A Simple yet Strong Baseline

NeurIPS 2022arXiv
0
citations

OpenLane-V2: A Topology Reasoning Benchmark for Unified 3D HD Mapping

NeurIPS 2023arXiv
0
citations

Leveraging Vision-Centric Multi-Modal Expertise for 3D Object Detection

NeurIPS 2023arXiv
0
citations