Kristen Grauman
15
Papers
146
Total Citations
Papers (15)
PerceptionLM: Open-Access Data and Models for Detailed Visual Understanding
NeurIPS 2025arXiv
40
citations
Learning Object State Changes in Videos: An Open-World Perspective
CVPR 2024
33
citations
Action2Sound: Ambient-Aware Generation of Action Sounds from Egocentric Videos
ECCV 2024
19
citations
SoundingActions: Learning How Actions Sound from Narrated Egocentric Videos
CVPR 2024
11
citations
ExpertAF: Expert Actionable Feedback from Video
CVPR 2025
11
citations
Learning Spatial Features from Audio-Visual Correspondence in Egocentric Videos
CVPR 2024
8
citations
Progress-Aware Video Frame Captioning
CVPR 2025
7
citations
Detours for Navigating Instructional Videos
CVPR 2024
7
citations
When Thinking Drifts: Evidential Grounding for Robust Video Reasoning
NeurIPS 2025
4
citations
FIction: 4D Future Interaction Prediction from Video
CVPR 2025
3
citations
Which Viewpoint Shows it Best? Language for Weakly Supervising View Selection in Multi-view Instructional Videos
CVPR 2025arXiv
3
citations
Switch-a-View: View Selection Learned from Unlabeled In-the-wild Videos
ICCV 2025
0
citations
Viewpoint Rosetta Stone: Unlocking Unpaired Ego-Exo Videos for View-invariant Representation Learning
CVPR 2025
0
citations
Ego-Exo4D: Understanding Skilled Human Activity from First- and Third-Person Perspectives
CVPR 2024
0
citations
Learning Skill-Attributes for Transferable Assessment in Video
NeurIPS 2025
0
citations