Dynamic Camera Poses and Where to Find Them
Chris Rockwell, Joseph Tung, Tsung-Yi Lin, Ming-Yu Liu, David F. Fouhey, Chen-Hsuan Lin
2025-04-25
Summary
This paper talks about DynPose-100K, a huge collection of real-world internet videos where each video has detailed information about how the camera was moving while filming.
What's the problem?
The problem is that most existing video datasets are either too small, use fake or staged videos, or only cover very specific situations like driving or pets. It's also really hard to figure out the camera's exact movement in real, dynamic videos because most internet videos aren't suitable for this kind of analysis, and tracking the camera in scenes with lots of movement is very challenging.
What's the solution?
The researchers solved this by creating a careful process to pick out only the best real-world videos for camera pose estimation, using a mix of expert models and AI to filter out bad ones. They then used advanced techniques to accurately calculate the camera's position and movement for each video, ending up with over 100,000 diverse, high-quality clips that are much more useful than previous datasets.
Why it matters?
This matters because having such a large and varied dataset helps scientists and engineers train and test new AI systems for things like video editing, special effects, robotics, and even self-driving cars, making these technologies smarter and more reliable.
Abstract
DynPose-100K is a large-scale dataset of dynamic Internet videos with annotated camera poses, collected using advanced filtering and pose estimation techniques.