PhD Proposal: Motion Modeling from Human to the World Around Us
IRB IRB-5137
Motion is a fundamental aspect of how humans and animals interact with the world, making motion modeling central to understanding dynamic phenomena. With recent advances in deep learning, research on motion modeling has rapidly advanced, spanning both reconstruction from real-world observations and the generation of new motions.
This proposal investigates how to model realistic motion conditioned on context. We first propose ShapeMove, a text-driven human motion synthesis framework that explicitly incorporates body shape. Unlike prior methods that assume a canonical body model, ShapeMove predicts both motion and body shape parameters, enabling the synthesis of shape-aware motions that capture how the same action is performed differently across diverse body morphologies. Next, we introduce PAD3R, a framework for reconstructing dynamic, arbitrary objects from casual monocular videos without relying on predefined templates. PAD3R disentangles object motion from camera motion and leverages long-term point tracking to regularize non-rigid deformations, enabling template-free dynamic 3D reconstruction directly from video.
Together, these contributions bridge generation and reconstruction under a unified lens of context, advancing controllable motion modeling across humans and general dynamic objects.