Omid Taheri
Omid Taheri
Home
News
Publications
Contact
CV
Thesis
Light
Dark
Automatic
1
CWGrasp: 3D Whole-Body Grasp Synthesis with Directional Controllability
Abstract Synthesizing 3D whole bodies that realistically grasp objects is crucial for animation, mixed reality, and robotics. Key challenges include natural coordination between hand, body, and environment, and the scarcity of training data.
Georgios Paschalidis
,
Romana Wilschut
,
Dimitrije Antić
,
Omid Taheri
,
Dimitrios Tzionas
Cite
arXiv
Video
Code
3DV 2025
HUMOS: Human Motion Model Conditioned on Body Shape
Abstract Generating realistic human motion is crucial for many computer vision and graphics applications. The rich diversity of human body shapes and sizes significantly influences how people move. However, existing motion models typically overlook these differences, using a normalized, average body instead.
Shashank Tripathi
,
Omid Taheri
,
Christoph Lassner
,
Michael Black
,
Daniel Holden
,
Carsten Stoll
Cite
PDF
arXiv
Video
Code
ECCV2024
WANDR: Intention-guided Human Motion Generation
Abstract Synthesizing natural human motions that enable a 3D human avatar to walk and reach for arbitrary goals in 3D space remains an unsolved problem with many applications. Existing methods (data-driven or using reinforcement learning) are limited in terms of generalization and motion naturalness.
Markos Diomataris
,
Nikos Athanasiou
,
Omid Taheri
,
Xi Wang
,
Otmar Hilliges
,
Michael J. Black
Cite
PDF
arXiv
Video
Code
Project
CVPR2024
GRIP: Generating Interaction Poses Using Spatial Cues and Latent Consistency
Abstract Hands are dexterous and highly versatile manipulators that are central to how humans interact with objects and their environment. Consequently, modeling realistic hand-object interactions, including the subtle motion of individual fingers, is critical for applications in computer graphics, computer vision, and mixed reality.
Omid Taheri
,
Yi Zhou
,
Dimitrios Tzionas
,
Yang Zhou
,
Duygu Ceylan
,
Soren Pirk
,
Michael J. Black
Cite
PDF
arXiv
Video
Code
Poster
Project
3DV2024
ARCTIC: A Dataset for Dexterous Bimanual Hand-Object Manipulation
Abstract Humans intuitively understand that inanimate objects do not move by themselves, but that state changes are typically caused by human manipulation (e.g., the opening of a book). This is not yet the case for machines.
Zicong Fan
,
Omid Taheri
,
Dimitrios Tzionas
,
Muhammed Kocabas
,
Manuel Kaufmann
,
Michael J. Black
,
Otmar Hilliges
Cite
PDF
arXiv
Video
Code/Competition
Data
Project
CVPR2023
IPMAN: 3D Human Pose Estimation via Intuitive Physics
Abstract The estimation of 3D human body shape and pose from images has advanced rapidly. While the results are often well aligned with image features in the camera view, the 3D pose is often physically implausible; bodies lean, float, or penetrate the floor.
Shashank Tripathi
,
Lea Müller
,
Chun-Hao P. Huang
,
Omid Taheri
,
Michael Black
,
Dimitrios Tzionas
Cite
PDF
Video
Code
Data (MoYo)
Poster
Project
CVPR2023
InterCap: Joint Markerless 3D Tracking of Humans and Objects in Interaction
Abstract Humans constantly interact with objects to accomplish tasks. To understand such interactions, computers need to reconstruct these in 3D from images of whole bodies manipulating objects, e.g., for grasping, moving, and using the latter.
Yinghao Huang
,
Omid Taheri
,
Michael J. Black
,
Dimitrios Tzionas
Cite
PDF
arXiv
Video
Code
Data
Poster
Project
GCPR2022
GOAL: Generating 4D Whole-Body Motion for Hand-Object Grasping
Abstract Generating digital humans that move realistically has many applications and is widely studied, but existing methods focus on the major limbs of the body, ignoring the hands and head. Hands have been separately studied but the focus has been on generating realistic static grasps of objects.
Omid Taheri
,
Vasileios Choutas
,
Michael J. Black
,
Dimitrios Tzionas
Cite
PDF
arXiv
Video
Code
Poster
Project
CVPR2022
GRAB: A Dataset of Whole-Body Human Grasping of Objects
Abstract Training computers to understand, model, and synthesize human grasping requires a rich dataset containing complex 3D object shapes, detailed contact information, hand pose and shape, and the 3D body motion over time.
Omid Taheri
,
Nima Ghorbani
,
Michael J. Black
,
Dimitrios Tzionas
Cite
arXiv
Video
GRAB
GrabNet
Data
Project
ECCV2020
«
Cite
×