The third Workshop on
Reconstruction of Human-Object Interactions (RHOBIN)


rhobin-logo

June 12, 0855 - 1230 @ CVPR 2025, Nashville TN, USA
Music City Center


  • Speakers
  • Schedule
  • Papers & Winners
  • Contact
  • 2nd RHOBIN @ CVPR'24

News

  • [Feb. 7, 2025] Rhobin says hello CVPR'25!

Introduction

Following the success of the first and second Rhobin workshop at CVPR'23 and CVPR'24, this third half-day Rhobin workshop will continue providing a venue to present and discuss state-of-the-art research in the reconstruction of human-object interactions from images. The focus of this third workshop will go beyond image-based interaction reconstruction, extend to interaction tracking over time, and seek connections to relevant topics such as egocentric vision and dynamic scene interactions. The third Rhobin challenge will feature five tracks in total with tasks on human-object interaction tracking and image-based contact estimation, using datasets InterCap, DAMON, and BEHAVE.

Humans are in constant contact with the world as they move through it and interact with it. To better understand how humans interact with the world, it is crucial to estimate human poses, shapes, and movements. 3D Human Pose and Motion estimation from images or videos has attracted a lot of interest. However, in most cases, the task does not explicitly involve objects and the interaction with them. Whether it is 2D detection and/or monocular 3D reconstruction, objects and humans have been mostly studied separately. Considering the interaction between them can marry the best of both worlds.

Participation details of the Rhobin Challenge and paper submission can be found below.

Invited Speakers (Check out the Full Schedule)

Federica Bogo

Eric Yi

Angela Yao

Jiajun Wu

Meta
Tsinghua University (IIIS)
National University of Singapore
Stanford University

Call for Papers

In this workshop, we invite papers on topics related to human-centered interaction modeling. This could include, but is not limited to:

  • Estimation of 3D human pose and shape from images or video
  • 3D human motion prediction
  • Interactive motion sequence generation
  • Interactive motion sequence generation/synthesis
  • Shape reconstruction from a single image
  • Object 6-DoF pose estimation and tracking
  • Human-centered object semantics and functionality modeling
  • Joint reconstruction of bodies and objects/scenes
  • Interaction modeling between humans and objects, e.g., contact, physics properties
  • Detection of human-object interaction semantics
  • New datasets or benchmarks that have 3D annotations of both humans and objects/scenes
  • Detection of holistic and/or semantic contact and downstream applications
  • Estimation and modeling of physics and biomechanical properties
  • Tracking human-object interactions in 2D/3D
  • Template free interaction reconstruction
  • Semantic 3D contact estimation

We invite submissions of a maximum of 8 pages, excluding references, using the CVPR template. Submissions should follow CVPR 2025 instructions. All papers will be subject to a double-blind review process, i.e. authors must not identify themselves on the submitted papers. The reviewing process is single-stage without rebuttals. We also invite 1-page abstract submissions of already published works or relevants works in progress.

Submission Instructions


Submissions are anonymous and should not include any author names, affiliations, and contact information in the PDF.
  • Online Submission System: https://openreview.net/group?id=thecvf.com/CVPR/2025/Workshop/RHOBIN
If you have any questions, feel free to reach out to us.

Timeline Table (11:59 PM, Pacific Time)

  • Submissions open: February 20, 2025
  • Submission deadline: March 17, 2025
  • Notification to authors: March 31, 2025
  • Camera-ready deadline: April 7, 2025

The Third Rhobin Challenge

Given the importance of human-object interaction, as also highlighted by this workshop, we propose a challenge on reconstructing 3D human and object and estimating 3D human-object and human-scene contact, from monocular RGB images. We have seen promising progress in reconstructing human body meshes or estimating 6DoF object pose from single images. However, most of these works focus on occlusion-free images which are not realistic for settings during close human-object interaction since humans and objects occlude each other. This makes inference more difficult and poses challenges to existing state-of-the-art methods. Similarly, methods estimating 3D contacts have also seen rapid progress, but are restricted to scanned or synthetic datasets, and struggle with generalization to in-the-wild scenarios. In this workshop, we want to examine how well the existing human and object reconstruction and contact estimation methods work under more realistic settings and more importantly, understand how they can benefit each other for accurate interaction reasoning. The recently released BEHAVE (CVPR'22), InterCap (GCPR’22) and DAMON (ICCV’23) datasets enable joint reasoning about human-object interactions in real settings and evaluating contact prediction in the wild.

Challenge website

  • 3D human reconstruction track
  • Object 6DoF pose estimation track
  • Joint human object reconstruction
  • Template-free interaction reconstruction
  • Tracking human-object interaction
  • 3D human contact prediction
  • Semantic 3D contact prediction from RGB images

Important dates

  • Challenge open: TBD
  • Submission deadline: TBD
  • Winner award: TBD

Workshop Organizers

Xi Wang

Xianghui Xie

Nikos Athanasiou

ETH Zurich & TUM
Max Planck Institute for Informatics,
Saarland Informatics Campus
Max-Planck-Institute for Intelligent Systems,
Tübingen

Dimitrios Tzionas

Bharat Lal Bhatnagar

Alexey Gavryushin

University of Amsterdam
Research Scientist, Meta Reality Labs
ETH Zurich

Thiemo Alldieck

Muhammed Kocabas

Luc Van Gool

Google DeepMind
Max-Planck-Institute for Intelligent Systems,
Tübingen & Meshcapade
INSAIT, Sofia University "St. Kliment Ohridski",
Bulgaria & ETH Zurich

Marc Pollefeys

Gerard Pons-Moll

ETH Zurich & Microsoft
University of Tübingen
Max Planck Institute for Informatics,
Saarland Informatics Campus

Challenge Organizers

Xianghui Xie

Shashank Tripathi

Nikos Athanasiou

Max Planck Institute for Informatics, Saarland Informatics Campus
Max-Planck-Institute for Intelligent Systems, Tübingen
Max-Planck-Institute for Intelligent Systems,
Tübingen

Dimitrios Tzionas

Chengfeng Zhao

Juze Zhang

University of Amsterdam
ShanghaiTech University, China
ShanghaiTech University, China

Lan Xu

Gerard Pons-Moll

ShanghaiTech University, China
University of Tübingen
Max Planck Institute for Informatics,
Saarland Informatics Campus

Committee Members

Nikos Athanasiou, MPI Intelligent Systems, Germany
Dimitrije Antic, University of Amsterdam, Netherlands
George Paschalidis, University of Amsterdam, Netherlands
Sammy Christen, Disney Research, Switzerland
Zicong Fan, ETH Zurich, Switzerland
Bharat Lal Bhatnagar, Meta, Switzerland
Muhammed Kocabas, Meshcapade, Germany
Otmar Hilliges, ETH Zurich, Switzerland
Chun-Hao Paul Huang, Adobe London, UK
Yannan He, University of Tübingen, Germany
Jiaxi Jiang, ETH Zurich, Switzerland
Chuqiao Li, University of Tübingen, Germany
Bowen Wen, NVIDIA, USA
Kaichun Mo, NVIDIA, USA
Ilya Petrov, University of Tübingen, Germany
Gerard Pons-Moll, University of Tübingen and MPI, Germany
Omid Taheri, MPI-IS, Germany
Purva Tendulkar, Columbia University, USA
Anh N. Thai, Georgia Tech, USA
Shashank Tripathi, MPI Intelligent System, Germany
Dimitrios Tzionas, University of Amsterdam, Netherlands
Julien Valentin, Microsoft, Switzerland
Luc Van Gool, ETH Zurich, Switzerland
Xi Wang, ETH Zurich, Switzerland
Xianghui Xie, MPI, Germany
Yuxuan Xue, University of Tübingen, Germany
Xiaohan Zhang, University of Tübingen, Germany
Keyang Zhou, University of Tübingen, Germany
Alexey Gavryushin, ETH Zurich, Switzerland
Marc Pollefeys, ETH Zurich, Switzerland
Chenfeng Zhao, ShanghaiTech University
Lan Xu, ShanghaiTech University
Juze Zhang, ShanghaiTech University

Contact Info

E-mail: rhobinchallenge@gmail.com

Sponsors

Acknowledgements

Website template borrowed from: https://futurecv.github.io/ (Thanks to Deepak Pathak)