TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/HOI4ABOT: Human-Object Interaction Anticipation for Human ...

HOI4ABOT: Human-Object Interaction Anticipation for Human Intention Reading Collaborative roBOTs

Esteve Valls Mascaro, Daniel Sliwowski, Dongheui Lee

2023-09-28Human-Object Interaction DetectionHuman-Object Interaction Anticipation
PaperPDF

Abstract

Robots are becoming increasingly integrated into our lives, assisting us in various tasks. To ensure effective collaboration between humans and robots, it is essential that they understand our intentions and anticipate our actions. In this paper, we propose a Human-Object Interaction (HOI) anticipation framework for collaborative robots. We propose an efficient and robust transformer-based model to detect and anticipate HOIs from videos. This enhanced anticipation empowers robots to proactively assist humans, resulting in more efficient and intuitive collaborations. Our model outperforms state-of-the-art results in HOI detection and anticipation in VidHOI dataset with an increase of 1.76% and 1.04% in mAP respectively while being 15.4 times faster. We showcase the effectiveness of our approach through experimental results in a real robot, demonstrating that the robot's ability to anticipate HOIs is key for better Human-Robot Interaction. More information can be found on our project webpage: https://evm7.github.io/HOI4ABOT_page/

Results

TaskDatasetMetricValueModel
Human-Object Interaction DetectionVidHOIDetection: Full (mAP@0.5)11.12HOI4ABOT
Human-Object Interaction DetectionVidHOIDetection: Non-Rare (mAP@0.5)18.48HOI4ABOT
Human-Object Interaction DetectionVidHOIDetection: Rare (mAP@0.5)5.61HOI4ABOT
Human-Object Interaction DetectionVidHOIOracle: Full (mAP@0.5)40.37HOI4ABOT
Human-Object Interaction DetectionVidHOIOracle: Non-Rare (mAP@0.5)54.52HOI4ABOT
Human-Object Interaction DetectionVidHOIOracle: Rare (mAP@0.5)29.5HOI4ABOT
Human-Object Interaction AnticipationVidHOIPerson-wise Top5: t=1(mAP@0.5)37.77HOI4ABOT
Human-Object Interaction AnticipationVidHOIPerson-wise Top5: t=3(mAP@0.5)34.75HOI4ABOT
Human-Object Interaction AnticipationVidHOIPerson-wise Top5: t=5(mAP@0.5)34.07HOI4ABOT

Related Papers

RoHOI: Robustness Benchmark for Human-Object Interaction Detection2025-07-12Bilateral Collaboration with Large Vision-Language Models for Open Vocabulary Human-Object Interaction Detection2025-07-09VolumetricSMPL: A Neural Volumetric Body Model for Efficient Interactions, Contacts, and Collisions2025-06-29HOIverse: A Synthetic Scene Graph Dataset With Human Object Interactions2025-06-24On the Robustness of Human-Object Interaction Detection against Distribution Shift2025-06-22Egocentric Human-Object Interaction Detection: A New Benchmark and Method2025-06-17InterActHuman: Multi-Concept Human Animation with Layout-Aligned Audio Conditions2025-06-11HunyuanVideo-HOMA: Generic Human-Object Interaction in Multimodal Driven Human Animation2025-06-10