TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/CholecTriplet2021: A benchmark challenge for surgical acti...

CholecTriplet2021: A benchmark challenge for surgical action triplet recognition

Chinedu Innocent Nwoye, Deepak Alapatt, Tong Yu, Armine Vardazaryan, Fangfang Xia, Zixuan Zhao, Tong Xia, Fucang Jia, Yuxuan Yang, Hao Wang, Derong Yu, Guoyan Zheng, Xiaotian Duan, Neil Getty, Ricardo Sanchez-Matilla, Maria Robu, Li Zhang, Huabin Chen, Jiacheng Wang, Liansheng Wang, Bokai Zhang, Beerend Gerats, Sista Raviteja, Rachana Sathish, Rong Tao, Satoshi Kondo, Winnie Pang, Hongliang Ren, Julian Ronald Abbing, Mohammad Hasan Sarhan, Sebastian Bodenstedt, Nithya Bhasker, Bruno Oliveira, Helena R. Torres, Li Ling, Finn Gaida, Tobias Czempiel, João L. Vilaça, Pedro Morais, Jaime Fonseca, Ruby Mae Egging, Inge Nicole Wijma, Chen Qian, GuiBin Bian, Zhen Li, Velmurugan Balasubramanian, Debdoot Sheet, Imanol Luengo, Yuanbo Zhu, Shuai Ding, Jakob-Anton Aschenbrenner, Nicolas Elini van der Kar, Mengya Xu, Mobarakol Islam, Lalithkumar Seenivasan, Alexander Jenke, Danail Stoyanov, Didier Mutter, Pietro Mascagni, Barbara Seeliger, Cristians Gonzalez, Nicolas Padoy

2022-04-10Action DetectionAction Triplet RecognitionActivity Recognition
PaperPDFCodeCode(official)CodeCodeCodeCode

Abstract

Context-aware decision support in the operating room can foster surgical safety and efficiency by leveraging real-time feedback from surgical workflow analysis. Most existing works recognize surgical activities at a coarse-grained level, such as phases, steps or events, leaving out fine-grained interaction details about the surgical activity; yet those are needed for more helpful AI assistance in the operating room. Recognizing surgical actions as triplets of <instrument, verb, target> combination delivers comprehensive details about the activities taking place in surgical videos. This paper presents CholecTriplet2021: an endoscopic vision challenge organized at MICCAI 2021 for the recognition of surgical action triplets in laparoscopic videos. The challenge granted private access to the large-scale CholecT50 dataset, which is annotated with action triplet information. In this paper, we present the challenge setup and assessment of the state-of-the-art deep learning methods proposed by the participants during the challenge. A total of 4 baseline methods from the challenge organizers and 19 new deep learning algorithms by competing teams are presented to recognize surgical action triplets directly from surgical videos, achieving mean average precision (mAP) ranging from 4.2% to 38.1%. This study also analyzes the significance of the results obtained by the presented approaches, performs a thorough methodological comparison between them, in-depth result analysis, and proposes a novel ensemble method for enhanced recognition. Our analysis shows that surgical workflow analysis is not yet solved, and also highlights interesting directions for future research on fine-grained surgical activity recognition which is of utmost importance for the development of AI in surgery.

Results

TaskDatasetMetricValueModel
Activity RecognitionCholecT50 (Challenge)mAP38.1Team Trequartista
Activity RecognitionCholecT50 (Challenge)mAP36.9Team 2Ai
Activity RecognitionCholecT50 (Challenge)mAP35.8Team SIAT CAMI
Activity RecognitionCholecT50 (Challenge)mAP32.9Team HFUT-MedIA
Activity RecognitionCholecT50 (Challenge)mAP32.7Rendezvous (TensorFlow v1)
Activity RecognitionCholecT50 (Challenge)mAP32Team CITI SJTU
Activity RecognitionCholecT50 (Challenge)mAP31.9Team ANL
Activity RecognitionCholecT50 (Challenge)mAP31.7Team Digital Surgery
Activity RecognitionCholecT50 (Challenge)mAP26.7Team Casia Robotics
Activity RecognitionCholecT50 (Challenge)mAP26.3Team Lsgroup
Activity RecognitionCholecT50 (Challenge)mAP25.6Team J&M
Activity RecognitionCholecT50 (Challenge)mAP25.5Attention Tripnet (TensorFlow v1)
Activity RecognitionCholecT50 (Challenge)mAP25.2Team Ceaiik
Activity RecognitionCholecT50 (Challenge)mAP24.8Team SJTU-IMR
Activity RecognitionCholecT50 (Challenge)mAP18.4Team SK
Activity RecognitionCholecT50 (Challenge)mAP18.1Team MMLAB
Activity RecognitionCholecT50 (Challenge)mAP16Team Band of Broeders
Activity RecognitionCholecT50 (Challenge)mAP10.4Team NCT-TSO
Activity RecognitionCholecT50 (Challenge)mAP9.8Team HFUT-NUS
Activity RecognitionCholecT50 (Challenge)mAP9.3Team CAMP
Activity RecognitionCholecT50 (Challenge)mAP4.2Team Med Recognizer
Action RecognitionCholecT50 (Challenge)mAP38.1Team Trequartista
Action RecognitionCholecT50 (Challenge)mAP36.9Team 2Ai
Action RecognitionCholecT50 (Challenge)mAP35.8Team SIAT CAMI
Action RecognitionCholecT50 (Challenge)mAP32.9Team HFUT-MedIA
Action RecognitionCholecT50 (Challenge)mAP32.7Rendezvous (TensorFlow v1)
Action RecognitionCholecT50 (Challenge)mAP32Team CITI SJTU
Action RecognitionCholecT50 (Challenge)mAP31.9Team ANL
Action RecognitionCholecT50 (Challenge)mAP31.7Team Digital Surgery
Action RecognitionCholecT50 (Challenge)mAP26.7Team Casia Robotics
Action RecognitionCholecT50 (Challenge)mAP26.3Team Lsgroup
Action RecognitionCholecT50 (Challenge)mAP25.6Team J&M
Action RecognitionCholecT50 (Challenge)mAP25.5Attention Tripnet (TensorFlow v1)
Action RecognitionCholecT50 (Challenge)mAP25.2Team Ceaiik
Action RecognitionCholecT50 (Challenge)mAP24.8Team SJTU-IMR
Action RecognitionCholecT50 (Challenge)mAP18.4Team SK
Action RecognitionCholecT50 (Challenge)mAP18.1Team MMLAB
Action RecognitionCholecT50 (Challenge)mAP16Team Band of Broeders
Action RecognitionCholecT50 (Challenge)mAP10.4Team NCT-TSO
Action RecognitionCholecT50 (Challenge)mAP9.8Team HFUT-NUS
Action RecognitionCholecT50 (Challenge)mAP9.3Team CAMP
Action RecognitionCholecT50 (Challenge)mAP4.2Team Med Recognizer

Related Papers

ZKP-FedEval: Verifiable and Privacy-Preserving Federated Evaluation using Zero-Knowledge Proofs2025-07-15CBF-AFA: Chunk-Based Multi-SSL Fusion for Automatic Fluency Assessment2025-06-25MultiHuman-Testbench: Benchmarking Image Generation for Multiple Humans2025-06-25SEZ-HARN: Self-Explainable Zero-shot Human Activity Recognition Network2025-06-25Distributed Activity Detection for Cell-Free Hybrid Near-Far Field Communications2025-06-17Efficient Retail Video Annotation: A Robust Key Frame Generation Approach for Product and Customer Interaction Analysis2025-06-17DeSPITE: Exploring Contrastive Deep Skeleton-Pointcloud-IMU-Text Embeddings for Advanced Point Cloud Human Activity Understanding2025-06-16MORIC: CSI Delay-Doppler Decomposition for Robust Wi-Fi-based Human Activity Recognition2025-06-15