TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/A Diffusion Model and Knowledge Distillation Framework for...

A Diffusion Model and Knowledge Distillation Framework for Robust Coral Detection in Complex Underwater Environments

Zhaoxuan Lu; Lyuchao Liao; Chuang Li;Xingang Xie; Hui Yuan;

2025-01-06SSRN 2025 1Transfer Learning2D Object DetectionKnowledge Distillation
PaperPDFCode

Abstract

Coral reefs play a crucial role in marine ecosystems, but their sustainability is increasingly threatened by climate change and human activities. To aid in the protection and monitoring of these ecosystems, developing advanced artificial intelligence (AI)-based automated detection technologies is essential. This paper introduces the MambaCoral-Diffusion Detection framework (MambaCoral-Diffusion Detection, MCDD), an AI-driven approach for robust coral detection, designed to enhance performance in complex underwater environments—a critical challenge in marine engineering. Key AI contributions include integrating a diffusion model to generate diverse training data to address data imbalance and employing an adaptive global-local attention mechanism within the Mamba architecture’s Spatial Sensing Dual Detection (SS2D) module to enhance feature extraction accuracy. Furthermore, a Channel-Aware Intersection over Union (IoU) knowledge distillation technique is presented, enabling efficient knowledge transfer from complex teacher models to lightweight student models, optimizing the framework for practical deployment with reduced computational demands. From an engineering perspective, MCDD significantly advances automated coral detection in challenging underwater conditions, providing a reliable solution for monitoring marine ecosystems. Experimental results show that MCDD achieves high performance on the Soft Coral dataset, reaching 31.5 frames per second (FPS), the mean average precision at 50\% IoU threshold of 0.843, and the mean average precision averaged across IoU thresholds from 50% to 95% of 0.566, with only 6.5 million parameters and 13.6 billion floating point operations per second (GFLOPs). The code and dataset are available at https://github.com/RDXiaoLu/MambaCoral-DiffDet.git

Related Papers

Visual-Language Model Knowledge Distillation Method for Image Quality Assessment2025-07-21RaMen: Multi-Strategy Multi-Modal Learning for Bundle Construction2025-07-18Disentangling coincident cell events using deep transfer learning and compressive sensing2025-07-17Uncertainty-Aware Cross-Modal Knowledge Distillation with Prototype Learning for Multimodal Brain-Computer Interfaces2025-07-17Best Practices for Large-Scale, Pixel-Wise Crop Mapping and Transfer Learning Workflows2025-07-16DVFL-Net: A Lightweight Distilled Video Focal Modulation Network for Spatio-Temporal Action Recognition2025-07-16Robust-Multi-Task Gradient Boosting2025-07-15HanjaBridge: Resolving Semantic Ambiguity in Korean LLMs via Hanja-Augmented Pre-Training2025-07-15