What is it about?

Camouflaged Object Detection (COD) aims to identify and segment objects that blend into their surroundings. Since the color and texture of the camouflaged objects are extremely similar to the surrounding environment, it is super challenging for vision models to precisely detect them. Inspired by research on biology and evolution, we introduce depth information as an additional cue to help break camouflage, which can provide spatial information and texture-free separation for foreground and background. To dig clues of camouflaged objects in both RGB and depth modalities, we innovatively propose Depth-aided Camouflaged Object Detection (DaCOD), which involves two key components. We firstly propose the Multi-modal Collaborative Learning (MCL) module, which aims to collaboratively learning deep features from both RGB and depth channels via a hybrid backbone. Then, we propose a novel Cross-modal Asymmetric Fusion (CAF) strategy, which asymmetrically fuse RGB and depth information for complementary depth feature enhancement to produce accurate predictions. We conducted numerous experiments of the proposed DaCOD on three widely-used challenging COD benchmark datasets, in which DaCOD outperforms the current state-of-the-arts by a large margin. All resources are available at https://github.com/qingwei-wang/DaCOD.

Featured Image

Why is it important?

We introduce depth cues for COD, validating that depth information can benefit the accurate localization and segmentation of camouflaged objects.

Perspectives

We hope this work will benefit researchers interested in multi-modal COD tasks. ^_^

Qingwei Wang

Read the Original

This page is a summary of: Depth-aided Camouflaged Object Detection, October 2023, ACM (Association for Computing Machinery),
DOI: 10.1145/3581783.3611874.
You can read the full text:

Read

Contributors

The following have contributed to this page