6533b7d4fe1ef96bd1262977

RESEARCH PRODUCT

Incorporating depth information into few-shot semantic segmentation

Désiré SidibéFabrice MeriaudeauOlivier MorelYifei Zhang

subject

[INFO.INFO-AI] Computer Science [cs]/Artificial Intelligence [cs.AI]Artificial neural networkComputer sciencebusiness.industry[INFO.INFO-TS] Computer Science [cs]/Signal and Image ProcessingComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION[INFO.INFO-CV]Computer Science [cs]/Computer Vision and Pattern Recognition [cs.CV]020206 networking & telecommunications02 engineering and technologyImage segmentationSemanticsVisualization[INFO.INFO-AI]Computer Science [cs]/Artificial Intelligence [cs.AI][INFO.INFO-CV] Computer Science [cs]/Computer Vision and Pattern Recognition [cs.CV][INFO.INFO-TS]Computer Science [cs]/Signal and Image ProcessingMetric (mathematics)0202 electrical engineering electronic engineering information engineeringEmbeddingRGB color modelSegmentationComputer visionArtificial intelligencebusiness

description

International audience; Few-shot segmentation presents a significant challengefor semantic scene understanding under limited supervision.Namely, this task targets at generalizing the segmentationability of the model to new categories given a few samples.In order to obtain complete scene information, we extend theRGB-centric methods to take advantage of complementary depthinformation. In this paper, we propose a two-stream deep neuralnetwork based on metric learning. Our method, known as RDNet,learns class-specific prototype representations within RGB anddepth embedding spaces, respectively. The learned prototypesprovide effective semantic guidance on the corresponding RGBand depth query image, leading to more accurate performance.Moreover, we build a novel outdoor scene dataset, known asCityscapes-3i, using labeled RGB images and depth imagesfrom the Cityscapes dataset. We also perform ablation studiesto explore the effective use of depth information in few-shotsegmentation tasks. Experiments on Cityscapes-3i show that ourmethod achieves excellent results with visual and complementarygeometric cues from only a few labeled examples.

https://hal-univ-evry.archives-ouvertes.fr/hal-02887063/file/ICPR_2020_YZ_DS_OM_FM.pdf