6533b7d4fe1ef96bd1262977
RESEARCH PRODUCT
Incorporating depth information into few-shot semantic segmentation
Désiré SidibéFabrice MeriaudeauOlivier MorelYifei Zhangsubject
[INFO.INFO-AI] Computer Science [cs]/Artificial Intelligence [cs.AI]Artificial neural networkComputer sciencebusiness.industry[INFO.INFO-TS] Computer Science [cs]/Signal and Image ProcessingComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION[INFO.INFO-CV]Computer Science [cs]/Computer Vision and Pattern Recognition [cs.CV]020206 networking & telecommunications02 engineering and technologyImage segmentationSemanticsVisualization[INFO.INFO-AI]Computer Science [cs]/Artificial Intelligence [cs.AI][INFO.INFO-CV] Computer Science [cs]/Computer Vision and Pattern Recognition [cs.CV][INFO.INFO-TS]Computer Science [cs]/Signal and Image ProcessingMetric (mathematics)0202 electrical engineering electronic engineering information engineeringEmbeddingRGB color modelSegmentationComputer visionArtificial intelligencebusinessdescription
International audience; Few-shot segmentation presents a significant challengefor semantic scene understanding under limited supervision.Namely, this task targets at generalizing the segmentationability of the model to new categories given a few samples.In order to obtain complete scene information, we extend theRGB-centric methods to take advantage of complementary depthinformation. In this paper, we propose a two-stream deep neuralnetwork based on metric learning. Our method, known as RDNet,learns class-specific prototype representations within RGB anddepth embedding spaces, respectively. The learned prototypesprovide effective semantic guidance on the corresponding RGBand depth query image, leading to more accurate performance.Moreover, we build a novel outdoor scene dataset, known asCityscapes-3i, using labeled RGB images and depth imagesfrom the Cityscapes dataset. We also perform ablation studiesto explore the effective use of depth information in few-shotsegmentation tasks. Experiments on Cityscapes-3i show that ourmethod achieves excellent results with visual and complementarygeometric cues from only a few labeled examples.
year | journal | country | edition | language |
---|---|---|---|---|
2021-01-10 |