Semantic Segmentation With Multi Scale Spatial Attention For Self Driving Cars

In this paper, we present a novel neural network using multi scale feature fusion at various scales for accurate and efficient semantic image segmentation. We used ResNet based feature extractor, dilated convolutional layers in down-sampling part, atrous convolutional layers in the upsampling part a...

Full description

Saved in:
Bibliographic Details
Published in:2021 IEEE/CVF International Conference on Computer Vision Workshops (ICCVW) pp. 2650 - 2656
Main Authors: Sagar, Abhinav, Soundrapandiyan, RajKumar
Format: Conference Proceeding
Language:English
Published: IEEE 01-10-2021
Subjects:
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:In this paper, we present a novel neural network using multi scale feature fusion at various scales for accurate and efficient semantic image segmentation. We used ResNet based feature extractor, dilated convolutional layers in down-sampling part, atrous convolutional layers in the upsampling part and used concat operation to merge them. A new attention module is proposed to encode more contextual information and enhance the receptive field of the network. We present an in depth theoretical analysis of our network with training and optimization details. Our network was trained and tested on the Camvid dataset and Cityscapes dataset using mean accuracy per class and Intersection Over Union (IOU) as the evaluation metrics. Our model outperforms previous state of the art methods on semantic segmentation achieving mean IOU value of 74.12 while running at >100 FPS.
ISSN:2473-9944
DOI:10.1109/ICCVW54120.2021.00299