Joint Soft–Hard Attention for Self-Supervised Monocular Depth Estimation
In recent years, self-supervised monocular depth estimation has gained popularity among researchers because it uses only a single camera at a much lower cost than the direct use of laser sensors to acquire depth. Although monocular self-supervised methods can obtain dense depths, the estimation accu...
Guardado en:
Autores principales: | , , , , |
---|---|
Formato: | article |
Lenguaje: | EN |
Publicado: |
MDPI AG
2021
|
Materias: | |
Acceso en línea: | https://doaj.org/article/004f2d06aaea4edcb6ef81127fdda6f3 |
Etiquetas: |
Agregar Etiqueta
Sin Etiquetas, Sea el primero en etiquetar este registro!
|
id |
oai:doaj.org-article:004f2d06aaea4edcb6ef81127fdda6f3 |
---|---|
record_format |
dspace |
spelling |
oai:doaj.org-article:004f2d06aaea4edcb6ef81127fdda6f32021-11-11T19:00:15ZJoint Soft–Hard Attention for Self-Supervised Monocular Depth Estimation10.3390/s212169561424-8220https://doaj.org/article/004f2d06aaea4edcb6ef81127fdda6f32021-10-01T00:00:00Zhttps://www.mdpi.com/1424-8220/21/21/6956https://doaj.org/toc/1424-8220In recent years, self-supervised monocular depth estimation has gained popularity among researchers because it uses only a single camera at a much lower cost than the direct use of laser sensors to acquire depth. Although monocular self-supervised methods can obtain dense depths, the estimation accuracy needs to be further improved for better applications in scenarios such as autonomous driving and robot perception. In this paper, we innovatively combine soft attention and hard attention with two new ideas to improve self-supervised monocular depth estimation: (1) a soft attention module and (2) a hard attention strategy. We integrate the soft attention module in the model architecture to enhance feature extraction in both spatial and channel dimensions, adding only a small number of parameters. Unlike traditional fusion approaches, we use the hard attention strategy to enhance the fusion of generated multi-scale depth predictions. Further experiments demonstrate that our method can achieve the best self-supervised performance both on the standard KITTI benchmark and the Make3D dataset.Chao FanZhenyu YinFulong XuAnying ChaiFeiqing ZhangMDPI AGarticlemonocular depth estimationself-supervised learningattention mechanismChemical technologyTP1-1185ENSensors, Vol 21, Iss 6956, p 6956 (2021) |
institution |
DOAJ |
collection |
DOAJ |
language |
EN |
topic |
monocular depth estimation self-supervised learning attention mechanism Chemical technology TP1-1185 |
spellingShingle |
monocular depth estimation self-supervised learning attention mechanism Chemical technology TP1-1185 Chao Fan Zhenyu Yin Fulong Xu Anying Chai Feiqing Zhang Joint Soft–Hard Attention for Self-Supervised Monocular Depth Estimation |
description |
In recent years, self-supervised monocular depth estimation has gained popularity among researchers because it uses only a single camera at a much lower cost than the direct use of laser sensors to acquire depth. Although monocular self-supervised methods can obtain dense depths, the estimation accuracy needs to be further improved for better applications in scenarios such as autonomous driving and robot perception. In this paper, we innovatively combine soft attention and hard attention with two new ideas to improve self-supervised monocular depth estimation: (1) a soft attention module and (2) a hard attention strategy. We integrate the soft attention module in the model architecture to enhance feature extraction in both spatial and channel dimensions, adding only a small number of parameters. Unlike traditional fusion approaches, we use the hard attention strategy to enhance the fusion of generated multi-scale depth predictions. Further experiments demonstrate that our method can achieve the best self-supervised performance both on the standard KITTI benchmark and the Make3D dataset. |
format |
article |
author |
Chao Fan Zhenyu Yin Fulong Xu Anying Chai Feiqing Zhang |
author_facet |
Chao Fan Zhenyu Yin Fulong Xu Anying Chai Feiqing Zhang |
author_sort |
Chao Fan |
title |
Joint Soft–Hard Attention for Self-Supervised Monocular Depth Estimation |
title_short |
Joint Soft–Hard Attention for Self-Supervised Monocular Depth Estimation |
title_full |
Joint Soft–Hard Attention for Self-Supervised Monocular Depth Estimation |
title_fullStr |
Joint Soft–Hard Attention for Self-Supervised Monocular Depth Estimation |
title_full_unstemmed |
Joint Soft–Hard Attention for Self-Supervised Monocular Depth Estimation |
title_sort |
joint soft–hard attention for self-supervised monocular depth estimation |
publisher |
MDPI AG |
publishDate |
2021 |
url |
https://doaj.org/article/004f2d06aaea4edcb6ef81127fdda6f3 |
work_keys_str_mv |
AT chaofan jointsofthardattentionforselfsupervisedmonoculardepthestimation AT zhenyuyin jointsofthardattentionforselfsupervisedmonoculardepthestimation AT fulongxu jointsofthardattentionforselfsupervisedmonoculardepthestimation AT anyingchai jointsofthardattentionforselfsupervisedmonoculardepthestimation AT feiqingzhang jointsofthardattentionforselfsupervisedmonoculardepthestimation |
_version_ |
1718431624921612288 |