CA‐PMG: Channel attention and progressive multi‐granularity training network for fine‐grained visual classification
Abstract Fine‐grained visual classification is challenging due to the inherently subtle intra‐class object variations. To solve this issue, a novel framework named channel attention and progressive multi‐granularity training network, is proposed. It first exploits meaningful feature maps through the...
Guardado en:
Autores principales: | , , , , , |
---|---|
Formato: | article |
Lenguaje: | EN |
Publicado: |
Wiley
2021
|
Materias: | |
Acceso en línea: | https://doaj.org/article/4cef360c02644bdb9315476c85c48414 |
Etiquetas: |
Agregar Etiqueta
Sin Etiquetas, Sea el primero en etiquetar este registro!
|
id |
oai:doaj.org-article:4cef360c02644bdb9315476c85c48414 |
---|---|
record_format |
dspace |
spelling |
oai:doaj.org-article:4cef360c02644bdb9315476c85c484142021-11-29T03:38:16ZCA‐PMG: Channel attention and progressive multi‐granularity training network for fine‐grained visual classification1751-96671751-965910.1049/ipr2.12238https://doaj.org/article/4cef360c02644bdb9315476c85c484142021-12-01T00:00:00Zhttps://doi.org/10.1049/ipr2.12238https://doaj.org/toc/1751-9659https://doaj.org/toc/1751-9667Abstract Fine‐grained visual classification is challenging due to the inherently subtle intra‐class object variations. To solve this issue, a novel framework named channel attention and progressive multi‐granularity training network, is proposed. It first exploits meaningful feature maps through the channel attention module and captures multi‐granularity features by the progressive multi‐granularity training module. For each feature map, the channel attention module is proposed to explore channel‐wise correlation. This allows the model to re‐weight the channels of the feature map according to the impact of their semantic information on performance. Furthermore, the progressive multi‐granularity training module is introduced to fuse features cross multi‐granularity. And the fused features pay more attention to the subtle differences between images. The model can be trained efficiently in an end‐to‐end manner without bounding box or part annotations. Finally, comprehensive experiments are conducted to show that the method achieves state‐of‐the‐art performances on the CUB‐200‐2011, Stanford Cars, and FGVC‐Aircraft datasets. Ablation studies demonstrate the effectiveness of each part in our module.Peipei ZhaoQiguang MiaoHang YaoXiangzeng LiuRuyi LiuMaoguo GongWileyarticlePhotographyTR1-1050Computer softwareQA76.75-76.765ENIET Image Processing, Vol 15, Iss 14, Pp 3718-3727 (2021) |
institution |
DOAJ |
collection |
DOAJ |
language |
EN |
topic |
Photography TR1-1050 Computer software QA76.75-76.765 |
spellingShingle |
Photography TR1-1050 Computer software QA76.75-76.765 Peipei Zhao Qiguang Miao Hang Yao Xiangzeng Liu Ruyi Liu Maoguo Gong CA‐PMG: Channel attention and progressive multi‐granularity training network for fine‐grained visual classification |
description |
Abstract Fine‐grained visual classification is challenging due to the inherently subtle intra‐class object variations. To solve this issue, a novel framework named channel attention and progressive multi‐granularity training network, is proposed. It first exploits meaningful feature maps through the channel attention module and captures multi‐granularity features by the progressive multi‐granularity training module. For each feature map, the channel attention module is proposed to explore channel‐wise correlation. This allows the model to re‐weight the channels of the feature map according to the impact of their semantic information on performance. Furthermore, the progressive multi‐granularity training module is introduced to fuse features cross multi‐granularity. And the fused features pay more attention to the subtle differences between images. The model can be trained efficiently in an end‐to‐end manner without bounding box or part annotations. Finally, comprehensive experiments are conducted to show that the method achieves state‐of‐the‐art performances on the CUB‐200‐2011, Stanford Cars, and FGVC‐Aircraft datasets. Ablation studies demonstrate the effectiveness of each part in our module. |
format |
article |
author |
Peipei Zhao Qiguang Miao Hang Yao Xiangzeng Liu Ruyi Liu Maoguo Gong |
author_facet |
Peipei Zhao Qiguang Miao Hang Yao Xiangzeng Liu Ruyi Liu Maoguo Gong |
author_sort |
Peipei Zhao |
title |
CA‐PMG: Channel attention and progressive multi‐granularity training network for fine‐grained visual classification |
title_short |
CA‐PMG: Channel attention and progressive multi‐granularity training network for fine‐grained visual classification |
title_full |
CA‐PMG: Channel attention and progressive multi‐granularity training network for fine‐grained visual classification |
title_fullStr |
CA‐PMG: Channel attention and progressive multi‐granularity training network for fine‐grained visual classification |
title_full_unstemmed |
CA‐PMG: Channel attention and progressive multi‐granularity training network for fine‐grained visual classification |
title_sort |
ca‐pmg: channel attention and progressive multi‐granularity training network for fine‐grained visual classification |
publisher |
Wiley |
publishDate |
2021 |
url |
https://doaj.org/article/4cef360c02644bdb9315476c85c48414 |
work_keys_str_mv |
AT peipeizhao capmgchannelattentionandprogressivemultigranularitytrainingnetworkforfinegrainedvisualclassification AT qiguangmiao capmgchannelattentionandprogressivemultigranularitytrainingnetworkforfinegrainedvisualclassification AT hangyao capmgchannelattentionandprogressivemultigranularitytrainingnetworkforfinegrainedvisualclassification AT xiangzengliu capmgchannelattentionandprogressivemultigranularitytrainingnetworkforfinegrainedvisualclassification AT ruyiliu capmgchannelattentionandprogressivemultigranularitytrainingnetworkforfinegrainedvisualclassification AT maoguogong capmgchannelattentionandprogressivemultigranularitytrainingnetworkforfinegrainedvisualclassification |
_version_ |
1718407641444646912 |