PointMTL: Multi-Transform Learning for Effective 3D Point Cloud Representations
Effectively learning and extracting the feature representations of 3D point clouds is an important yet challenging task. Most of existing works achieve reasonable performance in 3D vision tasks by modeling the relationships among points appropriately. However, the feature representations are only le...
Guardado en:
Autores principales: | , , , , , , , |
---|---|
Formato: | article |
Lenguaje: | EN |
Publicado: |
IEEE
2021
|
Materias: | |
Acceso en línea: | https://doaj.org/article/9e3f67a428074777ae56849b7057d65f |
Etiquetas: |
Agregar Etiqueta
Sin Etiquetas, Sea el primero en etiquetar este registro!
|
id |
oai:doaj.org-article:9e3f67a428074777ae56849b7057d65f |
---|---|
record_format |
dspace |
spelling |
oai:doaj.org-article:9e3f67a428074777ae56849b7057d65f2021-11-19T00:06:53ZPointMTL: Multi-Transform Learning for Effective 3D Point Cloud Representations2169-353610.1109/ACCESS.2021.3094624https://doaj.org/article/9e3f67a428074777ae56849b7057d65f2021-01-01T00:00:00Zhttps://ieeexplore.ieee.org/document/9474491/https://doaj.org/toc/2169-3536Effectively learning and extracting the feature representations of 3D point clouds is an important yet challenging task. Most of existing works achieve reasonable performance in 3D vision tasks by modeling the relationships among points appropriately. However, the feature representations are only learned with a specific transform through these methods, which are easy to overlap and thus limit the representation ability of the model. To address these issues, we propose a novel Multi-Transform Learning framework for point clouds (PointMTL), which can extract diverse features from multiple mapping transform to obtain richer representations. Specifically, we build a module named Multi-Transform Encoder (MTE), which encodes and aggregates local features from multiple non-linear transforms. To further explore global context representations, a module named Global Spatial Fusion (GSF) is proposed to capture global information and selectively fuse with local representations. Moreover, to guarantee the richness and diversity of learned representations, we further propose a Spatial Independence Criterion (SIC) strategy to enlarge the differences between the transforms and reduce information redundancies. In contrast to previous works, our approach fully exploits representations from multiple transforms, thus having strong expressiveness and good robustness for point clouds related tasks. The experiments on three typical tasks (i.e., semantic segmentation on S3DIS and ScanNet, part segmentation on ShapeNet and shape classification on ModelNet40) demonstrates the effectiveness of our method.Yifan JianYuwei YangZhi ChenXianguo QingYang ZhaoLiang HeXuekun ChenWei LuoIEEEarticle3D point cloudsfeature representationsmulti-transform learningsemantic segmentationElectrical engineering. Electronics. Nuclear engineeringTK1-9971ENIEEE Access, Vol 9, Pp 126241-126255 (2021) |
institution |
DOAJ |
collection |
DOAJ |
language |
EN |
topic |
3D point clouds feature representations multi-transform learning semantic segmentation Electrical engineering. Electronics. Nuclear engineering TK1-9971 |
spellingShingle |
3D point clouds feature representations multi-transform learning semantic segmentation Electrical engineering. Electronics. Nuclear engineering TK1-9971 Yifan Jian Yuwei Yang Zhi Chen Xianguo Qing Yang Zhao Liang He Xuekun Chen Wei Luo PointMTL: Multi-Transform Learning for Effective 3D Point Cloud Representations |
description |
Effectively learning and extracting the feature representations of 3D point clouds is an important yet challenging task. Most of existing works achieve reasonable performance in 3D vision tasks by modeling the relationships among points appropriately. However, the feature representations are only learned with a specific transform through these methods, which are easy to overlap and thus limit the representation ability of the model. To address these issues, we propose a novel Multi-Transform Learning framework for point clouds (PointMTL), which can extract diverse features from multiple mapping transform to obtain richer representations. Specifically, we build a module named Multi-Transform Encoder (MTE), which encodes and aggregates local features from multiple non-linear transforms. To further explore global context representations, a module named Global Spatial Fusion (GSF) is proposed to capture global information and selectively fuse with local representations. Moreover, to guarantee the richness and diversity of learned representations, we further propose a Spatial Independence Criterion (SIC) strategy to enlarge the differences between the transforms and reduce information redundancies. In contrast to previous works, our approach fully exploits representations from multiple transforms, thus having strong expressiveness and good robustness for point clouds related tasks. The experiments on three typical tasks (i.e., semantic segmentation on S3DIS and ScanNet, part segmentation on ShapeNet and shape classification on ModelNet40) demonstrates the effectiveness of our method. |
format |
article |
author |
Yifan Jian Yuwei Yang Zhi Chen Xianguo Qing Yang Zhao Liang He Xuekun Chen Wei Luo |
author_facet |
Yifan Jian Yuwei Yang Zhi Chen Xianguo Qing Yang Zhao Liang He Xuekun Chen Wei Luo |
author_sort |
Yifan Jian |
title |
PointMTL: Multi-Transform Learning for Effective 3D Point Cloud Representations |
title_short |
PointMTL: Multi-Transform Learning for Effective 3D Point Cloud Representations |
title_full |
PointMTL: Multi-Transform Learning for Effective 3D Point Cloud Representations |
title_fullStr |
PointMTL: Multi-Transform Learning for Effective 3D Point Cloud Representations |
title_full_unstemmed |
PointMTL: Multi-Transform Learning for Effective 3D Point Cloud Representations |
title_sort |
pointmtl: multi-transform learning for effective 3d point cloud representations |
publisher |
IEEE |
publishDate |
2021 |
url |
https://doaj.org/article/9e3f67a428074777ae56849b7057d65f |
work_keys_str_mv |
AT yifanjian pointmtlmultitransformlearningforeffective3dpointcloudrepresentations AT yuweiyang pointmtlmultitransformlearningforeffective3dpointcloudrepresentations AT zhichen pointmtlmultitransformlearningforeffective3dpointcloudrepresentations AT xianguoqing pointmtlmultitransformlearningforeffective3dpointcloudrepresentations AT yangzhao pointmtlmultitransformlearningforeffective3dpointcloudrepresentations AT lianghe pointmtlmultitransformlearningforeffective3dpointcloudrepresentations AT xuekunchen pointmtlmultitransformlearningforeffective3dpointcloudrepresentations AT weiluo pointmtlmultitransformlearningforeffective3dpointcloudrepresentations |
_version_ |
1718420626197184512 |