Shoukat, M. A., Sargano, A. B., Malyshev, A., You, L. and Habib, Z., 2024. SS3DNet-AF: A Single-Stage, Single-View 3D Reconstruction Network with Attention-Based Fusion. Applied Sciences, 14 (23), 11424.
Full text available as:
|
PDF (OPEN ACCESS ARTICLE)
applsci-14-11424-v2.pdf - Published Version Available under License Creative Commons Attribution. 3MB | |
Copyright to original material in this document is with the original owner(s). Access to this content through BURO is granted on condition that you use it only for research, scholarly or other non-commercial purposes. If you wish to use it for any other purposes, you must contact BU via BURO@bournemouth.ac.uk. Any third party copyright material in this document remains the property of its respective owner(s). BU grants no licence for further use of that third party material. |
DOI: 10.3390/app142311424
Abstract
Learning object shapes from a single image is challenging due to variations in scene content, geometric structures, and environmental factors, which create significant disparities between 2D image features and their corresponding 3D representations, hindering the effective training of deep learning models. Existing learning-based approaches can be divided into two-stage and single-stage methods, each with limitations. Two-stage methods often rely on generating intermediate proposals by searching for similar structures across the entire dataset, a process that is computationally expensive due to the large search space and high-dimensional feature-matching requirements, further limiting flexibility to predefined object categories. In contrast, single-stage methods directly reconstruct 3D shapes from images without intermediate steps, but they struggle to capture complex object geometries due to high feature loss between image features and 3D shapes and limit their ability to represent intricate details. To address these challenges, this paper introduces SS3DNet-AF, a single-stage, single-view 3D reconstruction network with an attention-based fusion (AF) mechanism to enhance focus on relevant image features, effectively capturing geometric details and generalizing across diverse object categories. The proposed method is quantitatively evaluated using the ShapeNet dataset, demonstrating its effectiveness in achieving accurate 3D reconstructions while overcoming the computational challenges associated with traditional approaches.
Item Type: | Article |
---|---|
ISSN: | 2076-3417 |
Uncontrolled Keywords: | SS3DNet-AF; 3D reconstruction; attention-based fusion; point clouds |
Group: | UNSPECIFIED |
ID Code: | 40668 |
Deposited By: | Symplectic RT2 |
Deposited On: | 08 Jan 2025 11:01 |
Last Modified: | 08 Jan 2025 11:01 |
Downloads
Downloads per month over past year
Repository Staff Only - |