The generation of high-quality 3D models from single 2D images remains challenging in terms of accuracy and completeness.Deep learning has emerged as a promising solution,offering new avenues for improvements.However,...The generation of high-quality 3D models from single 2D images remains challenging in terms of accuracy and completeness.Deep learning has emerged as a promising solution,offering new avenues for improvements.However,building models from scratch is computationally expensive and requires large datasets.This paper presents a transfer-learning-based approach for category-specific 3D reconstruction from a single 2D image.The core idea is to fine-tune a pre-trained model on specific object categories using new,unseen data,resulting in specialized versions of the model that are better adapted to reconstruct particular objects.The proposed approach utilizes a three-phase pipeline comprising image acquisition,3D reconstruction,and refinement.After ensuring the quality of the input image,a ResNet50 model is used for object recognition,directing the image to the corresponding category-specific model to generate a voxel-based representation.The voxel-based 3D model is then refined by transforming it into a detailed triangular mesh representation using the Marching Cubes algorithm and Laplacian smoothing.An experimental study,using the Pix2Vox model and the Pascal3D dataset,has been conducted to evaluate and validate the effectiveness of the proposed approach.Results demonstrate that category-specific fine-tuning of Pix2Vox significantly outperforms both the original model and the general model fine-tuned for all object categories,with substantial gains in Intersection over Union(IoU)scores.Visual assessments confirm improvements in geometric detail and surface realism.These findings indicate that combining transfer learning with category-specific fine tuning and refinement strategy of our approach leads to better-quality 3D model generation.展开更多
基金funded by the Research,Development,and Innovation Authority(RDIA)—Kingdom of Saudi Arabia—under supervision Energy,Industry,and Advanced Technologies Research Center,Taibah University,Madinah,Saudi Arabia with grant number(12979-iau-2023-TAU-R-3-1-EI-).
文摘The generation of high-quality 3D models from single 2D images remains challenging in terms of accuracy and completeness.Deep learning has emerged as a promising solution,offering new avenues for improvements.However,building models from scratch is computationally expensive and requires large datasets.This paper presents a transfer-learning-based approach for category-specific 3D reconstruction from a single 2D image.The core idea is to fine-tune a pre-trained model on specific object categories using new,unseen data,resulting in specialized versions of the model that are better adapted to reconstruct particular objects.The proposed approach utilizes a three-phase pipeline comprising image acquisition,3D reconstruction,and refinement.After ensuring the quality of the input image,a ResNet50 model is used for object recognition,directing the image to the corresponding category-specific model to generate a voxel-based representation.The voxel-based 3D model is then refined by transforming it into a detailed triangular mesh representation using the Marching Cubes algorithm and Laplacian smoothing.An experimental study,using the Pix2Vox model and the Pascal3D dataset,has been conducted to evaluate and validate the effectiveness of the proposed approach.Results demonstrate that category-specific fine-tuning of Pix2Vox significantly outperforms both the original model and the general model fine-tuned for all object categories,with substantial gains in Intersection over Union(IoU)scores.Visual assessments confirm improvements in geometric detail and surface realism.These findings indicate that combining transfer learning with category-specific fine tuning and refinement strategy of our approach leads to better-quality 3D model generation.