@Article{cmes.2022.021351, AUTHOR = {Abubakar Sulaiman Gezawa, Qicong Wang,2, Haruna Chiroma, Yunqi Lei}, TITLE = {A Deep Learning Approach to Mesh Segmentation}, JOURNAL = {Computer Modeling in Engineering \& Sciences}, VOLUME = {135}, YEAR = {2023}, NUMBER = {2}, PAGES = {1745--1763}, URL = {http://www.techscience.com/CMES/v135n2/50180}, ISSN = {1526-1506}, ABSTRACT = {In the shape analysis community, decomposing a 3D shape into meaningful parts has become a topic of interest. 3D model segmentation is largely used in tasks such as shape deformation, shape partial matching, skeleton extraction, shape correspondence, shape annotation and texture mapping. Numerous approaches have attempted to provide better segmentation solutions; however, the majority of the previous techniques used handcrafted features, which are usually focused on a particular attribute of 3D objects and so are difficult to generalize. In this paper, we propose a three-stage approach for using Multi-view recurrent neural network to automatically segment a 3D shape into visually meaningful sub-meshes. The first stage involves normalizing and scaling a 3D model to fit within the unit sphere and rendering the object into different views. Contrasting viewpoints, on the other hand, might not have been associated, and a 3D region could correlate into totally distinct outcomes depending on the viewpoint. To address this, we ran each view through (shared weights) CNN and Bolster block in order to create a probability boundary map. The Bolster block simulates the area relationships between different views, which helps to improve and refine the data. In stage two, the feature maps generated in the previous step are correlated using a Recurrent Neural network to obtain compatible fine detail responses for each view. Finally, a layer that is fully connected is used to return coherent edges, which are then back project to 3D objects to produce the final segmentation. Experiments on the Princeton Segmentation Benchmark dataset show that our proposed method is effective for mesh segmentation tasks.}, DOI = {10.32604/cmes.2022.021351} }