TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Subdivision-Based Mesh Convolution Networks

Subdivision-Based Mesh Convolution Networks

Shi-Min Hu, Zheng-Ning Liu, Meng-Hao Guo, Jun-Xiong Cai, Jiahui Huang, Tai-Jiang Mu, Ralph R. Martin

2021-06-043D ClassificationPose Estimation
PaperPDFCode(official)

Abstract

Convolutional neural networks (CNNs) have made great breakthroughs in 2D computer vision. However, their irregular structure makes it hard to harness the potential of CNNs directly on meshes. A subdivision surface provides a hierarchical multi-resolution structure, in which each face in a closed 2-manifold triangle mesh is exactly adjacent to three faces. Motivated by these two observations, this paper presents SubdivNet, an innovative and versatile CNN framework for 3D triangle meshes with Loop subdivision sequence connectivity. Making an analogy between mesh faces and pixels in a 2D image allows us to present a mesh convolution operator to aggregate local features from nearby faces. By exploiting face neighborhoods, this convolution can support standard 2D convolutional network concepts, e.g. variable kernel size, stride, and dilation. Based on the multi-resolution hierarchy, we make use of pooling layers which uniformly merge four faces into one and an upsampling method which splits one face into four. Thereby, many popular 2D CNN architectures can be easily adapted to process 3D meshes. Meshes with arbitrary connectivity can be remeshed to have Loop subdivision sequence connectivity via self-parameterization, making SubdivNet a general approach. Extensive evaluation and various applications demonstrate SubdivNet's effectiveness and efficiency.

Results

TaskDatasetMetricValueModel
Pose EstimationSALSAAccuracy93SubdivNet
Pose EstimationSALSAAccuracy87.7MeshCNN (Hanocka et al., 2019)
Pose EstimationSALSAAccuracy82.3Pointnet++ (Qi et al., [2017b])
Pose EstimationSALSAAccuracy74.7Pointnet (Qi et al., [2017a])
3DSALSAAccuracy93SubdivNet
3DSALSAAccuracy87.7MeshCNN (Hanocka et al., 2019)
3DSALSAAccuracy82.3Pointnet++ (Qi et al., [2017b])
3DSALSAAccuracy74.7Pointnet (Qi et al., [2017a])
1 Image, 2*2 StitchiSALSAAccuracy93SubdivNet
1 Image, 2*2 StitchiSALSAAccuracy87.7MeshCNN (Hanocka et al., 2019)
1 Image, 2*2 StitchiSALSAAccuracy82.3Pointnet++ (Qi et al., [2017b])
1 Image, 2*2 StitchiSALSAAccuracy74.7Pointnet (Qi et al., [2017a])

Related Papers

$π^3$: Scalable Permutation-Equivariant Visual Geometry Learning2025-07-17Revisiting Reliability in the Reasoning-based Pose Estimation Benchmark2025-07-17DINO-VO: A Feature-based Visual Odometry Leveraging a Visual Foundation Model2025-07-17From Neck to Head: Bio-Impedance Sensing for Head Pose Estimation2025-07-17AthleticsPose: Authentic Sports Motion Dataset on Athletic Field and Evaluation of Monocular 3D Pose Estimation Ability2025-07-17SpatialTrackerV2: 3D Point Tracking Made Easy2025-07-16SGLoc: Semantic Localization System for Camera Pose Estimation from 3D Gaussian Splatting Representation2025-07-16Efficient Calisthenics Skills Classification through Foreground Instance Selection and Depth Estimation2025-07-16