Back
Structured 3D Mesh Generation via Compositional Latent Diffusion Transformers
One‑shot generation of semantically decomposed 3D meshes from a single RGB image

PartCrafter: One‑Shot Structured 3D Mesh Generation from a Single Image
Imagine uploading a single RGB image and instantly receiving a full 3D model broken down into individual, semantically meaningful parts, ready for editing, printing, or animation. That’s exactly what PartCrafter accomplishes.
Paper & Repository
Research Paper
Title: “PartCrafter: Structured 3D Mesh Generation via Compositional Latent Diffusion Transformers”
Authors: Yuchen Lin, Chenguo Lin, Panwang Pan, Honglei Yan, Yiqiang Feng, Yadong Mu, Katerina Fragkiadaki
Published: June 5, 2025, on arXiv
Official GitHub Repository
Contains full code, with MIT license, and will release model weights + demos by mid-July
Project Page
Hosted by the authors, summarizing the method and linking to paper and code
Key Highlights
One‑Shot Multi-Part Generation: From a single RGB image, the model generates multiple 3D mesh parts simultaneously, no segmentation needed first.
Compositional Latent Space: Each part corresponds to a distinct set of latent tokens, plus a learned identity embedding.
Hierarchical Attention: A dual attention mechanism maintains detail within parts and coherence across the entire object.
Pretrained 3D Mesh Diffusion Transformer (DiT): Leverages pretrained DiT components for faster convergence and higher fidelity.
Performance & Dataset
Dataset: Curated ~130,000 objects with part annotations extracted from Objaverse, ShapeNet, ABO with about 300,000 parts.
Speed: Generates full part-aware mesh in ~30–34 seconds on a GPU, outperforming older two-stage pipelines (~18 minutes).
Quality: Achieves lower Chamfer distance and higher F-score (F ≈ 0.7472) than HoloPart and baseline DiT methods
Use Cases
3D Printing: Creates individual STL-ready parts for easy assembly (Tom’s Hardware feature).
CAD & Design: Enables part-level editing and modular asset creation.
AR/VR & Games: Facilitates generation of decomposable assets directly from reference images.
Robotics & Simulation: Provides structural understanding for object manipulation.
What's Available Now
Code & Checkpoints: Public GitHub repo available now. Model weights and demo expected before July 15, 2025.
Technical Demo: Live demonstrations and project page visuals show input-to-mesh workflow .
https://github.com/wgsxm/PartCrafter
https://arxiv.org/abs/2506.05573

MecAgent Inc.
Other articles you might like
Learn how to increase your productivity with calendars