Structured 3D Mesh Generation
from Single Images

Transform any image into multiple editable 3D part meshes using advanced AI. PartCrafter revolutionizes 3D modeling with compositional latent diffusion transformers, enabling unprecedented control over generated meshes.

30 Seconds

Generation Time

4-16 Parts

Editable Meshes

0.7472

F-Score (SOTA)

Quick Start Guide

Get PartCrafter running in minutes with these simple steps

Clone Repository

1
git clone https://github.com/wgsxm/PartCrafter.git

Download the latest version from GitHub

Install Dependencies

2
pip install -r requirements.txt

Install Python packages and CUDA dependencies

Download Models

3
python download_models.py

Download pre-trained checkpoint files

Run Generation

4
python generate.py --input image.jpg

Generate 3D meshes from your image

See PartCrafter in Action

Watch our comprehensive demonstration showing how PartCrafter generates multiple editable 3D meshes from a single image

LIVE DEMO

Official demonstration video showing the complete generation process and results

Powering Innovation Across Industries

PartCrafter addresses critical challenges across multiple domains with AI-powered structured 3D generation, enabling professionals to achieve unprecedented productivity and creativity.

🎨

3D Artists & Animators

Transform concept art into editable 3D models with structured part separation for seamless animation workflows

Key Applications

  • Character rigging
  • Asset creation
  • Animation preparation
🖨️

3D Printing Enthusiasts

Convert photos into printable STL files with automatic part separation, perfect for complex assemblies

Key Applications

  • Rapid prototyping
  • Custom part creation
  • Repair modeling
🤖

Robotics Researchers

Generate structured object representations for advanced robotic manipulation and scene understanding

Key Applications

  • Grasp planning
  • Object manipulation
  • Scene understanding

Revolutionary 3D Generation Technology

PartCrafter combines cutting-edge AI techniques to achieve unprecedented 3D generation quality and speed

🧩

Compositional Latent Space

Disentangles every part into its own latent token set, enabling independent editing and replacement while maintaining object coherence.

🎯

Local-Global Hierarchical Attention

Local blocks keep intra-part detail preservation while global blocks enforce coherence across parts, preventing intersections or gaps.

One-Step Multi-Mesh Output

Generates 4-16 separate meshes directly from a single RGB image, unlike traditional two-stage 'segment-then-reconstruct' pipelines.

🚀

Warm-Start from TripoSG-DiT

Leverages pre-trained weights for higher fidelity generation with reduced computational requirements and faster training convergence.

🔺

Explicit Triangle Meshes

Direct output of CAD-ready triangle meshes suitable for animation, physics simulation, and 3D printing without post-processing.

📊

Part-Annotated Training Data

Trained on ~130k meshes from Objaverse, ShapeNet, and ABO with preserved part hierarchy metadata for structured learning.

Technical Specifications

Comprehensive technical details about PartCrafter's architecture, performance metrics, and system requirements

Model Architecture

Base ModelCompositional Latent Diffusion Transformer
Attention MechanismLocal-Global Hierarchical
Warm-start FromTripoSG-DiT
Output FormatExplicit Triangle Meshes

Performance Metrics

Generation Time~30 seconds
Number of Parts4-16 meshes per object
Chamfer Distance0.1726 (SOTA)
F-Score0.7472 (SOTA)

Training Dataset

Dataset Size~130k part-annotated meshes
SourcesObjaverse + ShapeNet + ABO
Annotation TypePart hierarchy metadata
Input Resolution512×512 RGB images

System Requirements

GPU Memory≥16GB VRAM recommended
CUDA SupportCUDA 11.8+ required
Python Version3.8+ supported
Storage~10GB for models and dependencies

Performance Benchmarks

PartCrafter achieves state-of-the-art results in structured 3D generation while being significantly faster than competing methods

ModelChamfer Distance ↓F-Score ↑Generation TimeRequires Segmentation
PartCrafter0.17260.747234 secondsNo
HoloPart (2025)0.19160.691618 minutesYes
TripoSG (backbone)0.18210.711530 secondsNo

Benchmarks performed on H20 GPU (40GB). Numbers from official paper and industry evaluations.

Research Paper

arXiv:2506.05573Published June 2025

PartCrafter: Structured 3D Mesh Generation via Compositional Latent Diffusion Transformers

We present PartCrafter, a novel approach for generating structured 3D meshes from single RGB images. Unlike existing methods that treat 3D objects as monolithic entities, PartCrafter decomposes objects into semantically meaningful parts, enabling fine-grained editing and manipulation.

BibTeX Citation

@article{partcrafter2025,
  title={PartCrafter: Structured 3D Mesh Generation via 
         Compositional Latent Diffusion Transformers},
  author={Wang, Guangshun and Liu, Xiao and Chen, Wei},
  journal={arXiv preprint arXiv:2506.05573},
  year={2025}
}