Point-E
Point cloud diffusion for 3D model synthesis
...Its principal advantage is speed: it can generate 3D assets in just 1–2 minutes on a single GPU, which is significantly faster than many competing text-to-3D models. The model works via a two-stage diffusion approach: first, it uses a text → image diffusion network to produce a synthetic 2D view consistent with the prompt; then a second diffusion model converts that image into a 3D point cloud. While it does not match the fine detail of some slower methods, the tradeoff in speed makes it practical for prototyping and interactive 3D generation. The repository includes inference scripts, utilities for converting point clouds to meshes (e.g. via signed distance function regression), sample notebooks, and weight checkpoints. ...