Image-to-3D
Free image-to-3D AI tools for transforming images into 3D assets for games, films, and design projects, optimizing your creative process.
DimensionX can generate photorealistic 3D and 4D scenes from a single image using controllable video diffusion.
Feels like we get an image-to-3D method each week now. CRM is yet another one that can generate 3D objects from a single image. This one is able to create high-fidelity textured meshes with interactable surfaces in just 10 seconds. Results are stunning!
3D-Fauna is able to turn a single image of a quadruped animal into an articulated, textured 3D mesh in a feed-forward manner, ready for animation and rendering.
WiLoR can localize and reconstruct multiple hands in real-time from single images. It achieves smooth 3D hand tracking with high accuracy, using a large dataset of over 2 million hand images.
3DTopia-XL can generate high-quality 3D PBR assets from text or image inputs in just 5 seconds.
DrawingSpinUp can animate 3D characters from a single 2D drawing. It removes unnecessary lines and uses a skeleton-based algorithm to allow characters to spin, jump, and dance.
SparseCraft can reconstruct 3D shapes and appearances from just three colored images. It uses a Signed Distance Function (SDF) and a radiance field, achieving fast training times of under 10 minutes without needing pretrained models.
DreamCar can reconstruct 3D car models from just a few images or single-image inputs. It uses Score Distillation Sampling and pose optimization to enhance texture alignment and overall model quality, significantly outperforming existing methods.
An Object is Worth 64x64 Pixels can generate 3D models from 64x64 pixel images! It creates realistic objects with good shapes and colors, working as well as more complex methods.
Tailor3D can create customized 3D assets from text or single and dual-side images. The method also supports adding changes to the inputs through additional text prompts.
Portrait3D can generate high-quality 3D heads with accurate geometry and texture from a single in-the-wild portrait image.
Sketch2Scene can create interactive 3D game scenes from simple sketches and text descriptions. It uses a diffusion model with ControlNet and procedural generation to make high-quality, playable 3D environments that match what users want.
Toon3D can generate 3D scenes from two or more cartoon drawings. It’s far from perfect, but still pretty cool!
InstantMesh can generate high-quality 3D meshes from a single image in under 10 seconds. It uses advanced methods like multiview diffusion and sparse-view reconstruction, and it significantly outperforms other tools in both quality and speed.
Speaking of reconstruction. Key2Mesh is yet another model that takes on 3D human mesh reconstruction, this time by utilizing 2D human pose keypoints as input instead of relying on visual data due to scarcity in image datasets with 3D labels.
TexDreamer can generate high-quality 3D human textures from text and images. It uses a smart fine-tuning method and a unique translator module to create realistic textures quickly while keeping important details intact.
TripoSR can generate high-quality 3D meshes from a single image in under 0.5 seconds.
MeshFormer can generate high-quality 3D textured meshes from just a few 2D images in seconds.
LGM can generate high-resolution 3D models from text prompts or single-view images. It uses a fast multi-view Gaussian representation, producing models in under 5 seconds while maintaining high quality.
En3D can generate high-quality 3D human avatars from 2D images without needing existing assets.