Fast3D
Fast3D is a lightning‑fast AI‑powered 3D model generator that transforms text prompts or single/multi‑view images into professional‑grade mesh assets with customizable texture synthesis, mesh density, and style presets, all in under ten seconds without any modeling experience. It combines high‑fidelity PBR material generation with seamless tiling and intelligent style transfer, delivers precise geometric accuracy for realistic structures, and supports both text‑to‑3D and image‑to‑3D workflows. Outputs are compatible with any pipeline, offering export in GLB/GLTF, FBX, OBJ/MTL, and STL formats, while its intuitive web interface requires no login or setup. Whether for gaming, 3D printing, AR/VR, metaverse content, product design, or rapid prototyping, Fast3D’s AI core enables creators to explore diverse ideas through batch uploads, random inspiration galleries, and adjustable quality tiers, bringing concepts to 3D reality in seconds rather than days.
Learn more
GET3D
We generate a 3D SDF and a texture field via two latent codes. We utilize DMTet to extract a 3D surface mesh from the SDF and query the texture field at surface points to get colors. We train with adversarial losses defined on 2D images. In particular, we use a rasterization-based differentiable renderer to obtain RGB images and silhouettes. We utilize two 2D discriminators, each on RGB image, and silhouette, respectively, to classify whether the inputs are real or fake. The whole model is end-to-end trainable. As several industries are moving towards modeling massive 3D virtual worlds, the need for content creation tools that can scale in terms of the quantity, quality, and diversity of 3D content is becoming evident. In our work, we aim to train performant 3D generative models that synthesize textured meshes which can be directly consumed by 3D rendering engines, thus immediately usable in downstream applications.
Learn more
Magic3D
Together with image conditioning techniques as well as prompt-based editing approach, we provide users with new ways to control 3D synthesis, opening up new avenues to various creative applications. Magic3D can create high-quality 3D textured mesh models from input text prompts. It utilizes a coarse-to-fine strategy leveraging both low- and high-resolution diffusion priors for learning the 3D representation of the target content. Magic3D synthesizes 3D content with 8× higher-resolution supervision than DreamFusion while also being 2× faster. Given a coarse model generated with a base text prompt, we can modify parts of the text in the prompt, and then fine-tune the NeRF and 3D mesh models to obtain an edited high-resolution 3D mesh.
Learn more
Poly
Poly is an AI-enabled texture creation tool that lets you quickly generate customized, 8K HD, and seamlessly tile-able textures with up to 32-bit PBR maps using a simple prompt (text and/or image) in seconds. It's perfect for use in 3D applications such as 3D modeling, character design, architecture visualization, game development, AR/VR world-building, and much more. We're thrilled to share the result of our team's research work with the community and hope you will find it useful and fun. Type in a prompt, select a texture material type, and watch as Poly creates a fully-formed 32-bit EXR texture for you. You can use this to play around with Poly's AI, seeing what it is capable of and experimenting with prompting strategies. The dock at the bottom of the screen lets you switch views. You can view your past prompts, view a model in 3D, or view any of the six available physical-based rendering maps.
Learn more