OpenAI, the Elon Musk-founded artificial intelligence business that created the well-known DALL-E text-to-image generator, released POINT-E on Tuesday. POINT-E can create 3D point clouds directly from text commands. Point-E just requires one GPU and a few minutes, as opposed to existing systems like Google’s DreamFusion, which often take multiple hours and GPUs to make their visuals.
3D modeling is used across a variety of industries and applications. The CGI effects of modern movie blockbusters, video games, VR and AR, NASA’s moon crater mapping missions, Google’s heritage site preservation projects, and Meta’s vision for the Metaverse all hinge on 3D modeling capabilities. However, creating photorealistic 3D images is still a resource and time-consuming process, despite NVIDIA’s work to automate object generation and Epic Game’s RealityCapture mobile app, which allows anyone with an iOS phone to scan real-world objects as 3D images.
Text-to-Image systems like OpenAI’s DALL-E 2 and Crayon, DeepAI, Prisma Lab’s Lensa, and HuggingFace’s Stable Diffusion, have rapidly gained popularity, notoriety, and infamy in recent years. Text-to-3D is an offshoot of that research. Point-E, unlike similar systems, “leverages a large corpus of (text, image) pairs, allowing it to follow diverse and complex prompts, while our image-to-3D model is trained on a smaller dataset of (image, 3D) pairs,” the OpenAI research team led by Alex Nichol wrote in Point·E: A System for Generating 3D Point Clouds from Complex Prompts, published last week. “To produce a 3D object from a text prompt, we first sample an image using the text-to-image model and then sample a 3D object conditioned on the sampled image. Both of these steps can be performed in several seconds, and do not require expensive optimization procedures.”
If you were to input a text prompt, say, “A cat eating a burrito,” Point-E will first generate a synthetic view 3D rendering of said burrito-eating cat. It will then run that generated image through a series of diffusion models to create the 3D, an RGB point cloud of the initial image — first producing a coarse 1,024-point cloud model, then a finer 4,096-point. “In practice, we assume that the image contains the relevant information from the text, and do not explicitly condition the point clouds on the text,” the research team points out.
These diffusion models were each trained on “millions” of 3d models, all converted into a standardized format. “While our method performs worse on this evaluation than state-of-the-art techniques,” the team concedes, “it produces samples in a small fraction of the time.” If you’d like to try it out for yourself, OpenAI has posted the project’s open-source code on Github.
Subscribe to AFK Free Media on Google News.