Generating Text-to-3D Models Using OpenAI's Point-E, Gepetto AI & Unreal Engine

The meshes have room for improvement but look promising.

In December OpenAI, the developer of DALL-E and ChatGPT, announced a new AI model powered by Point-E, the company's system for generating 3D point clouds from complex prompts. Some developers have already started experimenting with it, integrating it into different apps. Here is one such experiment: TREE Industries presented some text-to-3D models generated in real time under 20 seconds in Gepetto AI – a technology layer aiming to create 3D worlds by speaking. It can now generate AI for avatars, 3D worlds from text prompts, and more.

The video demonstrates Point-E and Gepetto AI working together to make a dog, a bottle, a lamp, and a motorbike. TREE Industries is planning to use an image prompt at runtime to create the mesh and convert the vertex colors to a material.

Check out the progress on Twitter, follow Gepetto AI, and don't forget to join our 80 Level Talent platformour Reddit page, and our Telegram channel, follow us on Instagram and Twitter, where we share breakdowns, the latest news, awesome artworks, and more.

Keep reading

You may find this article interesting

Join discussion

Comments 3

  • AnyaEmberly AnyaEmberly

    I make over $25,000 doing a very easy and simple online job from home. Last month my friend got $50,000 from this job, giving only 2-3 hours a day. Everyone is starting to make q13 money online. Visit for more details…
    ..
    See ………..> 𝐖𝐰𝐰.𝐄𝐚𝐫𝐧𝐂𝐚𝐬𝐡𝟕.𝐜𝐨𝐦

    0

    AnyaEmberly AnyaEmberly

    ·3 hours ago·
  • Dubois Peter

    Currently it looks ridiculous, who needs this primitive chair? But they will develop this until it takes the jobs of artists. But before they need artists to train the software. It’s not a tool, it’s a replacement of artists.

    0

    Dubois Peter

    ·4 days ago·
  • Anonymous user

    Where can i go to use this?

    0

    Anonymous user

    ·28 days ago·

You might also like

We need your consent

We use cookies on this website to make your browsing experience better. By using the site you agree to our use of cookies.Learn more