Stanford researchers have developed an artificial intelligence (AI) model, EG3D, that can generate high-resolution arbitrary images of faces and other objects, along with underlying geometric structures.

Artificially intelligent models have recently advanced to the point where users will soon be able to use these models to build and alter near-photorealistic three-dimensional landscapes from the comfort of their laptops. Because these technologies make it easy to generate hyper-realistic avatars, they will revolutionize the way artists working on video games and CGI for movies approach their work. AIs have been able to create realistic 2D images for quite some time now. However, 3D scenarios have proven to be more challenging due to the enormous computing power required. Created by a team of Stanford academics, the AI ​​model EG3D can be used to create arbitrary high-resolution images of faces and other things with underlying geometric structure. This model is one of the first 3D models now in use to achieve a rendering quality close to photorealism.

A very popular machine learning method known as a generative adversarial network (GAN) is used by EG3D and its predecessors to create images. Using one neural network to create images and another to assess their accuracy, these systems pit two neural networks against each other. Until the result is achievable, this process is done several times. The researchers created a component that can transform these images for 3D space by combining features of pre-existing high-resolution 2D GANs. This two-part building achieves two goals at once. Plus, it’s fast enough to work on a laptop in real time and can be used to create complex 3D designs. It is compatible with current architectures and has efficient computing performance.

While it is possible to use tools like EG3D to create near-lifelike 3D images, the question still remains of how difficult it is to turn them into design software. This is because, despite the result being an image that can be seen, it is unclear how the GANs made it. A machine learning model called GiraffeHD, developed by researchers at the University of Wisconsin-Madison, may be helpful in this situation. This model is effective in removing manipulable features from 3D images. It allows the user to choose numerous elements including shape, color and the scene or background of the image. GiraffeHD was trained using numerous photos. To construct these images so that these many aspects behave as controllable variables, the model looks for latent factors in the image. Users could precisely modify attributes for desired scenarios by editing these manageable aspects in 3D-generated photos in the future. A more significant trend is the use of AI to create 3D photos, including EG3D and Giraffe HD. However, much work remains to be done with regard to algorithmic bias and wider applicability. The type of nutritional training data still limits these models. Research is still underway to address these issues.

Although it is still in its infancy, this research opens up possibilities for more realistic 3D images and models. It will be interesting to see where this line of research goes and how it can be applied in the future. I’d love to hear your thoughts on this new approach in our ML subreddit

This Article is written as a summary article by Marktechpost Staff based on the paper 'Efficient Geometry-aware 3D Generative Adversarial Networks'. All Credit For This Research Goes To Researchers on This Project. Checkout the paper, github and reference article.

Please Don't Forget To Join Our ML Subreddit

Leave a Comment

Your email address will not be published.