2-Min AI Newsletter #16
Asif Razzaq
AI Research Editor | CEO @ Marktechpost | 1 Million Monthly Readers and 80k+ ML Subreddit Members
Featured Post?????????
Nowadays, it is difficult to pass a day without reading/hearing about a new application of diffusion models if you are following the news about artificial intelligence/machine learning. The massive success of diffusion models like DALL-E and Stable Diffusion has attracted enormous attention to these applications.
What if we could go deeper? What if we use these generated images to train another AI model to achieve a task? How about teaching a robot to do something? That’s the question DALL-E-Bot tries to answer.
DALL-E-Bot tries to tackle the object rearrangement problem. Since diffusion models can generate realistic images, the authors wanted to examine their capabilities of arranging objects in the scene in a natural way. For example, “kitchen tabletop with utensils” prompt will generate a realistic-looking image where the utensils and the plate are neatly placed if you pass it to DALL-E. Based on this observation, DALL-E-Bot uses a diffusion model to generate the goal for the robot. Once the robot sees this image, it will know what the final object arrangement should look like.?