on June 6, hugging facea company that hosts open source artificial intelligence projects, saw traffic to an AI image generation tool called DALL-E Mini shoot up.
The seemingly simple app, which generates nine images in response to a typed text prompt, was launched almost a year ago by an independent developer. But after some recent improvements and a few viral tweets, the ability to roughly sketch all sorts of surreal, hilarious, and even nightmarish visions suddenly became meme magic. See the renditions of “Thanos is looking for his mother at Walmart,†drunken men without shirts who wander in Mordor,†CCTV camera footage of Darth Vader break dance,” and “a hamster Godzilla in a sombrero attacking Tokyo†
As more people created and shared DALL-E Mini Pictures on Twitter and reddit, and more new users arrived, Hugging Face saw its servers flooded with traffic. “Our engineers didn’t sleep the first night,” said Clément Delangue, CEO of Hugging Face, during a video call from his home in Miami. “It’s really hard to operate these models at scale; they had to fix everything.” For the past few weeks, DALL-E Mini has been displaying approximately 50,000 images per day.
The DALL-E Mini’s viral moment doesn’t just herald a new way to create memes. It also provides an early look at what could happen when AI tools make those on-demand images widely available, and a reminder of the uncertainties about their potential impact. Algorithms that generate custom photography and artwork can transform art and help businesses with marketing, but they can also have the power to manipulate and deceive. A warning on the DALL-E Mini webpage warns that it may “strengthen or exacerbate societal prejudices” or “generate images that contain stereotypes against minority groups.”
DALL-E Mini is inspired by a more powerful AI image creation tool called DALL-E (a portmanteau of Salvador Dali and WALL-E), unveiled by AI research firm OpenAI in January 2021. DALL-E is more powerful, but is not openly available due to concerns that it will be abused.
It has become common for breakthroughs in AI research to be replicated quickly elsewhere, often within months, and DALL-E was no exception. Boris Dayma, a machine learning consultant based in Houston, Texas, says he was fascinated by the original DALL-E research paper. Although OpenAI did not release any code, it managed to put together the first version of DALL-E Mini during a hackathon hosted by Hugging Face and Google in July 2021. The first version produced low-quality images that were often difficult to spot, but Dayma has continued to improve it ever since. Last week he renamed his project: Craiyon, after OpenAI requested it, he changed the name to avoid confusion with the original DALL-E project. Ads will run on the new site, and Dayma is also planning a premium version of its image generator.
DALL-E Mini images have a distinctive alien look. Objects are often distorted and blotchy, and people appear with faces or body parts that are missing or disfigured. But it’s usually possible to spot what it’s trying to display, and it’s often fun to compare the AI’s sometimes unhinged output to the original prompt.