DALL-E is an AI image generation model developed by OpenAI that creates images from text prompts using a transformer-based neural network architecture. Released in January 2021, it was one of the first major text-to-image models to demonstrate high-quality, coherent image synthesis from natural language descriptions, sparking widespread public interest in AI-generated imagery.
The model was trained on a large dataset of text-image pairs and learned to understand the relationships between language and visual concepts, allowing it to generate images of objects, scenes, and ideas that it had never explicitly seen during training. DALL-E could combine disparate concepts in creative ways, generating images of things like "an armchair in the shape of an avocado" or "a store front that has the word 'openai' written on it." While the original DALL-E had limitations in resolution and consistency, it proved the viability of large-scale text-to-image generation and set the stage for subsequent models.
DALL-E's release marked a turning point in public awareness of AI's creative capabilities, demonstrating that AI could generate novel, imaginative imagery rather than simply remixing existing content. The model's name is a portmanteau of the surrealist artist Salvador Dalí and the Pixar character WALL-E, reflecting both the creative and technological aspirations of the project.