In 2014, a PhD student at a Montreal bar sketched out an idea that would change art forever. Two years later, Google's DeepDream turned the internet psychedelic. By 2022, anyone with a keyboard could generate photorealistic images from a text description. Today, hundreds of millions of people use AI to create images every month.

The story of generative AI art is one of neural networks, viral algorithms, and a handful of breakthroughs that changed what it means to create. From the first AI system that could draw autonomously to the diffusion models powering today's image generators — this is that story.

1971
AARON — The First AI Artist
Harold Cohen begins building AARON at Stanford's AI Lab — a program that autonomously creates drawings and paintings for over 40 years.
1994
Evolving Virtual Creatures
Karl Sims uses genetic algorithms to evolve 3D creatures with bodies and behaviors generated entirely by computation — no human design.
2014
GANs Invented
Ian Goodfellow conceives Generative Adversarial Networks at a Montreal bar. Machines learn to generate images through competition.
2015
DeepDream & Neural Style Transfer
Google's DeepDream goes viral. Gatys et al. publish neural style transfer. Deep Dream Generator launches, making AI art accessible to everyone.
2018
StyleGAN & the $432,500 Portrait
NVIDIA's StyleGAN generates photorealistic faces. "Portrait of Edmond de Belamy" sells at Christie's for 43x its estimate.
2020
Diffusion Models
Ho et al. show that denoising diffusion can match and exceed GANs with stable training. The architecture that will power the revolution.
2022
DALL-E 2, Midjourney & Stable Diffusion
Three products launch within four months. Text-to-image generation becomes accessible to hundreds of millions of people.
2025
Nano Banana & AI Video
Google's Nano Banana goes viral with 200M+ image edits. Sora, Veo, and Kling extend generative art into video.
2026
Seedance 2.0 & the Hollywood Panic
ByteDance's multimodal video generator produces cinema-quality clips. The Dor Brothers' viral "$200M movie" triggers industry-wide alarm.
Prologue

Before AI Art: The Algorithmic Roots of Machine Creativity

Long before neural networks could generate images, artists and scientists were already asking the question: can a machine create art?

In the 1960s, mathematicians like Georg Nees, A. Michael Noll, and Frieder Nake — the "Three Ns" of computer art — used plotters and mainframes to generate visual compositions governed by mathematical rules with controlled randomness. Their work was algorithmic, not intelligent — but it planted the seed of an idea that would take decades to bloom.

AARON: The First AI Art Program (1971)

The first truly AI art system arrived in 1971, when British painter Harold Cohen visited Stanford's Artificial Intelligence Laboratory and began building AARON — a program designed to answer the question: "What are the minimum conditions under which a set of marks functions as an image?"

AARON evolved over more than forty years. It began with monochrome line drawings, gradually learned to produce foliage and human figures, and eventually chose and applied colors autonomously. Cohen described the program not as a tool but as a collaborator. When he died in 2016, AARON's development ended with him — its code was never open-sourced. A major retrospective was mounted at the Whitney Museum of American Art in 2024, finally cementing AARON's place in art history.

Evolving Virtual Creatures: AI-Driven Evolution (1994)

At SIGGRAPH 1994, computer scientist Karl Sims presented "Evolving Virtual Creatures" — a system that used genetic algorithms to evolve 3D creatures with morphologies and neural systems generated entirely by computation. Block-like beings learned to swim, walk, jump, and compete against each other for virtual resources. No human designed their bodies or programmed their behaviors. Evolution did.

It pointed toward an idea central to the AI art revolution that would follow: the most interesting generative art comes not from designing outputs directly, but from designing systems capable of surprise.

Chapter I

The Neural Network Revolution in AI Art (2014–2021)

Everything changed twice in rapid succession. First, a PhD student at a Montreal bar invented a new way for machines to learn to generate images. Then, a Google engineer woke from a nightmare at 2 AM and ran an experiment that would give the world its first taste of AI art.

Generative Adversarial Networks (GANs): How AI Learned to Create Images

In June 2014, Ian Goodfellow — a doctoral student at the University of Montreal — went drinking with colleagues at Les 3 Brasseurs, a local bar. A friend described a project on generating photorealistic images by computer. Goodfellow went home that night, coded his idea, and it worked on the first try.

The idea was Generative Adversarial Networks (GANs): two neural networks locked in competition. A Generator creates images. A Discriminator tries to tell them apart from real ones. The Generator improves by trying to fool the Discriminator. The Discriminator improves by getting better at detecting fakes. Through this adversarial process, both networks get better — and the generated images become startlingly realistic.

Generative Adversarial Network Architecture
Random
Noise
Generator
Creates images
Discriminator
Real or fake?
Real
Images
← Feedback loop: Discriminator trains Generator to improve →

MIT Technology Review dubbed Goodfellow "The GANfather" — the man who gave machines the gift of imagination. GANs dominated image generation for the next five years.

In December 2018, NVIDIA released StyleGAN, which could generate photorealistic human faces at unprecedented quality. Two months later, Uber engineer Phillip Wang used StyleGAN to create This Person Does Not Exist — a website that displayed a new AI-generated face on every page reload. It went viral instantly, introducing millions of people to the uncanny capabilities of generative AI.

NVIDIA StyleGAN — grid of AI-generated photorealistic faces that do not exist.
NVIDIA StyleGAN (2018). "These people are not real — they were produced by our generator." StyleGAN could generate photorealistic human faces indistinguishable from photographs, marking a watershed moment for GANs. Image courtesy of NVIDIA Research.

Google DeepDream: The AI Art Moment That Went Viral

Early in the morning of May 18, 2015, Google engineer Alexander Mordvintsev woke from a nightmare at 2 AM. Unable to sleep, he decided to try an experiment he had been thinking about — running an image recognition neural network in reverse. Instead of asking the network to identify what was in an image, he asked it to amplify whatever patterns it detected, then feed the result back through, again and again.

The results were hallucinatory. Dog eyes emerged from clouds. Pagodas sprouted from trees. Fractal spirals of animal faces cascaded across landscapes. The neural network was dreaming — projecting its training data (ImageNet, which contained 120 dog breed categories) onto whatever it saw.

One of Alexander Mordvintsev's earliest DeepDream images showing hallucinatory neural network patterns.
Google DeepDream (2015). One of the very first DeepDream images produced by Alexander Mordvintsev. Running a neural network in reverse amplified whatever patterns the network detected — the technique, published as "Inceptionism," became the first widely viral AI art phenomenon.
DeepDream image — hallucinatory neural network patterns and fractal details.
DeepDream (2015). Web platforms soon made the technique accessible to anyone — no coding or GPU hardware required. Users experimented with the algorithm, creating their own hallucinatory neural network visions.

Google published the results as "Inceptionism: Going Deeper into Neural Networks" on June 17, 2015. Mordvintsev's core experiment was just 30 lines of code. When Google open-sourced the software in July 2015, DeepDream images flooded the internet. For the first time in history, AI-generated art went mainstream.

The effect was even more striking in motion. When applied to video, DeepDream produced continuously morphing hallucinations — neural network dreams flowing through time:

Deep Dream Generator: The First AI Art Platform

DeepDream's psychedelic imagery captivated the internet, but creating these images required coding skills and GPU hardware most people did not have. Deep Dream Generator, built by developer Kaloyan Chernev in 2015, was arguably the first modern generative AI platform — the first website that let ordinary people create AI-generated images without writing code or configuring servers. While Harold Cohen's AARON had existed since the 1970s, it was a private research project. DDG was something new: a public, browser-based tool that put generative AI in anyone's hands.

The platform's trajectory mirrored the field itself. It evolved from a single-purpose DeepDream tool to incorporate neural style transfer, then text-to-image generation, and eventually grew into a multi-model environment — a pattern that other platforms would follow in the years ahead.

Neural Style Transfer: How AI Combines Art Styles

Two months after DeepDream, in August 2015, researchers Leon Gatys, Alexander Ecker, and Matthias Bethge published "A Neural Algorithm of Artistic Style" — a paper that demonstrated something remarkable. A neural network could separate the content of one image from the style of another, then recombine them. Your photograph, rendered in Van Gogh's brushwork. Your selfie, in Picasso's cubist vocabulary.

How Neural Style Transfer Works
Content
+
Style
=
Result
Content features from deeper layers + style features from texture correlations = stylized output

The implications were immediate and commercially explosive. The app Prisma, launched in 2016, brought neural style transfer to smartphones and was downloaded over 70 million times. On the web, Deep Dream Generator became one of the most popular destinations for style transfer — its community of artists pushing the technique far beyond simple filters, blending styles and iterating on results to produce genuinely striking work. Soon millions of people were transforming their photographs using the visual language of any artistic style.

Original input photograph
Input
+
Style reference painting
Style
=
Neural style transfer result — content of the input image rendered in the artistic style
Result
Neural Style Transfer in action. The algorithm separates the content of one image from the style of another, then recombines them — producing a new image that preserves the original scene but renders it in an entirely different artistic language.

Style transfer proved something important: neural networks had not just learned to recognize art. They had learned something about the structure of visual style itself — what makes a Monet look like a Monet, what makes a woodcut feel like a woodcut. That understanding, embedded in layers of mathematical weights, was itself a kind of aesthetic knowledge.

AI-Generated Art at Auction: StyleGAN and the $432,500 Portrait

On October 25, 2018, a blurry portrait of a fictional man in a dark coat sold at Christie's in New York.

Portrait of Edmond de Belamy — the first AI-generated artwork sold at a major auction house.
"Portrait of Edmond de Belamy" (2018). Generated by a GAN trained on 15,000 portraits spanning the 14th to 20th centuries. The blurred, ghostly quality is characteristic of early GAN outputs.
"Portrait of Edmond de Belamy" by Obvious
$432,500
Pre-sale estimate: $7,000 — $10,000
First AI artwork sold at a major auction house.
Created with a GAN trained on 15,000 portraits from WikiArt.

The sale made international headlines and sparked fierce debate. The French art collective Obvious had created the portrait using a GAN, but the underlying code was largely written by Robbie Barrat, a then-19-year-old AI artist from West Virginia who had published his work as open source. Questions about authorship, credit, and what it means to "create" AI art remain unresolved to this day.

The name itself was a wink: "Belamy" derives from "bel ami" — French for "good friend," a nod to Goodfellow, the inventor of GANs.

Pioneer AI Artists: Klingemann, Barrat, and Herndon

The late 2010s saw the emergence of artists who did not merely use AI as a tool but made the technology itself their medium.

Mario Klingemann, a German artist and self-described "neurographer," created Memories of Passersby I — an installation that uses neural networks to generate an infinite, never-repeating stream of portraits in real time. In March 2019, it became the first AI artwork sold at Sotheby's. Klingemann's central insight: the art is not the images, which disappear, but the code that creates them.

Refik Anadol brought AI art to monumental scale. His exhibition Unsupervised at the Museum of Modern Art in New York (November 2022) trained a StyleGAN on 138,151 images from MoMA's own collection, creating a constantly shifting data sculpture on a massive screen in the museum's lobby. Over 3 million people visited — one of MoMA's most popular exhibitions ever. MoMA acquired the work for its permanent collection.

Holly Herndon explored AI as collaborator rather than tool. Her 2019 album PROTO was created alongside Spawn, a neural network trained to recognize and replicate human voices through live "training ceremonies" — performances where participants sang to the AI. The project blurred the line between creator and creation in ways that no visual art had achieved.

Chapter II

Diffusion Models and Text-to-Image AI Art (2022–Present)

If GANs proved that machines could generate images, diffusion models proved they could generate whatever you asked for.

The core idea, formalized in a 2020 paper by Jonathan Ho, Ajay Jain, and Pieter Abbeel at UC Berkeley, was counterintuitive: train a neural network to remove noise from images. Start with a clear image, add noise gradually until it becomes pure static, then teach the network to reverse the process. At generation time, begin with pure noise and denoise step by step. What emerges is a new image — coherent, detailed, and controlled by a text description.

How Diffusion Models Work
t = 0
t = 250
t = 500
t = 750
t = 1000
Forward: a structured image is progressively corrupted with noise
Reverse: the model learns to recover structure from noise
t = 1000
t = 750
t = 500
t = 250
t = 0
"a cat sitting on a windowsill at sunset" → the text prompt guides each denoising step

Diffusion models were more stable to train than GANs, produced more diverse outputs, and — critically — could be guided by natural language. You could simply describe what you wanted to see.

DALL-E 2, Midjourney, and Stable Diffusion: The 2022 AI Image Generation Explosion

Three products launched within four months of each other in 2022, each approaching text-to-image generation differently. Together, they detonated.

DALL-E 2 (April 2022) — OpenAI's second-generation model combined diffusion with CLIP (Contrastive Language-Image Pre-training) to generate images at four times the resolution of the original DALL-E. The jump in quality from DALL-E to DALL-E 2 was described as jaw-dropping. Text-to-image generation was suddenly good.

DALL-E 2 generated image: Teddy bears working on new AI research underwater with 1990s technology.
DALL-E 2 (2022). "Teddy bears working on new AI research underwater with 1990s technology." A single text prompt could now conjure photorealistic scenes that combined absurd concepts with convincing detail — something no previous system could achieve.

Midjourney (July 2022) — Founded by David Holz (co-founder of Leap Motion), Midjourney launched its open beta in July 2022. Self-funded and profitable from day one, it developed a distinctively painterly, aesthetic output that attracted artists and designers. By mid-2023, it had over 16 million registered users.

In September 2022, a Midjourney-generated image made international headlines when Jason Allen's Théâtre D'opéra Spatial won first place in the digital art category at the Colorado State Fair. Traditional artists were outraged. Allen had spent 80+ hours refining over 600 prompt iterations — but the debate about whether AI-generated work should compete alongside human-made art had arrived, and it wasn't going away.

Théâtre D'opéra Spatial by Jason Allen, created with Midjourney — first AI artwork to win a fine arts competition.
Théâtre D'opéra Spatial (2022) by Jason Allen, created with Midjourney. The first AI-generated image to win a fine arts competition. The U.S. Copyright Office later denied copyright registration, ruling the AI-generated elements were not human-authored.

Stable Diffusion (August 2022) — Stability AI released Stable Diffusion as open source on August 22, 2022 — a move that changed everything. The model weights were public. An optimized version could run on a consumer GPU with 2.4 GB of VRAM. Within weeks, an ecosystem of community-built tools, fine-tuned models, and specialized applications exploded into existence.

The combined impact was seismic. Within months, AI-generated images were appearing in advertising, game development, publishing, and social media. Debates about art, authorship, copyright, and labor erupted across every creative industry.

Chapter III

AI Image Generation in 2025–2026: Flux, Nano Banana, and AI Video

By 2026, the field has matured rapidly. Models like Flux (from Black Forest Labs, founded by former Stability AI researchers) and Kling 3.0 (from Kuaishou) push photorealism and prompt accuracy to levels that would have seemed implausible in 2022. Text rendering — once a telltale weakness of AI-generated images — is now largely solved. Hands, faces, and complex spatial relationships are handled with increasing reliability.

AI-generated festive raspberry cupcake with photorealistic detail — Flux model
AI-generated photorealistic family portrait — Flux model
AI-generated cave-inspired bedroom interior design — Flux model
Modern AI image generation. Photorealistic images generated by Flux — product photography, portraiture, and architectural visualization. The level of detail in textures, lighting, and spatial coherence would have been impossible just two years earlier.

In August 2025, Google's Nano Banana — the playful codename for Gemini's image generation model — became the first AI image tool to truly go viral with mainstream audiences. An anonymous entry on the crowd-sourced AI evaluation platform LMArena, the model was unmasked after it dominated the leaderboard. Within weeks of its public release, the "3D figurine" trend it spawned swept Instagram and X: millions of users turned their selfies into photorealistic toy-like figurines. Google reported over 10 million new Gemini users and 200 million image edits in the weeks following launch. The upgraded Nano Banana Pro (Gemini 3 Pro Image) followed in November 2025, pushing text rendering and world knowledge further. It was a reminder that generative art's biggest moments are often driven not by technical benchmarks but by a single creative idea that captures public imagination.

Google Nano Banana — trending AI image generation examples including the viral 3D figurine trend
Google's Nano Banana trends. The "3D figurine" craze swept social media in 2025, with millions of users turning their photos into photorealistic toy-like figurines using Gemini's image generation. Image: Google.

Video generation has emerged as the next frontier. OpenAI's Sora, Google's Veo, and Kuaishou's Kling can generate video clips from text descriptions — a capability that extends generative art into time-based media.

In February 2026, ByteDance's Seedance 2.0 detonated across social media. The model introduced a unified multimodal architecture — accepting text, images, audio, and video as inputs simultaneously — and could generate multi-shot sequences with natural cuts and transitions in a single pass. When AI filmmaker Yonatan Dor (The Dor Brothers) used Seedance 2.0 to produce what he described as a "$200M movie in one day," the results went viral. A generated fight scene between AI versions of Tom Cruise and Brad Pitt accumulated millions of views and triggered cease-and-desist letters from Disney and Paramount. Hollywood panicked. The Motion Picture Association denounced the technology. But the genie was out of the bottle — generative video had crossed the threshold from curiosity to threat.

Perhaps most importantly, the platforms that deliver these capabilities have matured from single-model experiments into comprehensive creative environments. What began as one-trick tools — a DeepDream visualizer here, a style transfer app there — have evolved into multi-model platforms where users can access dozens of image, video, and audio generation models from a single interface. The barrier to entry has never been lower.

Epilogue

What Connects All Generative AI Art: Rules and Surprise

Step back far enough and a single thread runs through this entire history: the relationship between rules and surprise.

Georg Nees wrote rules for placing squares. The randomness within those rules produced results he did not expect. Vera Molnar created systems of constraints, then delighted in the configurations her algorithms discovered. Karl Sims designed fitness functions for virtual creatures, then watched evolution produce locomotion strategies no human engineer would invent. Ian Goodfellow set two neural networks against each other and let competition drive them toward photorealism.

Every generative art system works the same way: the artist designs a space of possibilities. The system explores it. The best generative art emerges from the tension between what the artist intended and what the system discovered — between control and chaos, between order and des ordres.

"The idea becomes a machine that makes the art."

— Sol LeWitt, "Paragraphs on Conceptual Art," 1967

Sol LeWitt wrote that sentence in 1967, decades before neural networks existed. He was talking about wall drawings executed by gallery assistants following written instructions. But the statement applies perfectly to a text prompt entered into an AI image generator in 2026. The idea — expressed as a set of instructions, a mathematical function, or a natural-language description — becomes a machine that makes the art.

What has changed is the machine's sophistication. A plotter drawing squares in 1965. A fractal algorithm mapping infinity in 1985. A genetic algorithm evolving creatures in 1994. A GAN generating faces in 2018. A diffusion model conjuring any scene from a sentence in 2025. The trajectory is unmistakable: the gap between intention and output is shrinking. The space of possibilities the machine can explore is expanding.

But the fundamental creative act remains the same. Someone has to decide what to explore. Someone has to recognize when the machine has found something worth keeping. Someone has to frame the question the system tries to answer.

Generative art has never been about removing the artist. It has been about giving the artist a more powerful collaborator — one that can search possibility spaces no human could navigate alone, that can produce variations at a pace no hand could match, that can discover visual ideas no one thought to look for.

Georg Nees, pinning his plotter drawings to a gallery wall in 1965, would recognize the practice immediately. The machines are faster now. The art is larger. The audience is everyone.

But the question at the heart of it — what happens when you set a good system loose and pay attention to what it finds? — has not changed at all.

Comments (10)