Skip to main content
Generative AI refers to a class of artificial intelligence systems that create new content—text, images, audio, code, or other media—by learning the underlying patterns of existing data. In this lesson you’ll learn what generative AI means, how it evolved from earlier AI techniques, and why it’s reshaping creative and productivity workflows today. We’ll start with a concise timeline showing how AI has progressed over time.
  • 1950s — Classical AI: rule-based systems and symbolic reasoning intended to encode expert knowledge explicitly.
  • 1990s — Machine Learning: statistical methods that learn patterns and relationships from data rather than relying solely on hand-coded rules.
  • 2010s — Deep Learning: multilayer neural networks that learn hierarchical features from very large datasets, enabling breakthroughs in vision, speech, and language.
  • 2020s — Generative AI: models that synthesize novel content (text, images, audio, code) by learning the distribution of training data and sampling from it.
A slide titled "What is Generative AI?" showing an evolutionary timeline of silhouettes from Artificial Intelligence (1950s) to Machine Learning (1990s), Deep Learning (2010s), and Generative AI (2020s). It also includes a brief definition noting machine learning is a subset of AI that learns from data to make decisions or predictions.
Deep learning, which rose to prominence in the 2010s, uses deep neural networks to process vast datasets and learn complex representations. These networks power image recognition, speech recognition, translation, and other applications that require understanding high-dimensional data. How generative AI differs
  • Traditional ML models are often discriminative: they classify or predict a label for input data (for example, “spam” vs “not spam”).
  • Generative models learn an approximation of the full data distribution and can sample from that distribution to produce entirely new examples that resemble the training data.
Common classes of generative models
  • Variational Autoencoders (VAEs): learn latent representations and generate samples by decoding from the latent space.
  • Generative Adversarial Networks (GANs): use a generator and discriminator in competition to produce highly realistic images and other media.
  • Transformer-based models and Large Language Models (LLMs): use attention mechanisms and massive training corpora to generate coherent text and support tasks like summarization, translation, and code generation.
Generative models approximate the data distribution and produce novel—but statistically plausible—outputs when sampling from that learned distribution. This enables creation of new images, text, audio, or code that resemble the training examples.
Practical examples and popular tools
  • ChatGPT — conversational text generation and assistants. OpenAI ChatGPT
  • DALL·E — image synthesis from text prompts. DALL·E
  • GitHub Copilot — AI-assisted code completion and generation. GitHub Copilot
Key capabilities enabled by generative AI
  • Content creation: synthetic images, text drafts, music, and video.
  • Code generation and automation: boilerplate, function suggestions, and auto-completion.
  • Data augmentation: generating synthetic examples for training or simulation.
  • Personalization: adapting content to user preferences at scale.
Risks and best practices Generative AI can produce realistic outputs that are fluent and persuasive, but important risks remain:
  • Hallucinations: models may assert incorrect facts as if they are true.
  • Biases: models can reproduce or amplify biases in their training data.
  • Copyright and provenance: generated content may inadvertently reproduce copyrighted material. Careful validation, human-in-the-loop review, and responsible deployment are essential.
Generative AI is powerful but not infallible. Outputs can be factually incorrect, biased, or inappropriate—always validate and apply safeguards before using generated content in critical or public contexts.
Quick reference table
EraCharacteristicTypical techniques
1950s — Classical AIRule-based, symbolic reasoningExpert systems, logic-based AI
1990s — Machine LearningStatistical pattern learningSVMs, decision trees, clustering
2010s — Deep LearningLearned hierarchical featuresCNNs, RNNs, deep neural networks
2020s — Generative AIContent synthesis from learned distributionsVAEs, GANs, Transformers / LLMs
Further reading and references This overview gives you the conceptual timeline and technical distinctions needed to understand why generative AI is a transformative area of modern AI research and product development.

Watch Video