The History of AI
An interactive timeline of the breakthroughs that brought us here
The Turing Test
Alan Turing publishes "Computing Machinery and Intelligence", proposing the Turing Test — can a machine fool a human into thinking it's a person?
AI Gets Its Name
The Dartmouth Conference takes place. The term "Artificial Intelligence" is coined for the first time, launching AI as a formal field of study.
ELIZA Chatbot
ELIZA, created at MIT, becomes the first conversational AI. It simulates a psychotherapist and convinces some users it truly understands them.
Deep Blue vs. Kasparov
IBM's Deep Blue defeats world chess champion Garry Kasparov, the first time a computer beats a reigning champion under standard tournament rules.
Watson & Siri
IBM Watson wins Jeopardy! against human champions. Apple launches Siri, bringing voice assistants to millions of pockets worldwide.
Deep Learning Breakthrough
AlexNet wins the ImageNet competition by a massive margin, proving deep neural networks can dramatically outperform traditional methods at image recognition.
GANs Invented
Ian Goodfellow invents Generative Adversarial Networks (GANs) — two neural networks that compete to generate increasingly realistic images, audio, and more.
AlphaGo Stuns the World
Google DeepMind's AlphaGo defeats world Go champion Lee Sedol 4–1. Go was considered too complex for computers — with more possible positions than atoms in the universe.
Attention Is All You Need
Google publishes the landmark paper introducing the Transformer architecture. This single innovation becomes the foundation for virtually all modern AI models.
GPT-1 & BERT
OpenAI releases GPT-1, the first Generative Pre-trained Transformer. Google releases BERT. The era of large language models begins.
GPT-3: Emergent Abilities
OpenAI releases GPT-3 with 175 billion parameters. It demonstrates surprising emergent abilities — few-shot learning, code generation, and creative writing — simply from scale.
AI Goes Mainstream
ChatGPT launches in November and reaches 100 million users in two months. Stable Diffusion open-sources image generation. AI becomes a household word.
The Cambrian Explosion
GPT-4 released. Anthropic launches Claude 2. Meta open-sources Llama 2. AI coding tools boom. Every major tech company races to build and deploy AI products.
Open Source Catches Up
Claude 3.5 Sonnet, GPT-4o, and Gemini push the frontier. Open-source models rival proprietary ones. AI agents begin to emerge as a new paradigm.
The Age of Agents
Claude 4 / Opus arrives. AI coding agents (Claude Code, Codex CLI) ship. Agentic AI becomes mainstream — AI systems that can plan, use tools, and take actions autonomously.
Key AI Concepts — In Plain English
Every term explained with one-sentence definitions, real-world analogies, and optional technical depth
Artificial Intelligence(AI)
Making computers do things that normally require human intelligence — recognizing images, understanding language, making decisions.
Teaching a calculator to not just add numbers, but understand what numbers mean.
Machine Learning(ML)
Instead of programming rules, you show the computer examples and it figures out the rules itself.
Learning to cook by tasting 10,000 dishes vs. reading a recipe. You eventually figure out what makes food taste good.
Neural Network
A system inspired by the human brain, made of layers of connected nodes (neurons) that process information.
A team of people where each person checks one thing, and they all vote on the answer. With enough people and layers, they can recognize anything.
Deep Learning
Machine learning using artificial neural networks with many layers, capable of learning incredibly complex patterns.
Like ML but with a much bigger brain — it can spot patterns humans never could, like recognizing faces from millions of pixels.
Transformer
The breakthrough architecture (2017) that made modern AI possible. Uses "attention" to understand which words in a sentence relate to each other.
Reading a sentence and highlighting which words are connected, instead of reading left to right. In "The cat sat on the mat because it was tired" — the Transformer knows "it" refers to "cat."
Training
The process of feeding data to a model so it learns patterns. For large models, this costs millions of dollars and takes weeks on thousands of GPUs.
Going to school for 12+ years. It's expensive and time-consuming, but once you've graduated, you can answer questions quickly.
Parameters
The internal "knobs" a model learns during training — each one is a number that controls how the model processes information. More parameters generally means more capability.
A mixing board in a music studio with billions of sliders. During training, the AI adjusts every slider until the music sounds right. A bigger board (more parameters) can produce richer, more nuanced music.
Tokens
The pieces that AI breaks text into before processing. Not always whole words — sometimes parts of words, punctuation, or spaces.
Lego bricks. You build sentences from smaller pieces. The AI reads and writes one piece at a time.
Large Language Model(LLM)
A neural network trained on massive amounts of text that can generate, understand, and reason about language.
Someone who has read every book, article, and website ever written, and can write in any style — from Shakespeare to code to legal contracts.
GPT(Generative Pre-trained Transformer)
A specific type of LLM architecture by OpenAI. G = generates text, P = pre-trained on lots of data, T = uses the Transformer architecture.
A specific brand of car engine. "LLM" is the category (all car engines), "GPT" is a particular design (like a V8 from a specific manufacturer).
Context Window
How much text the AI can "see" at once — both your input and its output must fit within this window.
The size of the AI's desk. A bigger desk means it can spread out more documents and see more information at once.
Inference
When the AI actually generates a response. The "using" phase as opposed to the "learning" phase.
Training is studying for the exam. Inference is taking the exam.
Prompt
The text you send to an AI — your question, instruction, or input. Everything the model sees before generating a response.
The question on an exam paper. The clearer and more specific the question, the better the answer you'll get.
Prompt Engineering
The art of writing prompts that get the best results from AI — including techniques like giving examples, assigning roles, and structuring instructions.
Knowing how to ask the right question to get the right answer. "Tell me about dogs" vs. "Compare 3 hypoallergenic dog breeds for apartments, in a table."
Chain of Thought(CoT)
Getting AI to show its reasoning step by step instead of jumping straight to an answer, which dramatically improves accuracy on complex problems.
Showing your work in math class. The AI "thinks out loud" and catches mistakes along the way.
Temperature
A setting that controls how creative/random vs. deterministic the AI's output is. Low temperature = predictable, high temperature = creative and varied.
A dial between "play it safe" and "get creative." Temperature 0 always picks the most likely word; temperature 1+ takes more risks.
Hallucination
When AI confidently generates incorrect or made-up information — citing fake sources, inventing facts, or getting details wrong.
A student who doesn't know the answer but writes something convincing anyway. It sounds right, but it's completely fabricated.
Fine-tuning
Taking a pre-trained model and training it further on specific data for a specific task — like specializing after a general education.
A doctor who went to medical school (pre-training) and then specialized in cardiology (fine-tuning). Same brain, focused expertise.
Reinforcement Learning(RL)
Training AI by giving it rewards for good behavior and penalties for bad behavior, so it learns what to do through trial and error.
Training a dog — treat for sitting, no treat for barking. Over time, the dog figures out what earns rewards.
RLHF(Reinforcement Learning from Human Feedback)
Humans rate AI outputs as good or bad, and the AI learns from that feedback to produce more helpful, harmless, and honest responses.
A chef getting reviews from food critics and adjusting recipes based on what diners actually enjoy.
Embedding
Converting text (or images, audio) into numbers (vectors) so the AI can understand meaning and compare similarity.
Giving every word a GPS coordinate in "meaning space." Similar words like "happy" and "joyful" end up close together on the map.
RAG(Retrieval-Augmented Generation)
Letting the AI search a knowledge base before answering, so it has access to up-to-date or private information it wasn't trained on.
An open-book exam vs. a closed-book exam. The AI looks up information in your documents before answering, instead of relying only on memory.
Multimodal
AI that can understand and generate multiple types of content — text, images, audio, video — not just one format.
A person who can read, see photos, listen to music, and watch videos — not just read text. They understand the world through multiple senses.
Agent
An AI system that can take actions, use tools, and make decisions autonomously — not just answer questions, but actually do things.
Regular AI answers questions. An agent can actually DO things — browse the web, write and run code, create files, send emails, and plan multi-step workflows.
Open Source vs. Closed Source Models
Open source means the model's code and weights are public and free to use (Llama, Mistral). Closed source means you can only access it via an API (GPT-4, Claude).
A recipe you can see, modify, and cook at home vs. a secret sauce you can only order at the restaurant.
MCP (Model Context Protocol)
A universal standard that lets AI assistants connect to external tools and data sources — like apps on your phone, but for AI.
USB for AI. Before USB, every device needed a different cable. MCP is the one standard plug that lets any AI tool connect to any service — GitHub, Slack, databases, cloud providers, and more.
AI Cheatsheet
Quick reference for models, companies, numbers, and buzzwords