2017
The Transformer Paper
Google publishes "Attention Is All You Need" — the architecture behind every modern LLM. Nobody outside AI research notices.
Research
2018 – 2019
GPT-1 & GPT-2
OpenAI shows that scaling up transformers with more data and parameters makes them surprisingly good at generating text. GPT-2 was considered "too dangerous to release."
Early models
2020
GPT-3: The Scale Breakthrough
175 billion parameters. For the first time, a language model could write essays, code, and translate — without being specifically trained for each task. The era of "general purpose" began.
Breakthrough
November 2022
ChatGPT: The Moment Everything Changed
OpenAI wraps GPT-3.5 in a simple chat interface. 100 million users in two months. Suddenly everyone — from students to CEOs — is talking to an AI. The "ChatGPT moment" from your podcast.
The tipping point
2023
The Race Begins
GPT-4 raises the bar. Google launches Gemini. Anthropic releases Claude. Meta open-sources Llama. China produces DeepSeek and Qwen. In one year, we go from one player to an entire industry.
Competition
2024
Models Get Specialised
Instead of one model fits all, providers release tiers: powerful models for complex work, fast models for quick tasks, tiny models you can run on your phone. Thinking vs Doing vs Checking.
Tiers emerge
2025
Reasoning Models
OpenAI's o1 and o3, DeepSeek R1 — models that "think step by step" before answering. They take longer but get hard problems right. A new tier: models that reason, not just predict.
New capability
2026
Where NYMG Is Today
The landscape moves fast. At NYMG, Anthropic Claude is primary (with different sizes for different agents), OpenAI Codex/GPT-5.4 is the fallback, and gpt-5.4 is used in the LangGraph translation pipeline.
You are here