What Are LLMs and Why Should You Care?
Listen First
Start with the podcast. Two hosts walk you through everything you need to know about LLMs - what they are, how they work, and why they sometimes get things wrong.
What You'll Learn
- The ChatGPT moment
- How LLMs are trained
- Prediction vs thinking
- LLM ≠ computer
- Tokens & costs
- Context window
- Knowledge cutoff
- Hallucination
Study the Visual
See how LLM prediction actually works - from training data to next-token output.
- An LLM is a prediction engine, not a thinking machine
- It's trained on text data - that data IS its worldview
- It's probabilistic (not a computer) - optimised for fluency, not truth
- Everything runs on tokens, with a hard context window limit
- On its own: no internet, no memory, no action
The ChatGPT Moment
In November 2022, OpenAI released ChatGPT. The technology wasn't new - Google invented it in 2017 - but the simple chat interface made it accessible to everyone. Like the web browser did for the internet.
How LLMs Are Built
An LLM is trained on enormous amounts of text - books, websites, papers, code. This training data IS its entire worldview. It has read about the world but never experienced it.
Predicts, Doesn't Think
An LLM predicts the most likely next word based on patterns. It's the world's most sophisticated autocomplete. It does not reason or understand.
Not a Computer
Computers are deterministic - 2+2 always = 4. LLMs are probabilistic - they deal in likelihood, not certainty. That's why they can't reliably count the R's in "strawberry" (they see tokens, not individual letters).
Tokens & Context Window
A token ≈ ¾ of a word. Everything costs tokens. The context window is the AI's desk - everything must fit. Some models handle up to 1 million tokens (~1,500 pages), but instructions, conversation, and answers all share that space.
Knowledge Cutoff
LLMs are frozen in time after training. They don't learn from new events. Ask about last week's news and they'll either say "I don't know" or confidently make something up.
Hallucination
When AI generates confident but completely made-up information. It's optimised for fluency, not truth. It has no fact-checker and no humility - it can't say "I'm not sure."
What LLMs Cannot Do (On Their Own)
- Browse the internet - frozen at training cutoff
- Remember past conversations - every chat starts from zero
- Take action - can write an email but can't send it
- Have you used ChatGPT or another AI tool? What was your experience?
- Can you think of a time you trusted information that turned out to be wrong? How does that relate to hallucination?
- What surprised you most about how LLMs actually work?
- What's one thing you'd want to verify before trusting an AI answer?
Day 1 Complete!
You've mastered the fundamentals. Before moving on - see how fast this all happened:
🎬 Bonus: Ryan Serhant on AI
You know him from Owning Manhattan - hear his take on AI and trust.
The Landscape: Who's Who and What to Use
AI is not one thing from one company. Different models for different jobs. Smart companies match the tool to the task.
Day 2 is locked
Complete Day 1 flashcards to unlock.
Listen First
Today we zoom out from the technology to the landscape. Who builds these AI models, why they're all different, and how NYMG navigates the choices.
What You'll Learn
- The major AI companies
- What "open source" means
- Why models feel different
- Model versions & upgrades
- What model "size" means
- Thinking / Doing / Checking
- NYMG's AI journey
- Passive vs Active AI
Study the Visual
The wider AI landscape at a glance - and the simple setup NYMG uses in practice.
- AI is not one company - it's a global ecosystem of providers
- Different models feel different because they're optimised for different goals
- Models come in tiers: thinking (expensive), doing (workhorse), checking (cheap)
- NYMG uses multiple tools - Claude for heavy lifting, ChatGPT for helpdesk
- Everything so far is passive AI - you ask, it answers
The Major Players
| Company | Model | Known For |
|---|---|---|
| OpenAI | GPT / ChatGPT | Started the revolution. Biggest consumer AI. |
| Anthropic | Claude | Safety-focused. Massive context window. NYMG's primary AI. |
| Gemini | Invented the technology (2017). Integrated with Workspace. | |
| Meta | Llama | Open source - free for anyone to download and use. |
| xAI | Grok | Elon Musk's company. Growing quickly. |
| DeepSeek | DeepSeek | Chinese. Open source. Competitive quality. Free. |
Why Models Feel Different
Same question, three different answers. OpenAI is designed to make you feel good (warm, conversational). Anthropic is designed to make you think (precise, careful). Each model has its own personality based on training.
Models Change
AI models get version updates (claude-sonnet-4-6 replacing earlier Sonnet generations). Behaviour, formatting, and tone can shift. Companies track versions and test after updates. Benchmarks compare models objectively, but don't tell the full story.
Model Tiers - Thinking / Doing / Checking
| Tier | Example | Use For | Cost |
|---|---|---|---|
| Thinking | claude-opus-4-6 | Complex reasoning, strategy | $$$ |
| Doing | claude-sonnet-4-6 | Everyday tasks (90% of work) | $$ |
| Checking | claude-haiku-4-5 | Quick grammar, formatting | $ |
Same task can cost 10× more on a thinking model vs a checking model.
How NYMG Uses AI
- Started on helpdesk → wasn't reliable
- Moved into technical operations and AI-assisted coding workflows
- A jump to newer Claude generations was the turning point for coding productivity
- Claude Code let Charlotte do tech tasks without coding knowledge
- Dev team shifted: writing code → reviewing AI-written code
- Bart uses ChatGPT for helpdesk; Charlotte & Daniëlle use Claude Code
Passive AI vs Active AI
Passive: You go to it. Ask a question, get an answer. (ChatGPT, Claude)
Active: It can take initiative - use tools, remember past conversations, work through steps independently.
- Have you noticed different results when using different AI tools? What was different?
- Which of the AI tools mentioned (ChatGPT, Claude) are you most curious to try?
- Can you think of tasks in your daily work that could use a 'checking model' vs a 'thinking model'?
- What does 'passive AI' vs 'active AI' mean to you based on what you heard?
Day 2 Complete!
You now know who builds AI models, why different tools exist for different jobs, and how to spot agent washing. Tomorrow we cross the line from passive to active AI.
From Chatbot to Agent
Day 3 is locked
Complete Day 2 flashcards to unlock.
Listen First
Today we cover the most important concept of the week: the difference between a chatbot and an agent. This is where passive AI becomes active AI - and where everything clicks.
What You'll Learn
- Chatbot vs agent
- Tools (the agent's hands)
- Memory & autonomy
- System prompts explained
- Guardrails & safety
- The agent loop
- Workflow automation
- Orchestrators
Bonus: What is an Agent?
A short explainer. The anatomy of an agent in 6 building blocks: model, memory, skills, tools, guardrails, identity.
Study the Visual
The key shift: from a chatbot that answers, to an agent that acts.
- Agentic = able to take action. A chatbot talks, an agent acts. Watch out for agent washing.
- Three pillars: tools (capabilities), memory (short-term + long-term), guided autonomy (freedom within a pipeline)
- Tools are individual actions. Skills are structured instructions that orchestrate multiple tools for complex tasks.
- Guardrails keep agents safe - rules + pipeline checkpoints. Joey's duplicate story proves why.
- 2026 = the agentic shift. Every major company is building agent capabilities. NYMG is part of this.
Chatbot vs Agent
A chatbot is passive AI - you ask, it answers, but it can't take action. An agent is active AI that can use tools, remember past interactions, and work autonomously toward goals.
The Three Pillars of an Agent
| Pillar | What It Means | Example |
|---|---|---|
| Tools | External capabilities to interact with the world | Browse web, edit files, send messages |
| Memory | Stores and recalls past conversations | Remembers your preferences from last week |
| Autonomy | Plans its own steps without instruction | Figures out how to complete a goal independently |
System Prompts
Hidden instructions that define an AI's persona, rules, and boundaries. The user never sees them, but they shape everything. The same AI becomes a completely different tool with different system prompts - that's how Claude becomes Claude Code, or a customer support bot.
Guardrails
Predefined rules and limits that prevent an agent from taking dangerous actions. Without guardrails, autonomy is a liability. With guardrails, it's a superpower.
The Agent Loop
- Observe - take in the current situation
- Think - decide what to do next
- Act - use a tool or take a step
- Check - evaluate the result
- Repeat until the job is done
Workflow Automation & Orchestrators
When multiple specialist agents work together under an orchestrator, you get workflow automation - entire processes running from start to finish. At NYMG, Jørgen's mystery tool from Day 2 is an agent, and the company is moving toward specialised agents coordinated by an orchestrator.
- Think about your daily work: which tasks do you do repeatedly that follow the same steps each time?
- If you could give an AI assistant three tools to help with your job, what would they be?
- What's one task where you'd want to keep human approval (guardrails) and one where you'd trust the AI to just do it?
- How does the chatbot vs agent distinction change how you think about AI in your work?
Day 3 Complete!
You now understand the shift from chatbot to agent. Day 4 is now unlocked - time to look under the hood.
Setup
Listen First
The most NYMG-specific episode. See how Atlas, Joey, and the other agents work behind the scenes, hear real war stories, and understand how everything connects.
What You'll Learn
- Why OpenClaw (local, open-source)
- SOUL file and more - how agent identity works
- The glossary - 15 years of existing institutional knowledge, used by agents automatically
- Meet Atlas - your primary day-to-day agent
- How Atlas uses skills, pipelines, and approval checkpoints
- How glossary knowledge is applied across markets
- Memory, heartbeats, and cron jobs (at a practical level)
- War stories - what went wrong and why
- Anthropomorphism - agents have names, not feelings
Study the Visual
Three visuals: how an agent is built, the overall system, and Atlas's content pipeline.
How is the "system prompt" configured in OpenClaw?
What happens when Atlas needs to translate a post into 17 languages?
What's the key difference between a guardrail-as-instruction and a guardrail-as-code?
📌 Key Takeaways
- Local = control. OpenClaw runs on Charlotte's Mac Mini. Your data stays in the building.
- System prompt = files. SOUL.md (personality), AGENTS.md (rules), USER.md (comms), MEMORY.md (context), TOOLS.md (tool guidance). Not a single text box.
- Atlas is your main agent. It handles content updates, translations, and quality checks for your market. Joey handles social media in a separate workflow.
- Skills auto-load. The right skill activates when the task matches - the agent doesn't need to be told which skill to use.
- Memory persists. pgmemory auto-captures and auto-recalls. Compaction summarises long conversations.
- Heartbeats = proactive. Agents check in without being asked. They work while you sleep.
- Guardrails: instruction vs code. Instructions can be ignored. Code gates physically block bad actions. We're upgrading from instructions to code.
- Things go wrong. Duplicate posts, ignored stops, crashes. That's normal. That's why agents need humans.
OpenClaw - A local, open-source platform running on Charlotte's Mac Mini as a separate user. Multi-agent by design. Data stays on our hardware.
System Prompt in Practice - Not a text box. Workspace files: SOUL.md defines personality, AGENTS.md sets rules, USER.md teaches communication style, MEMORY.md provides bootstrap context, and TOOLS.md explains tool usage.
The Agents - Atlas 🗺️ is the one country managers work with most (content ops, 17 languages, sub-agents per language). Joey 🗽 runs social workflows with Jørgen. Other technical agents exist in the background, but your day-to-day interaction is mainly Atlas.
Skills - Structured instructions that auto-load when tasks match. Atlas has 18 skills. Shared library means consistent processes across agents.
Memory - Short-term = current conversation (context window). Compaction = summarising when it gets long. Long-term = pgmemory (PostgreSQL, auto-capture, auto-recall).
Heartbeat - Proactive scheduled check-ins. Health checks, inbox monitoring. Joey: weekends only. Powered by cron jobs (scheduled tasks).
Pipelines (LangGraph) - Designed step sequences with checkpoints. Joey's posting pipeline. Atlas's translation pipeline. Hard gates physically block bad actions.
Fallback Chains - Automatic model switching when the primary AI provider is down. Ensures agents keep working.
War Stories - Duplicate reel (health monitor killed Joey mid-upload). Ignored STOP command (instruction vs code guardrail). These examples show why process and human review matter.
Glossary - Years of institutional knowledge, now stored in a pgmemory glossary table with 7,283 entries. Agents query what they need instead of loading giant per-language dumps.
- Which agent will you be working with most directly? What does it do?
- What's the difference between how Joey handles social media and how you'd do it manually?
- Why does it matter that NYMG's agents run locally instead of in the cloud?
- If Atlas makes a mistake in your language, what should you do?
🗺️ Meet Atlas - Your Content Agent
Atlas is the agent you'll be working with most. It lives in your Slack channel and handles content updates, translations, and quality checks for your market.
e.g., #atlas-nl for Dutch, #atlas-fi for Finnish, #atlas-de for German
Your First Task
🚧 Your first guided task will be provided in the meeting. Charlotte will walk you through it step by step.
⚠️ What to Do When Things Go Wrong
- Wrong translation? Tell Atlas directly: "That's incorrect. It should be [correct term]."
- Atlas isn't responding? Wait 30 seconds, then try again. If still nothing, message Charlotte.
- Atlas did something you didn't ask? Tell it to stop and revert. Then flag it to Charlotte.
- Not sure if it's right? Ask Atlas to show you what it changed before approving.
Day 4 Complete!
You now know how the agents work under the hood. Day 5 is now unlocked - the practical reality of working with them.
🎁 Bonus: Atlas Content Pipeline
A detailed look at every step Atlas follows when updating your pages.
Along for the Ride
Final lesson
This final day is the operator's guide: how to brief agents clearly, how to verify their work, what they can see, and how your corrections make the system better over time.
What You'll Learn
- Be specific, not clever — avoid assumption and ambiguity
- Trust, but verify — always check the output
- Your feedback trains the system permanently
- What the agent can see — privacy and boundaries
- When things go wrong — practical troubleshooting
- Philosophy: try it, watch carefully, improve every week
Team Reflection
Bring these questions to your next team meeting. There are no right answers — just honest ones.
- What surprised you most across all 5 days?
- What's one thing you'll do differently starting next week?
- Where do you see the biggest opportunity for agents in your specific role?
- What's your biggest remaining concern or question?
You've completed the Agentic Academy!
Five days of learning, from "what is AI?" to working alongside agents every day. The goal now: informed use, better reviews, and clearer workflows.