Ch 1 — What Is AI, Really?

Cutting through the noise — what artificial intelligence actually does, what it doesn’t, and why it matters now
High Level
help
Define
arrow_forward
psychology
How It Works
arrow_forward
category
Types
arrow_forward
trending_up
Why Now
arrow_forward
business
Impact
arrow_forward
rule
Limits
arrow_forward
map
Landscape
-
Click play or press Space to begin...
Step- / 8
psychology
A Working Definition
What AI means in practice, not in science fiction
The Core Idea
Artificial intelligence is software that learns patterns from data and uses those patterns to make predictions or decisions. That’s it. Not consciousness. Not sentience. Not a digital brain. It’s pattern recognition at scale, applied to problems that were previously too complex or too tedious for traditional software to handle.
What Makes It Different from Regular Software
Traditional software follows explicit rules written by programmers: “If X, then Y.” AI systems derive their own rules from data. A spam filter doesn’t have a list of every spam phrase — it learned what spam looks like by analyzing millions of emails. This is the fundamental shift: from programming rules to learning patterns.
IBM’s Definition
IBM defines AI as “technology that enables computers and machines to simulate human learning, comprehension, problem-solving, decision-making, creativity and autonomy.” The key word is simulate. These systems produce outputs that resemble intelligent behavior, but through fundamentally different mechanisms than human thought.
Key insight: When someone says “AI,” they could mean anything from a simple spam filter to GPT-4. The term covers an enormous range. This chapter gives you the framework to distinguish what’s what.
model_training
The Three-Step Process
Data in, patterns learned, predictions out
Step 1: Training
You feed the system large amounts of data — historical transactions, customer records, images, text. The system analyzes this data and identifies statistical patterns. A fraud detection model, for example, learns what distinguishes fraudulent transactions from legitimate ones by studying millions of past examples.
Step 2: The Model
The result of training is a model — a mathematical representation of the patterns found in the data. Think of it as compressed experience. The model doesn’t store every data point; it stores the relationships and patterns that matter for making predictions.
Step 3: Inference
Once trained, the model receives new, unseen data and produces predictions. This is called inference — the model applies what it learned to new situations. When your email client flags a message as spam, or Netflix recommends a show, or a bank declines a suspicious transaction, that’s inference happening in real time.
Why it matters: Training is expensive and happens periodically. Inference is cheap and happens constantly. When executives evaluate AI costs, understanding this distinction is critical — training costs are upfront investments; inference costs are ongoing operational expenses.
category
Narrow AI, General AI, Superintelligence
Only one of these exists today
Narrow AI (What Exists Today)
Every AI system in production today is narrow AI. It excels at a specific task — image recognition, language translation, fraud detection — but cannot generalize beyond its training domain. Siri can answer questions but can’t drive a car. A radiology AI can detect tumors but can’t write a legal brief. Each system is purpose-built.
Artificial General Intelligence (Theoretical)
AGI would match human cognitive abilities across domains — learning any intellectual task a human can, transferring knowledge between unrelated fields, and reasoning about novel situations. Despite headlines, AGI does not exist. There is no scientific consensus on when — or whether — it will be achieved.
Superintelligence (Speculative)
Artificial Superintelligence (ASI) would surpass human intelligence in virtually every domain. This is purely theoretical and is the subject of philosophical debate, not engineering roadmaps. It’s worth understanding the concept, but it should not drive near-term business decisions.
Key insight: When vendors claim “AI” capabilities, they mean narrow AI. The practical question isn’t “is this intelligent?” but “does it solve a specific business problem better than the alternatives?”
trending_up
Why AI Is Happening Now
Three forces converged at the same time
1. Data Explosion
The world generates roughly 120 zettabytes of data per year (as of 2023, per Statista). Digital transactions, social media, IoT sensors, and enterprise systems produce the raw material AI needs. Without massive datasets, even the best algorithms can’t learn meaningful patterns.
2. Compute Power
Modern GPUs can perform trillions of calculations per second. NVIDIA’s A100 GPU, released in 2020, delivers 312 teraflops of compute. Cloud platforms (AWS, Azure, GCP) made this power accessible without massive capital expenditure. Training runs that would have taken years in 2010 now take days.
3. Algorithmic Breakthroughs
Key innovations unlocked new capabilities: deep learning (2012, AlexNet), the transformer architecture (2017, Google), and reinforcement learning from human feedback (2022, OpenAI). These weren’t incremental improvements — each represented a step-change in what AI could do.
The connection: AI research has existed since the 1950s. What changed isn’t the ambition — it’s that data, compute, and algorithms finally reached critical mass simultaneously. Remove any one of the three, and the current AI wave doesn’t happen.
payments
The Scale of the Opportunity
What the numbers actually say
Market Size
Gartner projects global AI spending will reach $2.52 trillion in 2026, a 44% increase from 2025. Of that, $1.37 trillion is infrastructure (GPUs, data centers, cloud), $452 billion is software, and $589 billion is services. This isn’t a niche technology — it’s becoming a foundational layer of the global economy.
Enterprise Adoption
72% of enterprises have deployed AI in at least one business function (McKinsey, 2024). However, Gartner notes AI is currently in the “trough of disillusionment” phase — initial hype is fading, and organizations are now focused on proving measurable ROI rather than running pilots.
The Deployment Gap
Only one-third of AI pilot projects reach scaled deployment. The gap between experimentation and production value is where most organizations struggle. The bottleneck is rarely the technology itself — it’s data quality, organizational readiness, and clear problem definition.
Why it matters: The market is massive, but the winners won’t be those who adopt AI first. They’ll be those who deploy it effectively — on the right problems, with the right data, and with clear success metrics.
block
What AI Cannot Do
Understanding the limits is as important as understanding the capabilities
It Does Not Understand
AI processes statistical patterns. It does not comprehend meaning. A language model can produce a grammatically perfect sentence about grief without any experience of loss. It generates text that is statistically likely to follow the input — not text that reflects genuine understanding. This distinction matters when the stakes are high.
It Hallucinates
AI systems confidently produce false information. A language model will cite papers that don’t exist, invent statistics, or fabricate legal precedents — all with the same confident tone as accurate responses. This isn’t a bug that will be fixed; it’s a structural property of how these systems generate output.
It Reflects Its Training Data
AI systems inherit the biases present in their training data. If historical hiring data reflects gender or racial bias, an AI trained on that data will perpetuate those biases. Amazon famously scrapped an AI recruiting tool in 2018 after discovering it systematically downgraded resumes from women.
Critical for leaders: AI is a powerful tool with real limitations. Deploying it without understanding these constraints — hallucination, bias, lack of true understanding — creates legal, reputational, and operational risk.
map
The AI Landscape at a Glance
The major categories of AI you’ll encounter
Classical Machine Learning
Algorithms like decision trees, random forests, and logistic regression that learn from structured data. Still powers the majority of enterprise AI: fraud detection, credit scoring, demand forecasting, customer segmentation. Reliable, interpretable, and well-understood. Covered in Act II of this course.
Deep Learning
Neural networks with many layers that excel at unstructured data: images, audio, text. Enabled breakthroughs in computer vision (self-driving cars), speech recognition (voice assistants), and natural language processing. Requires more data and compute than classical ML. Covered in Act III.
Generative AI
Models that create new content — text, images, code, video — based on patterns learned from training data. GPT-4, Claude, Gemini, DALL-E, and Midjourney are all generative AI. This is the category driving the current wave of attention and investment. Covered in Act IV.
Agentic AI
AI systems that plan, use tools, and act autonomously. Rather than responding to a single prompt, agents break down complex tasks, call APIs, search databases, write code, and iterate on their own output. This is the emerging frontier. Covered in Act V.
lightbulb
The Mental Model to Carry Forward
A framework for every AI conversation you’ll have
Four Questions to Ask
When evaluating any AI initiative, product, or vendor claim, ask:

1. What specific problem does this solve? — Vague answers (“it uses AI”) are red flags.

2. What data does it need, and do we have it? — AI without quality data is a car without fuel.

3. What are the failure modes? — Every AI system fails. The question is how, and what’s the impact.

4. How do we measure success? — If you can’t define the metric, you can’t prove the value.
The Evolution Stack
This course walks through AI as an evolution, not a revolution:

Classical ML (1990s–2010s) → structured data, specific predictions
Deep Learning (2012–2020) → unstructured data, perception
Generative AI (2020–present) → content creation, language
Agentic AI (2024–present) → autonomous action, tool use

Each layer builds on the previous one. Understanding the stack helps you evaluate where any given AI product sits — and how mature the underlying technology is.
Rule of thumb: The further up the stack, the more powerful the capability — and the less mature the technology. Classical ML is battle-tested. Agentic AI is cutting-edge. Your risk tolerance should adjust accordingly.