Intelligent Machines Episode 842: "None Pizza Left Beef - AI On the Edge"
Date: October 23, 2025
Hosts: Paris Martineau, (filling in: Mike Elgin)
Guests: Joey de Villa (AI developer & advocate), Andrew Hawthorne (HP Product Manager, ZGX Nano)
Overview
This episode of Intelligent Machines focuses on the rise of local, edge-based AI—specifically, how large language models (LLMs) can now be efficiently run on compact, powerful personal hardware (with an emphasis on Nvidia’s Blackwell SoC and HP's new ZGX Nano AI station). The hosts discuss the implications for privacy, performance, and the evolving ecosystem of AI devices, touching on how this movement stands in contrast to cloud-based AI dominated by Big Tech. The latter half of the show delves into broader AI news: the launch and ethical concerns around OpenAI's Atlas (agentic browser), trends toward pushback against AI-generated content, ongoing challenges with AI accuracy, and the ever-present line between hype and substance in the field.
Key Discussion Points & Insights
1. Running LLMs Locally: Nvidia Blackwell & HP ZGX Nano
-
Edge AI Motivation
-
The Hardware
- HP ZGX Nano: A mini-form-factor AI station (proprietary, not mini-ITX) based on Nvidia’s Grace Blackwell GB10 SoC.
- Specs: 128GB of unified memory (accessible by both CPU/GPU) handles models up to 200 billion parameters; two units can be clustered for 405B params.
"You could connect two of them together...and get up to 405 billion parameters." —Andrew H., 05:39
- OS & Software: Runs Nvidia’s DGXOS (built on Ubuntu) with a VS Code extension toolkit for cross-platform development (14:10).
-
Local Development Workflows
-
Fine-Tuning vs. RAG:
- Fine-tuning = adding custom expertise to base models for long-term application.
- RAG (retrieval augmented generation) = short-term task-specific data injection; like "an open book test" (09:00).
"Fine tuning is where you go take a course on a particular topic... RAG is you’re about to do a talk and someone gives you some extra notes." —Joey, 09:00
-
Vast Model Marketplace:
- HuggingFace offers 2.1 million+ models, most being fine-tuned versions of foundational LLMs (07:35).
-
-
Performance & Price Market
-
Who’s Buying?
- Enterprises, enthusiasts, small businesses—basically anyone “looking to use AI to accelerate their business” (15:23).
Notable Quotes
- “Prompting is programming… it’s just the next level, a higher-level language” —Joey, 30:50
- “People expect everything digital and connected, and now they’ll expect every device to be AI-based” —Mike, 32:11
2. Tech Deep Dive: Model Architectures & Local-AI Ecosystem
-
Model Scaling & Specialization
- “The theory is, bigger the model, the better the answers... but there is a trend toward specialized models... models that are an expert on say, carpentry or news or protein powders” —Joey, 07:14.
-
Mixture of Experts (MoE) Approach
- LLMs like GPT-5 are structured as “a bunch of little models in a big trench coat,” with a “model picker” deciding which sub-model answers (20:20).
- The “MCP” (Model Context Protocol): lets local LLMs delegate tasks to specialized functions or servers (e.g., “Too Many Cats” demo for household pet ratio advice, 20:53).
Notable Moment
- Too Many Cats Demo:
3. Edge AI as a Societal Trend
-
Privacy & Latency Benefits
- Ideal use cases: home voice assistants, hospital room LLMs (for secure patient conversation), diagnostics equipment (31:15, 34:41).
- “Latency is one problem, intelligence is another... But the dream of having something in your house that could be an assistant... is real. People want that, even though Big Tech hasn't provided it.” —Leo, 33:49
-
Networking, Clustering, and Boosting
4. The Agentic Browser Controversy — OpenAI's Atlas
-
What Is Atlas?
- An “agentic browser”, i.e., integrates ChatGPT deeply into web browsing. Can act as a task-based browsing assistant but also “take over” the browser for actions (41:23, 43:17).
- It can perform clippy-like duties and actual UI manipulation/surfing on your behalf (44:38).
- Built on Chromium, but with strong integration to OpenAI’s platform and data-hungry features (41:45).
-
Skepticism from Hosts
- Cons:
- Sluggish/User Experience: “slower than a user”, “doesn’t add anything positive over existing workflows” (47:08).
- Privacy Concern: Case where the browser accesses shared show documents; “Jeff is gone one week and it’ll never leave” (56:06).
- Pitch?: For most users, value is unclear compared to established tools and workflows (47:46).
- Ethical Risks:
- Vulnerable to indirect prompt injection, e.g., malicious Amazon sellers inserting hidden prompts into product titles (61:04).
- Cons:
Notable Quote
"The idea is that ChatGPT will be your agent. In reality you are ChatGPT’s agent." —Anil Dash, summarized by Leo (101:24)
5. AI Accuracy Crisis and Media Study
-
BBC & EBU Study on LLMs and News Claims:
- 45% of AI answers contained significant issues.
- Gemini (Google) worst performer: nearly 76% responses had serious sourcing/accuracy issues.
- “One in five” (20%) of all answers across models had major accuracy issues (66:51).
-
Panel Reaction:
Notable Quote
- "Don’t let a chatbot do your thinking for you. Don’t let it do your research for you, but use it as a tool." —Mike, 73:31
6. Cultural Pushback Against AI-Generated Content
- Examples:
- Comics and Books: DC Comics pledges “no AI ever”; Michigan libraries ban AI books (citing poor quality).
- E-Commerce: Etsy and Pinterest now label or limit AI-generated items (80:51).
- Reddit/Art: Subreddits and social communities limiting/banning AI-generated content (80:51).
- Publishing Platforms: Amazon and Wikimedia Foundation instituting controls (79:06).
Memorable Moment
"You go to interior design subreddits, and people say ‘what’s this style?’—and it’s just AI slop, fake windows and furniture." —Paris, 81:19
7. The Overhype Problem: The Majority AI View
- Anil Dash’s Critique:
- Within the tech industry, the consensus is that LLMs and generative AI have real utility, but the overhype impedes progress and distracts from genuine value (96:57).
- The panel concurs: both blind hype and blind criticism fail users; measured, rational engagement and realistic assessment are needed (97:31).
Notable Quote
“AI is the most talked about topic in tech. But the opinion most common among those actually building it is rarely heard: it’s overhyped, being forced on everyone, and it’s very difficult to focus on legitimate uses where it might add value.” —Anil Dash, paraphrased (96:57)
8. Limits of LLMs vs. Human Intelligence
- AGI Skepticism:
- Andrej Karpathy (ex-OpenAI): AGI is still “at least 10 years away”, and LLMs are fundamentally different from humans—learning via brute force, lacking real-world grounding (105:22).
- Distinctly Human Factors:
- Human intelligence incorporates emotional, embodied, and experiential factors; LLMs “don’t really want a lollipop” (Gibson, 111:26).
Notable Quotes & Memorable Moments
-
Fine-Tuning vs. RAG Analogy:
"Fine-tuning is where you go take a course on a particular topic; RAG is you’re about to do a talk and someone gives you some extra notes.”
— Joey, 09:00 -
On Agentic Browsers:
“The idea is that ChatGPT will be your agent. In reality you are ChatGPT’s agent.”
— Anil Dash, summarized, 101:24 -
On Flooding the Internet with AI Slop:
“A single person can generate 10 AI books a day. There’s a podcast company with 5,000 shows, each a dollar to make. They sound natural, but they’re just flooding the zone with slop.”
— Mike, 82:26 -
On the Future of Embedded/Edge AI:
“Eventually, everything’s going to have AI in it… Not just big servers, but at the edge, everywhere—including the home.”
—Mike, 32:11
Important Timestamps
| Time | Segment / Topic | |-----------|-----------------------------------------------| | 02:43 | Privacy & edge AI motivation | | 04:17 | Intro to HP ZGX Nano & Blackwell SoC | | 07:35 | Marketplace of fine-tuned LLMs on HuggingFace | | 09:00 | Fine-Tuning vs. RAG analogy | | 20:18 | Mixture-of-Experts (MOE) in GPT-5 | | 21:15 | "Too Many Cats" MCP server example | | 26:48 | Hardware comparison: Mac vs. Blackwell/DGX | | 31:15 | Will local LLMs become common in the home? | | 34:41 | AI for hospital privacy use-case (Moffitt) | | 41:23 | Discussion of OpenAI’s Atlas agentic browser | | 66:51 | BBC/EBU LLM accuracy study | | 79:04 | First cultural bans on AI-generated content | | 96:57 | Anil Dash’s “Majority AI View” | |105:22 | Karpathy: AGI not imminent, LLMs ≠ human | |111:26 | Gibson quote—“LLMs don’t actually want a lollipop”| |134:14 | AI detection tools (Originality.AI) |
Additional References & Resources
-
Joey de Villa:
- Blog: GlobalNerdy.com
- YouTube tutorials (coming soon): YouTube.com/GlobalNerdy
-
Mike Elgin:
- AI newsletter: MachineSociety.ai
- Food/travel adventures: Gastronomad.net
Closing Notes
The episode blends hands-on hardware insights with a healthy skepticism of Big Tech’s latest AI innovations—balancing enthusiasm for real, useful edge AI with deep wariness about hype, privacy, and the flood of low-quality content. Throughout, the discussion maintains a light, conversational tone, peppered with memes (the “None Pizza Left Beef” anecdote) and relatable analogies. This is a must-listen for anyone trying to sort fact from fiction in our rapidly changing intelligent world.
[End of summary]