<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" encoding="UTF-8" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:sy="http://purl.org/rss/1.0/modules/syndication/" xmlns:admin="http://webns.net/mvcb/" xmlns:atom="http://www.w3.org/2005/Atom/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:fireside="http://fireside.fm/modules/rss/fireside">
  <channel>
    <fireside:hostname>web01.fireside.fm</fireside:hostname>
    <fireside:genDate>Sat, 07 Mar 2026 05:16:22 -0600</fireside:genDate>
    <generator>Fireside (https://fireside.fm)</generator>
    <title>Vanishing Gradients - Episodes Tagged with “Llms”</title>
    <link>https://vanishinggradients.fireside.fm/tags/llms</link>
    <pubDate>Tue, 30 Sep 2025 17:30:00 +1000</pubDate>
    <description>A podcast about all things data, brought to you by data scientist Hugo Bowne-Anderson.
It's time for more critical conversations about the challenges in our industry in order to build better compasses for the solution space! To this end, this podcast will consist of long-format conversations between Hugo and other people who work broadly in the data science, machine learning, and AI spaces. We'll dive deep into all the moving parts of the data world, so if you're new to the space, you'll have an opportunity to learn from the experts. And if you've been around for a while, you'll find out what's happening in many other parts of the data world.
</description>
    <language>en-us</language>
    <itunes:type>episodic</itunes:type>
    <itunes:subtitle>a data podcast with hugo bowne-anderson</itunes:subtitle>
    <itunes:author>Hugo Bowne-Anderson</itunes:author>
    <itunes:summary>A podcast about all things data, brought to you by data scientist Hugo Bowne-Anderson.
It's time for more critical conversations about the challenges in our industry in order to build better compasses for the solution space! To this end, this podcast will consist of long-format conversations between Hugo and other people who work broadly in the data science, machine learning, and AI spaces. We'll dive deep into all the moving parts of the data world, so if you're new to the space, you'll have an opportunity to learn from the experts. And if you've been around for a while, you'll find out what's happening in many other parts of the data world.
</itunes:summary>
    <itunes:image href="https://media24.fireside.fm/file/fireside-images-2024/podcasts/images/1/140c3904-8258-4c39-a698-a112b7077bd7/cover.jpg?v=1"/>
    <itunes:explicit>no</itunes:explicit>
    <itunes:keywords>data science, machine learning, AI</itunes:keywords>
    <itunes:owner>
      <itunes:name>Hugo Bowne-Anderson</itunes:name>
      <itunes:email>hugobowne@hey.com</itunes:email>
    </itunes:owner>
<itunes:category text="Technology"/>
<item>
  <title>Episode 60: 10 Things I Hate About AI Evals with Hamel Husain</title>
  <link>https://vanishinggradients.fireside.fm/60</link>
  <guid isPermaLink="false">0fbc2a65-3bfc-4f8a-83ac-d370f1a30e13</guid>
  <pubDate>Tue, 30 Sep 2025 17:30:00 +1000</pubDate>
  <author>Hugo Bowne-Anderson</author>
  <enclosure url="https://aphid.fireside.fm/d/1437767933/140c3904-8258-4c39-a698-a112b7077bd7/0fbc2a65-3bfc-4f8a-83ac-d370f1a30e13.mp3" length="105505355" type="audio/mpeg"/>
  <itunes:episodeType>full</itunes:episodeType>
  <itunes:author>Hugo Bowne-Anderson</itunes:author>
  <itunes:subtitle>Most AI teams find "evals" frustrating, but ML Engineer Hamel Husain argues they’re just using the wrong playbook. In this episode, he lays out a data-centric approach to systematically measure and improve AI, turning unreliable prototypes into robust, production-ready systems.
</itunes:subtitle>
  <itunes:duration>1:13:15</itunes:duration>
  <itunes:explicit>no</itunes:explicit>
  <itunes:image href="https://media24.fireside.fm/file/fireside-images-2024/podcasts/images/1/140c3904-8258-4c39-a698-a112b7077bd7/cover.jpg?v=1"/>
  <description>Most AI teams find "evals" frustrating, but ML Engineer Hamel Husain argues they’re just using the wrong playbook. In this episode, he lays out a data-centric approach to systematically measure and improve AI, turning unreliable prototypes into robust, production-ready systems.
Drawing from his experience getting countless teams unstuck, Hamel explains why the solution requires a "revenge of the data scientists." He details the essential mindset shifts, error analysis techniques, and practical steps needed to move beyond guesswork and build AI products you can actually trust.
We talk through:
  The 10(+1) critical mistakes that cause teams to waste time on evals
  Why "hallucination scores" are a waste of time (and what to measure instead)
  The manual review process that finds major issues in hours, not weeks
  A step-by-step method for building LLM judges you can actually trust
  How to use domain experts without getting stuck in endless review committees
  Guest Bryan Bischof's "Failure as a Funnel" for debugging complex AI agents
If you're tired of ambiguous "vibe checks" and want a clear process that delivers real improvement, this episode provides the definitive roadmap.
LINKS
Hamel's website and blog (https://hamel.dev/)
Hugo speaks with Philip Carter (Honeycomb) about aligning your LLM-as-a-judge with your domain expertise (https://vanishinggradients.fireside.fm/51)
Hamel Husain on Lenny's pocast, which includes a live demo of error analysis (https://www.lennysnewsletter.com/p/why-ai-evals-are-the-hottest-new-skill)
The episode of VG in which Hamel and Hugo talk about Hamel's "data consulting in Vegas" era (https://vanishinggradients.fireside.fm/9)
Upcoming Events on Luma (https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk)
Watch the podcast video on YouTube (https://youtube.com/live/QEk-XwrkqhI?feature=share)
Hamel's AI evals course, which he teaches with Shreya Shankar (UC Berkeley): starts Oct 6 and this link gives 35% off! (https://maven.com/parlance-labs/evals?promoCode=GOHUGORGOHOME) https://maven.com/parlance-labs/evals?promoCode=GOHUGORGOHOME
🎓 Learn more:
Hugo's course: Building LLM Applications for Data Scientists and Software Engineers (https://maven.com/s/course/d56067f338) — https://maven.com/s/course/d56067f338  
</description>
  <itunes:keywords>AI, GenAI, LLMs, data science, machine learning, evals</itunes:keywords>
  <content:encoded>
    <![CDATA[<p>Most AI teams find &quot;evals&quot; frustrating, but ML Engineer Hamel Husain argues they’re just using the wrong playbook. In this episode, he lays out a data-centric approach to systematically measure and improve AI, turning unreliable prototypes into robust, production-ready systems.</p>

<p>Drawing from his experience getting countless teams unstuck, Hamel explains why the solution requires a &quot;revenge of the data scientists.&quot; He details the essential mindset shifts, error analysis techniques, and practical steps needed to move beyond guesswork and build AI products you can actually trust.</p>

<p>We talk through:</p>

<ul>
<li>  The 10(+1) critical mistakes that cause teams to waste time on evals</li>
<li>  Why &quot;hallucination scores&quot; are a waste of time (and what to measure instead)</li>
<li>  The manual review process that finds major issues in hours, not weeks</li>
<li>  A step-by-step method for building LLM judges you can actually trust</li>
<li>  How to use domain experts without getting stuck in endless review committees</li>
<li>  Guest Bryan Bischof&#39;s &quot;Failure as a Funnel&quot; for debugging complex AI agents</li>
</ul>

<p>If you&#39;re tired of ambiguous &quot;vibe checks&quot; and want a clear process that delivers real improvement, this episode provides the definitive roadmap.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://hamel.dev/" rel="nofollow">Hamel&#39;s website and blog</a></li>
<li><a href="https://vanishinggradients.fireside.fm/51" rel="nofollow">Hugo speaks with Philip Carter (Honeycomb) about aligning your LLM-as-a-judge with your domain expertise</a></li>
<li><a href="https://www.lennysnewsletter.com/p/why-ai-evals-are-the-hottest-new-skill" rel="nofollow">Hamel Husain on Lenny&#39;s pocast, which includes a live demo of error analysis</a></li>
<li><a href="https://vanishinggradients.fireside.fm/9" rel="nofollow">The episode of VG in which Hamel and Hugo talk about Hamel&#39;s &quot;data consulting in Vegas&quot; era</a></li>
<li><a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">Upcoming Events on Luma</a></li>
<li><a href="https://youtube.com/live/QEk-XwrkqhI?feature=share" rel="nofollow">Watch the podcast video on YouTube</a></li>
<li><a href="https://maven.com/parlance-labs/evals?promoCode=GOHUGORGOHOME" rel="nofollow">Hamel&#39;s AI evals course, which he teaches with Shreya Shankar (UC Berkeley): starts Oct 6 and this link gives 35% off!</a> <a href="https://maven.com/parlance-labs/evals?promoCode=GOHUGORGOHOME" rel="nofollow">https://maven.com/parlance-labs/evals?promoCode=GOHUGORGOHOME</a></li>
</ul>

<p>🎓 Learn more:</p>

<ul>
<li><strong>Hugo&#39;s course:</strong> <a href="https://maven.com/s/course/d56067f338" rel="nofollow">Building LLM Applications for Data Scientists and Software Engineers</a> — <a href="https://maven.com/s/course/d56067f338" rel="nofollow">https://maven.com/s/course/d56067f338</a> </li>
</ul>]]>
  </content:encoded>
  <itunes:summary>
    <![CDATA[<p>Most AI teams find &quot;evals&quot; frustrating, but ML Engineer Hamel Husain argues they’re just using the wrong playbook. In this episode, he lays out a data-centric approach to systematically measure and improve AI, turning unreliable prototypes into robust, production-ready systems.</p>

<p>Drawing from his experience getting countless teams unstuck, Hamel explains why the solution requires a &quot;revenge of the data scientists.&quot; He details the essential mindset shifts, error analysis techniques, and practical steps needed to move beyond guesswork and build AI products you can actually trust.</p>

<p>We talk through:</p>

<ul>
<li>  The 10(+1) critical mistakes that cause teams to waste time on evals</li>
<li>  Why &quot;hallucination scores&quot; are a waste of time (and what to measure instead)</li>
<li>  The manual review process that finds major issues in hours, not weeks</li>
<li>  A step-by-step method for building LLM judges you can actually trust</li>
<li>  How to use domain experts without getting stuck in endless review committees</li>
<li>  Guest Bryan Bischof&#39;s &quot;Failure as a Funnel&quot; for debugging complex AI agents</li>
</ul>

<p>If you&#39;re tired of ambiguous &quot;vibe checks&quot; and want a clear process that delivers real improvement, this episode provides the definitive roadmap.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://hamel.dev/" rel="nofollow">Hamel&#39;s website and blog</a></li>
<li><a href="https://vanishinggradients.fireside.fm/51" rel="nofollow">Hugo speaks with Philip Carter (Honeycomb) about aligning your LLM-as-a-judge with your domain expertise</a></li>
<li><a href="https://www.lennysnewsletter.com/p/why-ai-evals-are-the-hottest-new-skill" rel="nofollow">Hamel Husain on Lenny&#39;s pocast, which includes a live demo of error analysis</a></li>
<li><a href="https://vanishinggradients.fireside.fm/9" rel="nofollow">The episode of VG in which Hamel and Hugo talk about Hamel&#39;s &quot;data consulting in Vegas&quot; era</a></li>
<li><a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">Upcoming Events on Luma</a></li>
<li><a href="https://youtube.com/live/QEk-XwrkqhI?feature=share" rel="nofollow">Watch the podcast video on YouTube</a></li>
<li><a href="https://maven.com/parlance-labs/evals?promoCode=GOHUGORGOHOME" rel="nofollow">Hamel&#39;s AI evals course, which he teaches with Shreya Shankar (UC Berkeley): starts Oct 6 and this link gives 35% off!</a> <a href="https://maven.com/parlance-labs/evals?promoCode=GOHUGORGOHOME" rel="nofollow">https://maven.com/parlance-labs/evals?promoCode=GOHUGORGOHOME</a></li>
</ul>

<p>🎓 Learn more:</p>

<ul>
<li><strong>Hugo&#39;s course:</strong> <a href="https://maven.com/s/course/d56067f338" rel="nofollow">Building LLM Applications for Data Scientists and Software Engineers</a> — <a href="https://maven.com/s/course/d56067f338" rel="nofollow">https://maven.com/s/course/d56067f338</a> </li>
</ul>]]>
  </itunes:summary>
</item>
<item>
  <title>Episode 57: AI Agents and LLM Judges at Scale: Processing Millions of Documents (Without Breaking the Bank)</title>
  <link>https://vanishinggradients.fireside.fm/57</link>
  <guid isPermaLink="false">60db26a1-cad5-4c3d-9661-bbc51a3a0b27</guid>
  <pubDate>Fri, 29 Aug 2025 21:00:00 +1000</pubDate>
  <author>Hugo Bowne-Anderson</author>
  <enclosure url="https://aphid.fireside.fm/d/1437767933/140c3904-8258-4c39-a698-a112b7077bd7/60db26a1-cad5-4c3d-9661-bbc51a3a0b27.mp3" length="81037068" type="audio/mpeg"/>
  <itunes:episodeType>full</itunes:episodeType>
  <itunes:author>Hugo Bowne-Anderson</itunes:author>
  <itunes:subtitle>While many people talk about “agents,” **Shreya Shankar** (UC Berkeley) has been building the systems that make them reliable. In this episode, she shares how AI agents and LLM judges can be used to process millions of documents accurately and cheaply.  

Drawing from work on projects ranging from databases of police misconduct reports to large-scale customer transcripts, Shreya explains the frameworks, error analysis, and guardrails needed to turn flaky LLM outputs into trustworthy pipelines</itunes:subtitle>
  <itunes:duration>41:27</itunes:duration>
  <itunes:explicit>no</itunes:explicit>
  <itunes:image href="https://media24.fireside.fm/file/fireside-images-2024/podcasts/images/1/140c3904-8258-4c39-a698-a112b7077bd7/cover.jpg?v=1"/>
  <description>While many people talk about “agents,” Shreya Shankar (UC Berkeley) has been building the systems that make them reliable. In this episode, she shares how AI agents and LLM judges can be used to process millions of documents accurately and cheaply.  
Drawing from work on projects ranging from databases of police misconduct reports to large-scale customer transcripts, Shreya explains the frameworks, error analysis, and guardrails needed to turn flaky LLM outputs into trustworthy pipelines.  
We talk through:  
- Treating LLM workflows as ETL pipelines for unstructured text  
- Error analysis: why you need humans reviewing the first 50–100 traces  
- Guardrails like retries, validators, and “gleaning”  
- How LLM judges work — rubrics, pairwise comparisons, and cost trade-offs  
- Cheap vs. expensive models: when to swap for savings  
- Where agents fit in (and where they don’t)  
If you’ve ever wondered how to move beyond unreliable demos, this episode shows how to scale LLMs to millions of documents — without breaking the bank.
LINKS
Shreya's website (https://www.sh-reya.com/)
DocETL, A system for LLM-powered data processing (https://www.docetl.org/)
Upcoming Events on Luma (https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk)
Watch the podcast video on YouTube (https://youtu.be/3r_Hsjy85nk)
Shreya's AI evals course, which she teaches with Hamel "Evals" Husain (https://maven.com/parlance-labs/evals?promoCode=GOHUGORGOHOME)
🎓 Learn more:
Hugo's course: Building LLM Applications for Data Scientists and Software Engineers (https://maven.com/s/course/d56067f338) — https://maven.com/s/course/d56067f338 
</description>
  <itunes:keywords>LLMs, Agents, RAG, Machine Learning</itunes:keywords>
  <content:encoded>
    <![CDATA[<p>While many people talk about “agents,” <strong>Shreya Shankar</strong> (UC Berkeley) has been building the systems that make them reliable. In this episode, she shares how AI agents and LLM judges can be used to process millions of documents accurately and cheaply.  </p>

<p>Drawing from work on projects ranging from databases of police misconduct reports to large-scale customer transcripts, Shreya explains the frameworks, error analysis, and guardrails needed to turn flaky LLM outputs into trustworthy pipelines.  </p>

<p><strong>We talk through:</strong>  </p>

<ul>
<li>Treating LLM workflows as ETL pipelines for unstructured text<br></li>
<li>Error analysis: why you need humans reviewing the first 50–100 traces<br></li>
<li>Guardrails like retries, validators, and “gleaning”<br></li>
<li>How LLM judges work — rubrics, pairwise comparisons, and cost trade-offs<br></li>
<li>Cheap vs. expensive models: when to swap for savings<br></li>
<li>Where agents fit in (and where they don’t)<br></li>
</ul>

<p>If you’ve ever wondered how to move beyond unreliable demos, this episode shows how to scale LLMs to millions of documents — without breaking the bank.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://www.sh-reya.com/" rel="nofollow">Shreya&#39;s website</a></li>
<li><a href="https://www.docetl.org/" rel="nofollow">DocETL, A system for LLM-powered data processing</a></li>
<li><a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">Upcoming Events on Luma</a></li>
<li><a href="https://youtu.be/3r_Hsjy85nk" rel="nofollow">Watch the podcast video on YouTube</a></li>
<li><a href="https://maven.com/parlance-labs/evals?promoCode=GOHUGORGOHOME" rel="nofollow">Shreya&#39;s AI evals course, which she teaches with Hamel &quot;Evals&quot; Husain</a></li>
</ul>

<p>🎓 Learn more:</p>

<ul>
<li><strong>Hugo&#39;s course:</strong> <a href="https://maven.com/s/course/d56067f338" rel="nofollow">Building LLM Applications for Data Scientists and Software Engineers</a> — <a href="https://maven.com/s/course/d56067f338" rel="nofollow">https://maven.com/s/course/d56067f338</a> </li>
</ul>]]>
  </content:encoded>
  <itunes:summary>
    <![CDATA[<p>While many people talk about “agents,” <strong>Shreya Shankar</strong> (UC Berkeley) has been building the systems that make them reliable. In this episode, she shares how AI agents and LLM judges can be used to process millions of documents accurately and cheaply.  </p>

<p>Drawing from work on projects ranging from databases of police misconduct reports to large-scale customer transcripts, Shreya explains the frameworks, error analysis, and guardrails needed to turn flaky LLM outputs into trustworthy pipelines.  </p>

<p><strong>We talk through:</strong>  </p>

<ul>
<li>Treating LLM workflows as ETL pipelines for unstructured text<br></li>
<li>Error analysis: why you need humans reviewing the first 50–100 traces<br></li>
<li>Guardrails like retries, validators, and “gleaning”<br></li>
<li>How LLM judges work — rubrics, pairwise comparisons, and cost trade-offs<br></li>
<li>Cheap vs. expensive models: when to swap for savings<br></li>
<li>Where agents fit in (and where they don’t)<br></li>
</ul>

<p>If you’ve ever wondered how to move beyond unreliable demos, this episode shows how to scale LLMs to millions of documents — without breaking the bank.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://www.sh-reya.com/" rel="nofollow">Shreya&#39;s website</a></li>
<li><a href="https://www.docetl.org/" rel="nofollow">DocETL, A system for LLM-powered data processing</a></li>
<li><a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">Upcoming Events on Luma</a></li>
<li><a href="https://youtu.be/3r_Hsjy85nk" rel="nofollow">Watch the podcast video on YouTube</a></li>
<li><a href="https://maven.com/parlance-labs/evals?promoCode=GOHUGORGOHOME" rel="nofollow">Shreya&#39;s AI evals course, which she teaches with Hamel &quot;Evals&quot; Husain</a></li>
</ul>

<p>🎓 Learn more:</p>

<ul>
<li><strong>Hugo&#39;s course:</strong> <a href="https://maven.com/s/course/d56067f338" rel="nofollow">Building LLM Applications for Data Scientists and Software Engineers</a> — <a href="https://maven.com/s/course/d56067f338" rel="nofollow">https://maven.com/s/course/d56067f338</a> </li>
</ul>]]>
  </itunes:summary>
</item>
<item>
  <title>Episode 56: DeepMind Just Dropped Gemma 270M... And Here’s Why It Matters</title>
  <link>https://vanishinggradients.fireside.fm/56</link>
  <guid isPermaLink="false">4f0a10fa-b411-458f-91b4-b68784e2d557</guid>
  <pubDate>Fri, 15 Aug 2025 02:00:00 +1000</pubDate>
  <author>Hugo Bowne-Anderson</author>
  <enclosure url="https://aphid.fireside.fm/d/1437767933/140c3904-8258-4c39-a698-a112b7077bd7/4f0a10fa-b411-458f-91b4-b68784e2d557.mp3" length="88573733" type="audio/mpeg"/>
  <itunes:episodeType>full</itunes:episodeType>
  <itunes:author>Hugo Bowne-Anderson</itunes:author>
  <itunes:subtitle>While much of the AI world chases ever-larger models, Ravin Kumar (Google DeepMind) and his team build across the size spectrum, from billions of parameters down to this week’s release: **Gemma 270M**, the smallest member yet of the Gemma 3 open-weight family. At just 270 million parameters, a quarter the size of Gemma 1B, it’s designed for speed, efficiency, and fine-tuning.  

We explore what makes 270M special, where it fits alongside its billion-parameter siblings, and why you might reach for it in production even if you think “small” means “just for experiments.”  </itunes:subtitle>
  <itunes:duration>45:40</itunes:duration>
  <itunes:explicit>no</itunes:explicit>
  <itunes:image href="https://media24.fireside.fm/file/fireside-images-2024/podcasts/images/1/140c3904-8258-4c39-a698-a112b7077bd7/cover.jpg?v=1"/>
  <description>While much of the AI world chases ever-larger models, Ravin Kumar (Google DeepMind) and his team build across the size spectrum, from billions of parameters down to this week’s release: Gemma 270M, the smallest member yet of the Gemma 3 open-weight family. At just 270 million parameters, a quarter the size of Gemma 1B,  it’s designed for speed, efficiency, and fine-tuning.  
We explore what makes 270M special, where it fits alongside its billion-parameter siblings, and why you might reach for it in production even if you think “small” means “just for experiments.”  
We talk through:  
- Where 270M fits into the Gemma 3 lineup — and why it exists  
- On-device use cases where latency, privacy, and efficiency matter  
- How smaller models open up rapid, targeted fine-tuning  
- Running multiple models in parallel without heavyweight hardware  
- Why “small” models might drive the next big wave of AI adoption  
If you’ve ever wondered what you’d do with a model this size (or how to squeeze the most out of it) this episode will show you how small can punch far above its weight.
LINKS
Introducing Gemma 3 270M: The compact model for hyper-efficient AI (Google Developer Blog) (https://developers.googleblog.com/en/introducing-gemma-3-270m/)
Full Model Fine-Tune Guide using Hugging Face Transformers (https://ai.google.dev/gemma/docs/core/huggingface_text_full_finetune)
The Gemma 270M model on HuggingFace (https://huggingface.co/google/gemma-3-270m)
The Gemma 270M model on Ollama (https://ollama.com/library/gemma3:270m)
Building AI Agents with Gemma 3, a workshop with Ravin and Hugo (https://www.youtube.com/live/-IWstEStqok) (Code here (https://github.com/canyon289/ai_agent_basics))
From Images to Agents: Building and Evaluating Multimodal AI Workflows, a workshop with Ravin and Hugo (https://www.youtube.com/live/FNlM7lSt8Uk)(Code here (https://github.com/canyon289/ai_image_agent))
Evaluating AI Agents: From Demos to Dependability, an upcoming workshop with Ravin and Hugo (https://lu.ma/ezgny3dl)
Upcoming Events on Luma (https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk)
Watch the podcast video on YouTube (https://youtu.be/VZDw6C2A_8E)
🎓 Learn more:
Hugo's course: Building LLM Applications for Data Scientists and Software Engineers (https://maven.com/s/course/d56067f338) — https://maven.com/s/course/d56067f338 ($600 off early bird discount for November cohort availiable until August 16) 
</description>
  <itunes:keywords>LLMs, GenAI, machine learning</itunes:keywords>
  <content:encoded>
    <![CDATA[<p>While much of the AI world chases ever-larger models, Ravin Kumar (Google DeepMind) and his team build across the size spectrum, from billions of parameters down to this week’s release: <strong>Gemma 270M</strong>, the smallest member yet of the Gemma 3 open-weight family. At just 270 million parameters, a quarter the size of Gemma 1B,  it’s designed for speed, efficiency, and fine-tuning.  </p>

<p>We explore what makes 270M special, where it fits alongside its billion-parameter siblings, and why you might reach for it in production even if you think “small” means “just for experiments.”  </p>

<p><strong>We talk through:</strong>  </p>

<ul>
<li>Where 270M fits into the Gemma 3 lineup — and why it exists<br></li>
<li>On-device use cases where latency, privacy, and efficiency matter<br></li>
<li>How smaller models open up rapid, targeted fine-tuning<br></li>
<li>Running multiple models in parallel without heavyweight hardware<br></li>
<li>Why “small” models might drive the next big wave of AI adoption<br></li>
</ul>

<p>If you’ve ever wondered what you’d do with a model this size (or how to squeeze the most out of it) this episode will show you how small can punch far above its weight.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://developers.googleblog.com/en/introducing-gemma-3-270m/" rel="nofollow">Introducing Gemma 3 270M: The compact model for hyper-efficient AI (Google Developer Blog)</a></li>
<li><a href="https://ai.google.dev/gemma/docs/core/huggingface_text_full_finetune" rel="nofollow">Full Model Fine-Tune Guide using Hugging Face Transformers</a></li>
<li><a href="https://huggingface.co/google/gemma-3-270m" rel="nofollow">The Gemma 270M model on HuggingFace</a></li>
<li><a href="https://ollama.com/library/gemma3:270m" rel="nofollow">The Gemma 270M model on Ollama</a></li>
<li><a href="https://www.youtube.com/live/-IWstEStqok" rel="nofollow">Building AI Agents with Gemma 3, a workshop with Ravin and Hugo</a> (<a href="https://github.com/canyon289/ai_agent_basics" rel="nofollow">Code here</a>)</li>
<li><a href="https://www.youtube.com/live/FNlM7lSt8Uk" rel="nofollow">From Images to Agents: Building and Evaluating Multimodal AI Workflows, a workshop with Ravin and Hugo</a>(<a href="https://github.com/canyon289/ai_image_agent" rel="nofollow">Code here</a>)</li>
<li><a href="https://lu.ma/ezgny3dl" rel="nofollow">Evaluating AI Agents: From Demos to Dependability, an upcoming workshop with Ravin and Hugo</a></li>
<li><a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">Upcoming Events on Luma</a></li>
<li><a href="https://youtu.be/VZDw6C2A_8E" rel="nofollow">Watch the podcast video on YouTube</a></li>
</ul>

<p>🎓 Learn more:</p>

<ul>
<li><strong>Hugo&#39;s course:</strong> <a href="https://maven.com/s/course/d56067f338" rel="nofollow">Building LLM Applications for Data Scientists and Software Engineers</a> — <a href="https://maven.com/s/course/d56067f338" rel="nofollow">https://maven.com/s/course/d56067f338</a> ($600 off early bird discount for November cohort availiable until August 16)</li>
</ul>]]>
  </content:encoded>
  <itunes:summary>
    <![CDATA[<p>While much of the AI world chases ever-larger models, Ravin Kumar (Google DeepMind) and his team build across the size spectrum, from billions of parameters down to this week’s release: <strong>Gemma 270M</strong>, the smallest member yet of the Gemma 3 open-weight family. At just 270 million parameters, a quarter the size of Gemma 1B,  it’s designed for speed, efficiency, and fine-tuning.  </p>

<p>We explore what makes 270M special, where it fits alongside its billion-parameter siblings, and why you might reach for it in production even if you think “small” means “just for experiments.”  </p>

<p><strong>We talk through:</strong>  </p>

<ul>
<li>Where 270M fits into the Gemma 3 lineup — and why it exists<br></li>
<li>On-device use cases where latency, privacy, and efficiency matter<br></li>
<li>How smaller models open up rapid, targeted fine-tuning<br></li>
<li>Running multiple models in parallel without heavyweight hardware<br></li>
<li>Why “small” models might drive the next big wave of AI adoption<br></li>
</ul>

<p>If you’ve ever wondered what you’d do with a model this size (or how to squeeze the most out of it) this episode will show you how small can punch far above its weight.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://developers.googleblog.com/en/introducing-gemma-3-270m/" rel="nofollow">Introducing Gemma 3 270M: The compact model for hyper-efficient AI (Google Developer Blog)</a></li>
<li><a href="https://ai.google.dev/gemma/docs/core/huggingface_text_full_finetune" rel="nofollow">Full Model Fine-Tune Guide using Hugging Face Transformers</a></li>
<li><a href="https://huggingface.co/google/gemma-3-270m" rel="nofollow">The Gemma 270M model on HuggingFace</a></li>
<li><a href="https://ollama.com/library/gemma3:270m" rel="nofollow">The Gemma 270M model on Ollama</a></li>
<li><a href="https://www.youtube.com/live/-IWstEStqok" rel="nofollow">Building AI Agents with Gemma 3, a workshop with Ravin and Hugo</a> (<a href="https://github.com/canyon289/ai_agent_basics" rel="nofollow">Code here</a>)</li>
<li><a href="https://www.youtube.com/live/FNlM7lSt8Uk" rel="nofollow">From Images to Agents: Building and Evaluating Multimodal AI Workflows, a workshop with Ravin and Hugo</a>(<a href="https://github.com/canyon289/ai_image_agent" rel="nofollow">Code here</a>)</li>
<li><a href="https://lu.ma/ezgny3dl" rel="nofollow">Evaluating AI Agents: From Demos to Dependability, an upcoming workshop with Ravin and Hugo</a></li>
<li><a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">Upcoming Events on Luma</a></li>
<li><a href="https://youtu.be/VZDw6C2A_8E" rel="nofollow">Watch the podcast video on YouTube</a></li>
</ul>

<p>🎓 Learn more:</p>

<ul>
<li><strong>Hugo&#39;s course:</strong> <a href="https://maven.com/s/course/d56067f338" rel="nofollow">Building LLM Applications for Data Scientists and Software Engineers</a> — <a href="https://maven.com/s/course/d56067f338" rel="nofollow">https://maven.com/s/course/d56067f338</a> ($600 off early bird discount for November cohort availiable until August 16)</li>
</ul>]]>
  </itunes:summary>
</item>
<item>
  <title>Episode 52: Why Most LLM Products Break at Retrieval (And How to Fix Them) </title>
  <link>https://vanishinggradients.fireside.fm/52</link>
  <guid isPermaLink="false">258dd611-e817-4971-a655-f07343b967e4</guid>
  <pubDate>Thu, 03 Jul 2025 02:00:00 +1000</pubDate>
  <author>Hugo Bowne-Anderson</author>
  <enclosure url="https://aphid.fireside.fm/d/1437767933/140c3904-8258-4c39-a698-a112b7077bd7/258dd611-e817-4971-a655-f07343b967e4.mp3" length="27489267" type="audio/mpeg"/>
  <itunes:episodeType>full</itunes:episodeType>
  <itunes:season>1</itunes:season>
  <itunes:author>Hugo Bowne-Anderson</itunes:author>
  <itunes:subtitle>Most LLM-powered features do not break at the model. They break at the context. So how do you retrieve the right information to get useful results, even under vague or messy user queries?

In this episode, we hear from Eric Ma, who leads data science research in the Data Science and AI group at Moderna. He shares what it takes to move beyond toy demos and ship LLM features that actually help people do their jobs.</itunes:subtitle>
  <itunes:duration>28:38</itunes:duration>
  <itunes:explicit>no</itunes:explicit>
  <itunes:image href="https://media24.fireside.fm/file/fireside-images-2024/podcasts/images/1/140c3904-8258-4c39-a698-a112b7077bd7/cover.jpg?v=1"/>
  <description>Most LLM-powered features do not break at the model. They break at the context. So how do you retrieve the right information to get useful results, even under vague or messy user queries?
In this episode, we hear from Eric Ma, who leads data science research in the Data Science and AI group at Moderna. He shares what it takes to move beyond toy demos and ship LLM features that actually help people do their jobs.
We cover:
• How to align retrieval with user intent and why cosine similarity is not the answer
• How a dumb YAML-based system outperformed so-called smart retrieval pipelines
• Why vague queries like “what is this all about” expose real weaknesses in most systems
• When vibe checks are enough and when formal evaluation is worth the effort
• How retrieval workflows can evolve alongside your product and user needs
If you are building LLM-powered systems and care about how they work, not just whether they work, this one is for you.
LINKS
Eric's website (https://ericmjl.github.io/)
Upcoming Events on Luma (https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk)
Hugo's recent newsletter about upcoming events and more! (https://hugobowne.substack.com/p/stop-building-agents)
🎓 Learn more:
Hugo's course: Building LLM Applications for Data Scientists and Software Engineers (https://maven.com/s/course/d56067f338) — next cohort starts July 8: https://maven.com/s/course/d56067f338
📺 Watch the video version on YouTube: YouTube link (https://youtu.be/d-FaR5Ywd5k)
</description>
  <itunes:keywords>data science, machine learning, AI, LLMs, GenAI</itunes:keywords>
  <content:encoded>
    <![CDATA[<p>Most LLM-powered features do not break at the model. They break at the context. So how do you retrieve the right information to get useful results, even under vague or messy user queries?</p>

<p>In this episode, we hear from Eric Ma, who leads data science research in the Data Science and AI group at Moderna. He shares what it takes to move beyond toy demos and ship LLM features that actually help people do their jobs.</p>

<p>We cover:<br>
• How to align retrieval with user intent and why cosine similarity is not the answer<br>
• How a dumb YAML-based system outperformed so-called smart retrieval pipelines<br>
• Why vague queries like “what is this all about” expose real weaknesses in most systems<br>
• When vibe checks are enough and when formal evaluation is worth the effort<br>
• How retrieval workflows can evolve alongside your product and user needs</p>

<p>If you are building LLM-powered systems and care about how they work, not just whether they work, this one is for you.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://ericmjl.github.io/" rel="nofollow">Eric&#39;s website</a></li>
<li><a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">Upcoming Events on Luma</a></li>
<li><a href="https://hugobowne.substack.com/p/stop-building-agents" rel="nofollow">Hugo&#39;s recent newsletter about upcoming events and more!</a></li>
</ul>

<p>🎓 Learn more:</p>

<ul>
<li><strong>Hugo&#39;s course:</strong> <a href="https://maven.com/s/course/d56067f338" rel="nofollow">Building LLM Applications for Data Scientists and Software Engineers</a> — next cohort starts July 8: <a href="https://maven.com/s/course/d56067f338" rel="nofollow">https://maven.com/s/course/d56067f338</a></li>
</ul>

<p>📺 <strong>Watch the video version on YouTube:</strong> <a href="https://youtu.be/d-FaR5Ywd5k" rel="nofollow">YouTube link</a></p>]]>
  </content:encoded>
  <itunes:summary>
    <![CDATA[<p>Most LLM-powered features do not break at the model. They break at the context. So how do you retrieve the right information to get useful results, even under vague or messy user queries?</p>

<p>In this episode, we hear from Eric Ma, who leads data science research in the Data Science and AI group at Moderna. He shares what it takes to move beyond toy demos and ship LLM features that actually help people do their jobs.</p>

<p>We cover:<br>
• How to align retrieval with user intent and why cosine similarity is not the answer<br>
• How a dumb YAML-based system outperformed so-called smart retrieval pipelines<br>
• Why vague queries like “what is this all about” expose real weaknesses in most systems<br>
• When vibe checks are enough and when formal evaluation is worth the effort<br>
• How retrieval workflows can evolve alongside your product and user needs</p>

<p>If you are building LLM-powered systems and care about how they work, not just whether they work, this one is for you.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://ericmjl.github.io/" rel="nofollow">Eric&#39;s website</a></li>
<li><a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">Upcoming Events on Luma</a></li>
<li><a href="https://hugobowne.substack.com/p/stop-building-agents" rel="nofollow">Hugo&#39;s recent newsletter about upcoming events and more!</a></li>
</ul>

<p>🎓 Learn more:</p>

<ul>
<li><strong>Hugo&#39;s course:</strong> <a href="https://maven.com/s/course/d56067f338" rel="nofollow">Building LLM Applications for Data Scientists and Software Engineers</a> — next cohort starts July 8: <a href="https://maven.com/s/course/d56067f338" rel="nofollow">https://maven.com/s/course/d56067f338</a></li>
</ul>

<p>📺 <strong>Watch the video version on YouTube:</strong> <a href="https://youtu.be/d-FaR5Ywd5k" rel="nofollow">YouTube link</a></p>]]>
  </itunes:summary>
</item>
<item>
  <title>Episode 51: Why We Built an MCP Server and What Broke First</title>
  <link>https://vanishinggradients.fireside.fm/51</link>
  <guid isPermaLink="false">c45cdd9e-56a6-4b90-8ccf-3acd0c697415</guid>
  <pubDate>Fri, 27 Jun 2025 03:00:00 +1000</pubDate>
  <author>Hugo Bowne-Anderson</author>
  <enclosure url="https://aphid.fireside.fm/d/1437767933/140c3904-8258-4c39-a698-a112b7077bd7/c45cdd9e-56a6-4b90-8ccf-3acd0c697415.mp3" length="45788781" type="audio/mpeg"/>
  <itunes:episodeType>full</itunes:episodeType>
  <itunes:season>1</itunes:season>
  <itunes:author>Hugo Bowne-Anderson</itunes:author>
  <itunes:subtitle>What does it take to actually ship LLM-powered features, and what breaks when you connect them to real production data?

In this episode, we hear from Philip Carter — then a Principal PM at Honeycomb and now a Product Management Director at Salesforce. In early 2023, he helped build one of the first LLM-powered SaaS features to ship to real users. More recently, he and his team built a production-ready MCP server.</itunes:subtitle>
  <itunes:duration>47:41</itunes:duration>
  <itunes:explicit>no</itunes:explicit>
  <itunes:image href="https://media24.fireside.fm/file/fireside-images-2024/podcasts/images/1/140c3904-8258-4c39-a698-a112b7077bd7/cover.jpg?v=1"/>
  <description>What does it take to actually ship LLM-powered features, and what breaks when you connect them to real production data?
In this episode, we hear from Philip Carter — then a Principal PM at Honeycomb and now a Product Management Director at Salesforce. In early 2023, he helped build one of the first LLM-powered SaaS features to ship to real users. More recently, he and his team built a production-ready MCP server.
We cover:
    • How to evaluate LLM systems using human-aligned judges
    • The spreadsheet-driven process behind shipping Honeycomb’s first LLM feature
    • The challenges of tool usage, prompt templates, and flaky model behavior
    • Where MCP shows promise, and where it breaks in the real world
If you’re working on LLMs in production, this one’s for you!
LINKS
So We Shipped an AI Product: Did it Work? by Philip Carter (https://www.honeycomb.io/blog/we-shipped-ai-product)
Vanishing Gradients YouTube Channel (https://www.youtube.com/channel/UC_NafIo-Ku2loOLrzm45ABA)  
Upcoming Events on Luma (https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk)
Hugo's recent newsletter about upcoming events and more! (https://hugobowne.substack.com/p/ai-as-a-civilizational-technology)
🎓 Learn more:
Hugo's course: Building LLM Applications for Data Scientists and Software Engineers (https://maven.com/s/course/d56067f338) — next cohort starts July 8: https://maven.com/s/course/d56067f338
📺 Watch the video version on YouTube: YouTube link (https://youtu.be/JDMzdaZh9Ig) 
</description>
  <itunes:keywords>data science, machine learning, AI, LLMs</itunes:keywords>
  <content:encoded>
    <![CDATA[<p>What does it take to actually ship LLM-powered features, and what breaks when you connect them to real production data?</p>

<p>In this episode, we hear from Philip Carter — then a Principal PM at Honeycomb and now a Product Management Director at Salesforce. In early 2023, he helped build one of the first LLM-powered SaaS features to ship to real users. More recently, he and his team built a production-ready MCP server.</p>

<p>We cover:<br>
    • How to evaluate LLM systems using human-aligned judges<br>
    • The spreadsheet-driven process behind shipping Honeycomb’s first LLM feature<br>
    • The challenges of tool usage, prompt templates, and flaky model behavior<br>
    • Where MCP shows promise, and where it breaks in the real world</p>

<p>If you’re working on LLMs in production, this one’s for you!</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://www.honeycomb.io/blog/we-shipped-ai-product" rel="nofollow">So We Shipped an AI Product: Did it Work? by Philip Carter</a></li>
<li><a href="https://www.youtube.com/channel/UC_NafIo-Ku2loOLrzm45ABA" rel="nofollow">Vanishing Gradients YouTube Channel</a><br></li>
<li><a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">Upcoming Events on Luma</a></li>
<li><a href="https://hugobowne.substack.com/p/ai-as-a-civilizational-technology" rel="nofollow">Hugo&#39;s recent newsletter about upcoming events and more!</a></li>
</ul>

<p>🎓 Learn more:</p>

<ul>
<li><strong>Hugo&#39;s course:</strong> <a href="https://maven.com/s/course/d56067f338" rel="nofollow">Building LLM Applications for Data Scientists and Software Engineers</a> — next cohort starts July 8: <a href="https://maven.com/s/course/d56067f338" rel="nofollow">https://maven.com/s/course/d56067f338</a></li>
</ul>

<p>📺 <strong>Watch the video version on YouTube:</strong> <a href="https://youtu.be/JDMzdaZh9Ig" rel="nofollow">YouTube link</a></p>]]>
  </content:encoded>
  <itunes:summary>
    <![CDATA[<p>What does it take to actually ship LLM-powered features, and what breaks when you connect them to real production data?</p>

<p>In this episode, we hear from Philip Carter — then a Principal PM at Honeycomb and now a Product Management Director at Salesforce. In early 2023, he helped build one of the first LLM-powered SaaS features to ship to real users. More recently, he and his team built a production-ready MCP server.</p>

<p>We cover:<br>
    • How to evaluate LLM systems using human-aligned judges<br>
    • The spreadsheet-driven process behind shipping Honeycomb’s first LLM feature<br>
    • The challenges of tool usage, prompt templates, and flaky model behavior<br>
    • Where MCP shows promise, and where it breaks in the real world</p>

<p>If you’re working on LLMs in production, this one’s for you!</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://www.honeycomb.io/blog/we-shipped-ai-product" rel="nofollow">So We Shipped an AI Product: Did it Work? by Philip Carter</a></li>
<li><a href="https://www.youtube.com/channel/UC_NafIo-Ku2loOLrzm45ABA" rel="nofollow">Vanishing Gradients YouTube Channel</a><br></li>
<li><a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">Upcoming Events on Luma</a></li>
<li><a href="https://hugobowne.substack.com/p/ai-as-a-civilizational-technology" rel="nofollow">Hugo&#39;s recent newsletter about upcoming events and more!</a></li>
</ul>

<p>🎓 Learn more:</p>

<ul>
<li><strong>Hugo&#39;s course:</strong> <a href="https://maven.com/s/course/d56067f338" rel="nofollow">Building LLM Applications for Data Scientists and Software Engineers</a> — next cohort starts July 8: <a href="https://maven.com/s/course/d56067f338" rel="nofollow">https://maven.com/s/course/d56067f338</a></li>
</ul>

<p>📺 <strong>Watch the video version on YouTube:</strong> <a href="https://youtu.be/JDMzdaZh9Ig" rel="nofollow">YouTube link</a></p>]]>
  </itunes:summary>
</item>
<item>
  <title>Episode 50: A Field Guide to Rapidly Improving AI Products -- With Hamel Husain</title>
  <link>https://vanishinggradients.fireside.fm/50</link>
  <guid isPermaLink="false">3851d92b-389c-4690-90c3-8a54ad73b7d8</guid>
  <pubDate>Tue, 17 Jun 2025 18:30:00 +1000</pubDate>
  <author>Hugo Bowne-Anderson</author>
  <enclosure url="https://aphid.fireside.fm/d/1437767933/140c3904-8258-4c39-a698-a112b7077bd7/3851d92b-389c-4690-90c3-8a54ad73b7d8.mp3" length="54176426" type="audio/mpeg"/>
  <itunes:episodeType>full</itunes:episodeType>
  <itunes:season>1</itunes:season>
  <itunes:author>Hugo Bowne-Anderson</itunes:author>
  <itunes:subtitle>Hugo talks with Hamel Hussain (ex-Airbnb, GitHub, DataRobot) about how to improve AI products through evaluation, error analysis, and iteration. They discuss why most teams overlook debugging LLM systems, how to prioritize what to fix, and why evals are not just metrics—but a full development process.</itunes:subtitle>
  <itunes:duration>27:42</itunes:duration>
  <itunes:explicit>no</itunes:explicit>
  <itunes:image href="https://media24.fireside.fm/file/fireside-images-2024/podcasts/images/1/140c3904-8258-4c39-a698-a112b7077bd7/cover.jpg?v=1"/>
  <description>If we want AI systems that actually work, we need to get much better at evaluating them, not just building more pipelines, agents, and frameworks.
In this episode, Hugo talks with Hamel Hussain (ex-Airbnb, GitHub, DataRobot) about how teams can improve AI products by focusing on error analysis, data inspection, and systematic iteration. The conversation is based on Hamel’s blog post A Field Guide to Rapidly Improving AI Products, which he joined Hugo’s class to discuss.
They cover:
🔍 Why most teams struggle to measure whether their systems are actually improving  
📊 How error analysis helps you prioritize what to fix (and when to write evals)  
🧮 Why evaluation isn’t just a metric — but a full development process  
⚠️ Common mistakes when debugging LLM and agent systems  
🛠️ How to think about the tradeoffs in adding more evals vs. fixing obvious issues  
👥 Why enabling domain experts — not just engineers — can accelerate iteration
If you’ve ever built an AI system and found yourself unsure how to make it better, this conversation is for you.
LINKS
* A Field Guide to Rapidly Improving AI Products by Hamel Husain (https://hamel.dev/blog/posts/field-guide/)
* Vanishing Gradients YouTube Channel (https://www.youtube.com/channel/UC_NafIo-Ku2loOLrzm45ABA)  
* Upcoming Events on Luma (https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk)
* Hugo's recent newsletter about upcoming events and more! (https://hugobowne.substack.com/p/ai-as-a-civilizational-technology)
🎓 Learn more:
Hugo's course: Building LLM Applications for Data Scientists and Software Engineers (https://maven.com/s/course/d56067f338) — next cohort starts July 8: https://maven.com/s/course/d56067f338
Hamel &amp;amp; Shreya's course: AI Evals For Engineers &amp;amp; PMs (https://maven.com/parlance-labs/evals?promoCode=GOHUGORGOHOME) — use code GOHUGORGOHOME for $800 off
📺 Watch the video version on YouTube: YouTube link (https://youtu.be/rWToRi2_SeY) 
</description>
  <itunes:keywords>data science, machine learning, AI, LLMs, evas</itunes:keywords>
  <content:encoded>
    <![CDATA[<p>If we want AI systems that actually work, we need to get much better at evaluating them, not just building more pipelines, agents, and frameworks.</p>

<p>In this episode, Hugo talks with Hamel Hussain (ex-Airbnb, GitHub, DataRobot) about how teams can improve AI products by focusing on error analysis, data inspection, and systematic iteration. The conversation is based on Hamel’s blog post <em>A Field Guide to Rapidly Improving AI Products</em>, which he joined Hugo’s class to discuss.</p>

<p>They cover:<br>
🔍 Why most teams struggle to measure whether their systems are actually improving<br><br>
📊 How error analysis helps you prioritize what to fix (and when to write evals)<br><br>
🧮 Why evaluation isn’t just a metric — but a full development process<br><br>
⚠️ Common mistakes when debugging LLM and agent systems<br><br>
🛠️ How to think about the tradeoffs in adding more evals vs. fixing obvious issues<br><br>
👥 Why enabling domain experts — not just engineers — can accelerate iteration</p>

<p>If you’ve ever built an AI system and found yourself unsure how to make it better, this conversation is for you.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://hamel.dev/blog/posts/field-guide/" rel="nofollow">A Field Guide to Rapidly Improving AI Products by Hamel Husain</a></li>
<li><a href="https://www.youtube.com/channel/UC_NafIo-Ku2loOLrzm45ABA" rel="nofollow">Vanishing Gradients YouTube Channel</a><br></li>
<li><a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">Upcoming Events on Luma</a></li>
<li><a href="https://hugobowne.substack.com/p/ai-as-a-civilizational-technology" rel="nofollow">Hugo&#39;s recent newsletter about upcoming events and more!</a></li>
</ul>

<hr>

<p>🎓 Learn more:</p>

<ul>
<li><strong>Hugo&#39;s course:</strong> <a href="https://maven.com/s/course/d56067f338" rel="nofollow">Building LLM Applications for Data Scientists and Software Engineers</a> — next cohort starts July 8: <a href="https://maven.com/s/course/d56067f338" rel="nofollow">https://maven.com/s/course/d56067f338</a></li>
<li><strong>Hamel &amp; Shreya&#39;s course:</strong> <a href="https://maven.com/parlance-labs/evals?promoCode=GOHUGORGOHOME" rel="nofollow">AI Evals For Engineers &amp; PMs</a> — use code <code>GOHUGORGOHOME</code> for $800 off</li>
</ul>

<p>📺 <strong>Watch the video version on YouTube:</strong> <a href="https://youtu.be/rWToRi2_SeY" rel="nofollow">YouTube link</a></p>]]>
  </content:encoded>
  <itunes:summary>
    <![CDATA[<p>If we want AI systems that actually work, we need to get much better at evaluating them, not just building more pipelines, agents, and frameworks.</p>

<p>In this episode, Hugo talks with Hamel Hussain (ex-Airbnb, GitHub, DataRobot) about how teams can improve AI products by focusing on error analysis, data inspection, and systematic iteration. The conversation is based on Hamel’s blog post <em>A Field Guide to Rapidly Improving AI Products</em>, which he joined Hugo’s class to discuss.</p>

<p>They cover:<br>
🔍 Why most teams struggle to measure whether their systems are actually improving<br><br>
📊 How error analysis helps you prioritize what to fix (and when to write evals)<br><br>
🧮 Why evaluation isn’t just a metric — but a full development process<br><br>
⚠️ Common mistakes when debugging LLM and agent systems<br><br>
🛠️ How to think about the tradeoffs in adding more evals vs. fixing obvious issues<br><br>
👥 Why enabling domain experts — not just engineers — can accelerate iteration</p>

<p>If you’ve ever built an AI system and found yourself unsure how to make it better, this conversation is for you.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://hamel.dev/blog/posts/field-guide/" rel="nofollow">A Field Guide to Rapidly Improving AI Products by Hamel Husain</a></li>
<li><a href="https://www.youtube.com/channel/UC_NafIo-Ku2loOLrzm45ABA" rel="nofollow">Vanishing Gradients YouTube Channel</a><br></li>
<li><a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">Upcoming Events on Luma</a></li>
<li><a href="https://hugobowne.substack.com/p/ai-as-a-civilizational-technology" rel="nofollow">Hugo&#39;s recent newsletter about upcoming events and more!</a></li>
</ul>

<hr>

<p>🎓 Learn more:</p>

<ul>
<li><strong>Hugo&#39;s course:</strong> <a href="https://maven.com/s/course/d56067f338" rel="nofollow">Building LLM Applications for Data Scientists and Software Engineers</a> — next cohort starts July 8: <a href="https://maven.com/s/course/d56067f338" rel="nofollow">https://maven.com/s/course/d56067f338</a></li>
<li><strong>Hamel &amp; Shreya&#39;s course:</strong> <a href="https://maven.com/parlance-labs/evals?promoCode=GOHUGORGOHOME" rel="nofollow">AI Evals For Engineers &amp; PMs</a> — use code <code>GOHUGORGOHOME</code> for $800 off</li>
</ul>

<p>📺 <strong>Watch the video version on YouTube:</strong> <a href="https://youtu.be/rWToRi2_SeY" rel="nofollow">YouTube link</a></p>]]>
  </itunes:summary>
</item>
<item>
  <title>Episode 49: Why Data and AI Still Break at Scale (and What to Do About It)</title>
  <link>https://vanishinggradients.fireside.fm/49</link>
  <guid isPermaLink="false">309762f9-59cd-4f24-bea5-8e692a0d870f</guid>
  <pubDate>Thu, 05 Jun 2025 14:00:00 +1000</pubDate>
  <author>Hugo Bowne-Anderson</author>
  <enclosure url="https://aphid.fireside.fm/d/1437767933/140c3904-8258-4c39-a698-a112b7077bd7/309762f9-59cd-4f24-bea5-8e692a0d870f.mp3" length="117738811" type="audio/mpeg"/>
  <itunes:episodeType>full</itunes:episodeType>
  <itunes:season>1</itunes:season>
  <itunes:author>Hugo Bowne-Anderson</itunes:author>
  <itunes:subtitle>Hugo talks with Akshay Agrawal (Marimo, ex-Google Brain, Netflix, Stanford) about why data and AI systems still break at scale—and what it takes to fix them. They dive into the limits of existing workflows, the importance of reproducibility and reactive execution, and how Marimo reimagines notebooks for modern software development.</itunes:subtitle>
  <itunes:duration>1:21:45</itunes:duration>
  <itunes:explicit>no</itunes:explicit>
  <itunes:image href="https://media24.fireside.fm/file/fireside-images-2024/podcasts/images/1/140c3904-8258-4c39-a698-a112b7077bd7/cover.jpg?v=1"/>
  <description>If we want AI systems that actually work in production, we need better infrastructure—not just better models.
In this episode, Hugo talks with Akshay Agrawal (Marimo, ex-Google Brain, Netflix, Stanford) about why data and AI pipelines still break down at scale, and how we can fix the fundamentals: reproducibility, composability, and reliable execution.
They discuss:
🔁 Why reactive execution matters—and how current tools fall short
🛠️ The design goals behind Marimo, a new kind of Python notebook
⚙️ The hidden costs of traditional workflows (and what breaks at scale)
📦 What it takes to build modular, maintainable AI apps
🧪 Why debugging LLM systems is so hard—and what better tooling looks like
🌍 What we can learn from decades of tools built for and by data practitioners
Toward the end of the episode, Hugo and Akshay walk through two live demos: Hugo shares how he’s been using Marimo to prototype an app that extracts structured data from world leader bios, and Akshay shows how Marimo handles agentic workflows with memory and tool use—built entirely in a notebook.
This episode is about tools, but it’s also about culture. If you’ve ever hit a wall with your current stack—or felt like your tools were working against you—this one’s for you.
LINKS
* marimo | a next-generation Python notebook (https://marimo.io/)
* SciPy conference, 2025 (https://www.scipy2025.scipy.org/)
* Hugo's face Marimo World Leader Face Embedding demo (https://www.youtube.com/watch?v=DO21QEcLOxM)
* Vanishing Gradients YouTube Channel (https://www.youtube.com/channel/UC_NafIo-Ku2loOLrzm45ABA)  
* Upcoming Events on Luma (https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk)
* Hugo's recent newsletter about upcoming events and more! (https://hugobowne.substack.com/p/ai-as-a-civilizational-technology)
* Watch the podcast here on YouTube! (https://youtube.com/live/WVxAz19tgZY?feature=share)
🎓 Want to go deeper?
Check out Hugo's course: Building LLM Applications for Data Scientists and Software Engineers.
Learn how to design, test, and deploy production-grade LLM systems — with observability, feedback loops, and structure built in.
This isn’t about vibes or fragile agents. It’s about making LLMs reliable, testable, and actually useful.
Includes over $800 in compute credits and guest lectures from experts at DeepMind, Moderna, and more.
Cohort starts July 8 — Use this link for a 10% discount (https://maven.com/hugo-stefan/building-llm-apps-ds-and-swe-from-first-principles?promoCode=LLM10) 
</description>
  <itunes:keywords>data science, machine learning, AI, LLMs</itunes:keywords>
  <content:encoded>
    <![CDATA[<p>If we want AI systems that actually work in production, we need better infrastructure—not just better models.</p>

<p>In this episode, Hugo talks with Akshay Agrawal (Marimo, ex-Google Brain, Netflix, Stanford) about why data and AI pipelines still break down at scale, and how we can fix the fundamentals: reproducibility, composability, and reliable execution.</p>

<p>They discuss:<br>
🔁 Why reactive execution matters—and how current tools fall short<br>
🛠️ The design goals behind Marimo, a new kind of Python notebook<br>
⚙️ The hidden costs of traditional workflows (and what breaks at scale)<br>
📦 What it takes to build modular, maintainable AI apps<br>
🧪 Why debugging LLM systems is so hard—and what better tooling looks like<br>
🌍 What we can learn from decades of tools built for and by data practitioners</p>

<p>Toward the end of the episode, Hugo and Akshay walk through two live demos: Hugo shares how he’s been using Marimo to prototype an app that extracts structured data from world leader bios, and Akshay shows how Marimo handles agentic workflows with memory and tool use—built entirely in a notebook.</p>

<p>This episode is about tools, but it’s also about culture. If you’ve ever hit a wall with your current stack—or felt like your tools were working against you—this one’s for you.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://marimo.io/" rel="nofollow">marimo | a next-generation Python notebook</a></li>
<li><a href="https://www.scipy2025.scipy.org/" rel="nofollow">SciPy conference, 2025</a></li>
<li><a href="https://www.youtube.com/watch?v=DO21QEcLOxM" rel="nofollow">Hugo&#39;s face Marimo World Leader Face Embedding demo</a></li>
<li><a href="https://www.youtube.com/channel/UC_NafIo-Ku2loOLrzm45ABA" rel="nofollow">Vanishing Gradients YouTube Channel</a><br></li>
<li><a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">Upcoming Events on Luma</a></li>
<li><a href="https://hugobowne.substack.com/p/ai-as-a-civilizational-technology" rel="nofollow">Hugo&#39;s recent newsletter about upcoming events and more!</a></li>
<li><a href="https://youtube.com/live/WVxAz19tgZY?feature=share" rel="nofollow">Watch the podcast here on YouTube!</a></li>
</ul>

<p>🎓 Want to go deeper?<br>
Check out Hugo&#39;s course: <em>Building LLM Applications for Data Scientists and Software Engineers.</em><br>
Learn how to design, test, and deploy production-grade LLM systems — with observability, feedback loops, and structure built in.<br>
This isn’t about vibes or fragile agents. It’s about making LLMs reliable, testable, and actually useful.</p>

<p>Includes over $800 in compute credits and guest lectures from experts at DeepMind, Moderna, and more.<br>
Cohort starts July 8 — <a href="https://maven.com/hugo-stefan/building-llm-apps-ds-and-swe-from-first-principles?promoCode=LLM10" rel="nofollow">Use this link for a 10% discount</a></p>]]>
  </content:encoded>
  <itunes:summary>
    <![CDATA[<p>If we want AI systems that actually work in production, we need better infrastructure—not just better models.</p>

<p>In this episode, Hugo talks with Akshay Agrawal (Marimo, ex-Google Brain, Netflix, Stanford) about why data and AI pipelines still break down at scale, and how we can fix the fundamentals: reproducibility, composability, and reliable execution.</p>

<p>They discuss:<br>
🔁 Why reactive execution matters—and how current tools fall short<br>
🛠️ The design goals behind Marimo, a new kind of Python notebook<br>
⚙️ The hidden costs of traditional workflows (and what breaks at scale)<br>
📦 What it takes to build modular, maintainable AI apps<br>
🧪 Why debugging LLM systems is so hard—and what better tooling looks like<br>
🌍 What we can learn from decades of tools built for and by data practitioners</p>

<p>Toward the end of the episode, Hugo and Akshay walk through two live demos: Hugo shares how he’s been using Marimo to prototype an app that extracts structured data from world leader bios, and Akshay shows how Marimo handles agentic workflows with memory and tool use—built entirely in a notebook.</p>

<p>This episode is about tools, but it’s also about culture. If you’ve ever hit a wall with your current stack—or felt like your tools were working against you—this one’s for you.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://marimo.io/" rel="nofollow">marimo | a next-generation Python notebook</a></li>
<li><a href="https://www.scipy2025.scipy.org/" rel="nofollow">SciPy conference, 2025</a></li>
<li><a href="https://www.youtube.com/watch?v=DO21QEcLOxM" rel="nofollow">Hugo&#39;s face Marimo World Leader Face Embedding demo</a></li>
<li><a href="https://www.youtube.com/channel/UC_NafIo-Ku2loOLrzm45ABA" rel="nofollow">Vanishing Gradients YouTube Channel</a><br></li>
<li><a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">Upcoming Events on Luma</a></li>
<li><a href="https://hugobowne.substack.com/p/ai-as-a-civilizational-technology" rel="nofollow">Hugo&#39;s recent newsletter about upcoming events and more!</a></li>
<li><a href="https://youtube.com/live/WVxAz19tgZY?feature=share" rel="nofollow">Watch the podcast here on YouTube!</a></li>
</ul>

<p>🎓 Want to go deeper?<br>
Check out Hugo&#39;s course: <em>Building LLM Applications for Data Scientists and Software Engineers.</em><br>
Learn how to design, test, and deploy production-grade LLM systems — with observability, feedback loops, and structure built in.<br>
This isn’t about vibes or fragile agents. It’s about making LLMs reliable, testable, and actually useful.</p>

<p>Includes over $800 in compute credits and guest lectures from experts at DeepMind, Moderna, and more.<br>
Cohort starts July 8 — <a href="https://maven.com/hugo-stefan/building-llm-apps-ds-and-swe-from-first-principles?promoCode=LLM10" rel="nofollow">Use this link for a 10% discount</a></p>]]>
  </itunes:summary>
</item>
<item>
  <title>Episode 47: The Great Pacific Garbage Patch of Code Slop with Joe Reis</title>
  <link>https://vanishinggradients.fireside.fm/47</link>
  <guid isPermaLink="false">decc9c1a-f18a-41e9-947a-e58fa0957f1e</guid>
  <pubDate>Mon, 07 Apr 2025 10:00:00 +1000</pubDate>
  <author>Hugo Bowne-Anderson</author>
  <enclosure url="https://aphid.fireside.fm/d/1437767933/140c3904-8258-4c39-a698-a112b7077bd7/decc9c1a-f18a-41e9-947a-e58fa0957f1e.mp3" length="76045085" type="audio/mpeg"/>
  <itunes:episodeType>full</itunes:episodeType>
  <itunes:season>1</itunes:season>
  <itunes:author>Hugo Bowne-Anderson</itunes:author>
  <itunes:subtitle>What if the cost of writing code dropped to zero — but the cost of understanding it skyrocketed?

In this episode, Hugo sits down with Joe Reis to unpack how AI tooling is reshaping the software development lifecycle — from experimentation and prototyping to deployment, maintainability, and everything in between.</itunes:subtitle>
  <itunes:duration>1:19:12</itunes:duration>
  <itunes:explicit>yes</itunes:explicit>
  <itunes:image href="https://media24.fireside.fm/file/fireside-images-2024/podcasts/images/1/140c3904-8258-4c39-a698-a112b7077bd7/cover.jpg?v=1"/>
  <description>What if the cost of writing code dropped to zero — but the cost of understanding it skyrocketed?
In this episode, Hugo sits down with Joe Reis to unpack how AI tooling is reshaping the software development lifecycle — from experimentation and prototyping to deployment, maintainability, and everything in between.
Joe is the co-author of Fundamentals of Data Engineering and a longtime voice on the systems side of modern software. He’s also one of the sharpest critics of “vibe coding” — the emerging pattern of writing software by feel, with heavy reliance on LLMs and little regard for structure or quality.
We dive into:
    • Why “vibe coding” is more than a meme — and what it says about how we build today
    • How AI tools expand the surface area of software creation — for better and worse
    • What happens to technical debt, testing, and security when generation outpaces understanding
    • The changing definition of “production” in a world of ephemeral, internal, or just-good-enough tools
    • How AI is flattening the learning curve — and threatening the talent pipeline
    • Joe’s view on what real craftsmanship means in an age of disposable code
This conversation isn’t about doom, and it’s not about hype. It’s about mapping the real, messy terrain of what it means to build software today — and how to do it with care.
LINKS
* Joe's Practical Data Modeling Newsletter on Substack (https://practicaldatamodeling.substack.com/)
* Joe's Practical Data Modeling Server on Discord (https://discord.gg/HhSZVvWDBb)
* Vanishing Gradients YouTube Channel (https://www.youtube.com/channel/UC_NafIo-Ku2loOLrzm45ABA)  
* Upcoming Events on Luma (https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk)
🎓 Want to go deeper?
Check out my course: Building LLM Applications for Data Scientists and Software Engineers.
Learn how to design, test, and deploy production-grade LLM systems — with observability, feedback loops, and structure built in.
This isn’t about vibes or fragile agents. It’s about making LLMs reliable, testable, and actually useful.
Includes over $800 in compute credits and guest lectures from experts at DeepMind, Moderna, and more.
Cohort starts July 8 — Use this link for a 10% discount (https://maven.com/hugo-stefan/building-llm-apps-ds-and-swe-from-first-principles?promoCode=LLM10) 
</description>
  <itunes:keywords>AI, LLMs, data science, machine learning, data science, GenAI, vibe coding</itunes:keywords>
  <content:encoded>
    <![CDATA[<p>What if the cost of writing code dropped to zero — but the cost of understanding it skyrocketed?</p>

<p>In this episode, Hugo sits down with Joe Reis to unpack how AI tooling is reshaping the software development lifecycle — from experimentation and prototyping to deployment, maintainability, and everything in between.</p>

<p>Joe is the co-author of Fundamentals of Data Engineering and a longtime voice on the systems side of modern software. He’s also one of the sharpest critics of “vibe coding” — the emerging pattern of writing software by feel, with heavy reliance on LLMs and little regard for structure or quality.</p>

<p>We dive into:<br>
    • Why “vibe coding” is more than a meme — and what it says about how we build today<br>
    • How AI tools expand the surface area of software creation — for better and worse<br>
    • What happens to technical debt, testing, and security when generation outpaces understanding<br>
    • The changing definition of “production” in a world of ephemeral, internal, or just-good-enough tools<br>
    • How AI is flattening the learning curve — and threatening the talent pipeline<br>
    • Joe’s view on what real craftsmanship means in an age of disposable code</p>

<p>This conversation isn’t about doom, and it’s not about hype. It’s about mapping the real, messy terrain of what it means to build software today — and how to do it with care.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://practicaldatamodeling.substack.com/" rel="nofollow">Joe&#39;s Practical Data Modeling Newsletter on Substack</a></li>
<li><a href="https://discord.gg/HhSZVvWDBb" rel="nofollow">Joe&#39;s Practical Data Modeling Server on Discord</a></li>
<li><a href="https://www.youtube.com/channel/UC_NafIo-Ku2loOLrzm45ABA" rel="nofollow">Vanishing Gradients YouTube Channel</a><br></li>
<li><a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">Upcoming Events on Luma</a></li>
</ul>

<p>🎓 Want to go deeper?<br>
Check out my course: <em>Building LLM Applications for Data Scientists and Software Engineers.</em><br>
Learn how to design, test, and deploy production-grade LLM systems — with observability, feedback loops, and structure built in.<br>
This isn’t about vibes or fragile agents. It’s about making LLMs reliable, testable, and actually useful.</p>

<p>Includes over $800 in compute credits and guest lectures from experts at DeepMind, Moderna, and more.<br>
Cohort starts July 8 — <a href="https://maven.com/hugo-stefan/building-llm-apps-ds-and-swe-from-first-principles?promoCode=LLM10" rel="nofollow">Use this link for a 10% discount</a></p>]]>
  </content:encoded>
  <itunes:summary>
    <![CDATA[<p>What if the cost of writing code dropped to zero — but the cost of understanding it skyrocketed?</p>

<p>In this episode, Hugo sits down with Joe Reis to unpack how AI tooling is reshaping the software development lifecycle — from experimentation and prototyping to deployment, maintainability, and everything in between.</p>

<p>Joe is the co-author of Fundamentals of Data Engineering and a longtime voice on the systems side of modern software. He’s also one of the sharpest critics of “vibe coding” — the emerging pattern of writing software by feel, with heavy reliance on LLMs and little regard for structure or quality.</p>

<p>We dive into:<br>
    • Why “vibe coding” is more than a meme — and what it says about how we build today<br>
    • How AI tools expand the surface area of software creation — for better and worse<br>
    • What happens to technical debt, testing, and security when generation outpaces understanding<br>
    • The changing definition of “production” in a world of ephemeral, internal, or just-good-enough tools<br>
    • How AI is flattening the learning curve — and threatening the talent pipeline<br>
    • Joe’s view on what real craftsmanship means in an age of disposable code</p>

<p>This conversation isn’t about doom, and it’s not about hype. It’s about mapping the real, messy terrain of what it means to build software today — and how to do it with care.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://practicaldatamodeling.substack.com/" rel="nofollow">Joe&#39;s Practical Data Modeling Newsletter on Substack</a></li>
<li><a href="https://discord.gg/HhSZVvWDBb" rel="nofollow">Joe&#39;s Practical Data Modeling Server on Discord</a></li>
<li><a href="https://www.youtube.com/channel/UC_NafIo-Ku2loOLrzm45ABA" rel="nofollow">Vanishing Gradients YouTube Channel</a><br></li>
<li><a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">Upcoming Events on Luma</a></li>
</ul>

<p>🎓 Want to go deeper?<br>
Check out my course: <em>Building LLM Applications for Data Scientists and Software Engineers.</em><br>
Learn how to design, test, and deploy production-grade LLM systems — with observability, feedback loops, and structure built in.<br>
This isn’t about vibes or fragile agents. It’s about making LLMs reliable, testable, and actually useful.</p>

<p>Includes over $800 in compute credits and guest lectures from experts at DeepMind, Moderna, and more.<br>
Cohort starts July 8 — <a href="https://maven.com/hugo-stefan/building-llm-apps-ds-and-swe-from-first-principles?promoCode=LLM10" rel="nofollow">Use this link for a 10% discount</a></p>]]>
  </itunes:summary>
</item>
<item>
  <title>Episode 46: Software Composition Is the New Vibe Coding</title>
  <link>https://vanishinggradients.fireside.fm/46</link>
  <guid isPermaLink="false">dcb8396f-ece2-4636-951c-8ad44d698d15</guid>
  <pubDate>Thu, 03 Apr 2025 13:00:00 +1100</pubDate>
  <author>Hugo Bowne-Anderson</author>
  <enclosure url="https://aphid.fireside.fm/d/1437767933/140c3904-8258-4c39-a698-a112b7077bd7/dcb8396f-ece2-4636-951c-8ad44d698d15.mp3" length="99299288" type="audio/mpeg"/>
  <itunes:episodeType>full</itunes:episodeType>
  <itunes:season>1</itunes:season>
  <itunes:author>Hugo Bowne-Anderson</itunes:author>
  <itunes:subtitle>What if building software felt more like composing than coding?

In this episode, Hugo and Greg explore how LLMs are reshaping the way we think about software development—from deterministic programming to a more flexible, prompt-driven, and collaborative style of building. It’s not just hype or grift—it’s a real shift in how we express intent, reason about systems, and collaborate across roles.

Hugo speaks with Greg Ceccarelli—co-founder of SpecStory, former CPO at Pluralsight, and Director of Data Science at GitHub—about the rise of software composition and how it changes the way individuals and teams create with LLMs.</itunes:subtitle>
  <itunes:duration>1:08:57</itunes:duration>
  <itunes:explicit>no</itunes:explicit>
  <itunes:image href="https://media24.fireside.fm/file/fireside-images-2024/podcasts/images/1/140c3904-8258-4c39-a698-a112b7077bd7/cover.jpg?v=1"/>
  <description>What if building software felt more like composing than coding?
In this episode, Hugo and Greg explore how LLMs are reshaping the way we think about software development—from deterministic programming to a more flexible, prompt-driven, and collaborative style of building. It’s not just hype or grift—it’s a real shift in how we express intent, reason about systems, and collaborate across roles.
Hugo speaks with Greg Ceccarelli—co-founder of SpecStory, former CPO at Pluralsight, and Director of Data Science at GitHub—about the rise of software composition and how it changes the way individuals and teams create with LLMs.
We dive into:
- Why software composition is emerging as a serious alternative to traditional coding
- The real difference between vibe coding and production-minded prototyping
- How LLMs are expanding who gets to build software—and how
- What changes when you focus on intent, not just code
- What Greg is building with SpecStory to support collaborative, traceable AI-native workflows
- The challenges (and joys) of debugging and exploring with agentic tools like Cursor and Claude
We’ve removed the visual demos from the audio—but you can catch our live-coded Chrome extension and JFK document explorer on YouTube. Links below.
JFK Docs Vibe Coding Demo (YouTube) (https://youtu.be/JpXCkuV58QE)  
Chrome Extension Vibe Coding Demo (YouTube) (https://youtu.be/ESVKp37jDwc)  
Meditations on Tech (Greg’s Substack) (https://www.meditationsontech.com/)  
Simon Willison on Vibe Coding (https://simonwillison.net/2025/Mar/19/vibe-coding/)  
Johnno Whitaker: On Vibe Coding (https://johnowhitaker.dev/essays/vibe_coding.html)  
Tim O’Reilly – The End of Programming (https://www.oreilly.com/radar/the-end-of-programming-as-we-know-it/)  
Vanishing Gradients YouTube Channel (https://www.youtube.com/channel/UC_NafIo-Ku2loOLrzm45ABA)  
Upcoming Events on Luma (https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk)  
Greg Ceccarelli on LinkedIn (https://www.linkedin.com/in/gregceccarelli/)  
Greg’s Hacker News Post on GOOD (https://news.ycombinator.com/item?id=43557698)  
SpecStory: GOOD – Git Companion for AI Workflows (https://github.com/specstoryai/getspecstory/blob/main/GOOD.md)
🎓 Want to go deeper?
Check out my course: Building LLM Applications for Data Scientists and Software Engineers.
Learn how to design, test, and deploy production-grade LLM systems — with observability, feedback loops, and structure built in.
This isn’t about vibes or fragile agents. It’s about making LLMs reliable, testable, and actually useful.
Includes over $2,500 in compute credits and guest lectures from experts at DeepMind, Moderna, and more.
Cohort starts April 7 — Use this link for a 10% discount (https://maven.com/hugo-stefan/building-llm-apps-ds-and-swe-from-first-principles?promoCode=LLM10)
🔍 Want to help shape the future of SpecStory?
Greg and the team are looking for design partners for their new SpecStory Teams product—built for collaborative, AI-native software development.
If you're working with LLMs in a team setting and want to influence the next wave of developer tools, you can apply here:  
👉 specstory.com/teams (https://specstory.com/teams) 
</description>
  <itunes:keywords>AI, LLMs, data science, machine learning, data science, GenAI, vibe coding</itunes:keywords>
  <content:encoded>
    <![CDATA[<p>What if building software felt more like composing than coding?</p>

<p>In this episode, Hugo and Greg explore how LLMs are reshaping the way we think about software development—from deterministic programming to a more flexible, prompt-driven, and collaborative style of building. It’s not just hype or grift—it’s a real shift in how we express intent, reason about systems, and collaborate across roles.</p>

<p>Hugo speaks with Greg Ceccarelli—co-founder of SpecStory, former CPO at Pluralsight, and Director of Data Science at GitHub—about the rise of software composition and how it changes the way individuals and teams create with LLMs.</p>

<p>We dive into:</p>

<ul>
<li>Why software composition is emerging as a serious alternative to traditional coding</li>
<li>The real difference between vibe coding and production-minded prototyping</li>
<li>How LLMs are expanding who gets to build software—and how</li>
<li>What changes when you focus on intent, not just code</li>
<li>What Greg is building with SpecStory to support collaborative, traceable AI-native workflows</li>
<li>The challenges (and joys) of debugging and exploring with agentic tools like Cursor and Claude</li>
</ul>

<p>We’ve removed the visual demos from the audio—but you can catch our live-coded Chrome extension and JFK document explorer on YouTube. Links below.</p>

<ul>
<li><a href="https://youtu.be/JpXCkuV58QE" rel="nofollow">JFK Docs Vibe Coding Demo (YouTube)</a><br></li>
<li><a href="https://youtu.be/ESVKp37jDwc" rel="nofollow">Chrome Extension Vibe Coding Demo (YouTube)</a><br></li>
<li><a href="https://www.meditationsontech.com/" rel="nofollow">Meditations on Tech (Greg’s Substack)</a><br></li>
<li><a href="https://simonwillison.net/2025/Mar/19/vibe-coding/" rel="nofollow">Simon Willison on Vibe Coding</a><br></li>
<li><a href="https://johnowhitaker.dev/essays/vibe_coding.html" rel="nofollow">Johnno Whitaker: On Vibe Coding</a><br></li>
<li><a href="https://www.oreilly.com/radar/the-end-of-programming-as-we-know-it/" rel="nofollow">Tim O’Reilly – The End of Programming</a><br></li>
<li><a href="https://www.youtube.com/channel/UC_NafIo-Ku2loOLrzm45ABA" rel="nofollow">Vanishing Gradients YouTube Channel</a><br></li>
<li><a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">Upcoming Events on Luma</a><br></li>
<li><a href="https://www.linkedin.com/in/gregceccarelli/" rel="nofollow">Greg Ceccarelli on LinkedIn</a><br></li>
<li><a href="https://news.ycombinator.com/item?id=43557698" rel="nofollow">Greg’s Hacker News Post on GOOD</a><br></li>
<li><a href="https://github.com/specstoryai/getspecstory/blob/main/GOOD.md" rel="nofollow">SpecStory: GOOD – Git Companion for AI Workflows</a></li>
</ul>

<p>🎓 Want to go deeper?<br>
Check out my course: <em>Building LLM Applications for Data Scientists and Software Engineers.</em><br>
Learn how to design, test, and deploy production-grade LLM systems — with observability, feedback loops, and structure built in.<br>
This isn’t about vibes or fragile agents. It’s about making LLMs reliable, testable, and actually useful.</p>

<p>Includes over $2,500 in compute credits and guest lectures from experts at DeepMind, Moderna, and more.<br>
Cohort starts April 7 — <a href="https://maven.com/hugo-stefan/building-llm-apps-ds-and-swe-from-first-principles?promoCode=LLM10" rel="nofollow">Use this link for a 10% discount</a></p>

<h3>🔍 Want to help shape the future of SpecStory?</h3>

<p>Greg and the team are looking for <strong>design partners</strong> for their new SpecStory Teams product—built for collaborative, AI-native software development.</p>

<p>If you&#39;re working with LLMs in a team setting and want to influence the next wave of developer tools, you can apply here:<br><br>
👉 <a href="https://specstory.com/teams" rel="nofollow">specstory.com/teams</a></p>]]>
  </content:encoded>
  <itunes:summary>
    <![CDATA[<p>What if building software felt more like composing than coding?</p>

<p>In this episode, Hugo and Greg explore how LLMs are reshaping the way we think about software development—from deterministic programming to a more flexible, prompt-driven, and collaborative style of building. It’s not just hype or grift—it’s a real shift in how we express intent, reason about systems, and collaborate across roles.</p>

<p>Hugo speaks with Greg Ceccarelli—co-founder of SpecStory, former CPO at Pluralsight, and Director of Data Science at GitHub—about the rise of software composition and how it changes the way individuals and teams create with LLMs.</p>

<p>We dive into:</p>

<ul>
<li>Why software composition is emerging as a serious alternative to traditional coding</li>
<li>The real difference between vibe coding and production-minded prototyping</li>
<li>How LLMs are expanding who gets to build software—and how</li>
<li>What changes when you focus on intent, not just code</li>
<li>What Greg is building with SpecStory to support collaborative, traceable AI-native workflows</li>
<li>The challenges (and joys) of debugging and exploring with agentic tools like Cursor and Claude</li>
</ul>

<p>We’ve removed the visual demos from the audio—but you can catch our live-coded Chrome extension and JFK document explorer on YouTube. Links below.</p>

<ul>
<li><a href="https://youtu.be/JpXCkuV58QE" rel="nofollow">JFK Docs Vibe Coding Demo (YouTube)</a><br></li>
<li><a href="https://youtu.be/ESVKp37jDwc" rel="nofollow">Chrome Extension Vibe Coding Demo (YouTube)</a><br></li>
<li><a href="https://www.meditationsontech.com/" rel="nofollow">Meditations on Tech (Greg’s Substack)</a><br></li>
<li><a href="https://simonwillison.net/2025/Mar/19/vibe-coding/" rel="nofollow">Simon Willison on Vibe Coding</a><br></li>
<li><a href="https://johnowhitaker.dev/essays/vibe_coding.html" rel="nofollow">Johnno Whitaker: On Vibe Coding</a><br></li>
<li><a href="https://www.oreilly.com/radar/the-end-of-programming-as-we-know-it/" rel="nofollow">Tim O’Reilly – The End of Programming</a><br></li>
<li><a href="https://www.youtube.com/channel/UC_NafIo-Ku2loOLrzm45ABA" rel="nofollow">Vanishing Gradients YouTube Channel</a><br></li>
<li><a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">Upcoming Events on Luma</a><br></li>
<li><a href="https://www.linkedin.com/in/gregceccarelli/" rel="nofollow">Greg Ceccarelli on LinkedIn</a><br></li>
<li><a href="https://news.ycombinator.com/item?id=43557698" rel="nofollow">Greg’s Hacker News Post on GOOD</a><br></li>
<li><a href="https://github.com/specstoryai/getspecstory/blob/main/GOOD.md" rel="nofollow">SpecStory: GOOD – Git Companion for AI Workflows</a></li>
</ul>

<p>🎓 Want to go deeper?<br>
Check out my course: <em>Building LLM Applications for Data Scientists and Software Engineers.</em><br>
Learn how to design, test, and deploy production-grade LLM systems — with observability, feedback loops, and structure built in.<br>
This isn’t about vibes or fragile agents. It’s about making LLMs reliable, testable, and actually useful.</p>

<p>Includes over $2,500 in compute credits and guest lectures from experts at DeepMind, Moderna, and more.<br>
Cohort starts April 7 — <a href="https://maven.com/hugo-stefan/building-llm-apps-ds-and-swe-from-first-principles?promoCode=LLM10" rel="nofollow">Use this link for a 10% discount</a></p>

<h3>🔍 Want to help shape the future of SpecStory?</h3>

<p>Greg and the team are looking for <strong>design partners</strong> for their new SpecStory Teams product—built for collaborative, AI-native software development.</p>

<p>If you&#39;re working with LLMs in a team setting and want to influence the next wave of developer tools, you can apply here:<br><br>
👉 <a href="https://specstory.com/teams" rel="nofollow">specstory.com/teams</a></p>]]>
  </itunes:summary>
</item>
<item>
  <title>Episode 44: The Future of AI Coding Assistants: Who’s Really in Control?</title>
  <link>https://vanishinggradients.fireside.fm/44</link>
  <guid isPermaLink="false">78988fdd-0e05-4e24-82dd-c0a406dd12a1</guid>
  <pubDate>Tue, 04 Feb 2025 13:00:00 +1100</pubDate>
  <author>Hugo Bowne-Anderson</author>
  <enclosure url="https://aphid.fireside.fm/d/1437767933/140c3904-8258-4c39-a698-a112b7077bd7/78988fdd-0e05-4e24-82dd-c0a406dd12a1.mp3" length="90430405" type="audio/mpeg"/>
  <itunes:episodeType>full</itunes:episodeType>
  <itunes:season>1</itunes:season>
  <itunes:author>Hugo Bowne-Anderson</itunes:author>
  <itunes:subtitle>AI coding assistants are reshaping how developers write, debug, and maintain code—but who’s really in control? In this episode, Hugo speaks with **Tyler Dunn**, CEO and co-founder of **Continue**, an open-source AI-powered code assistant that gives developers more customization and flexibility in their workflows.</itunes:subtitle>
  <itunes:duration>1:34:11</itunes:duration>
  <itunes:explicit>no</itunes:explicit>
  <itunes:image href="https://media24.fireside.fm/file/fireside-images-2024/podcasts/images/1/140c3904-8258-4c39-a698-a112b7077bd7/cover.jpg?v=1"/>
  <description>AI coding assistants are reshaping how developers write, debug, and maintain code—but who’s really in control? In this episode, Hugo speaks with Tyler Dunn, CEO and co-founder of Continue, an open-source AI-powered code assistant that gives developers more customization and flexibility in their workflows.
In this episode, we dive into:
- The trade-offs between proprietary vs. open-source AI coding assistants—why open-source might be the future.
- How structured workflows, modular AI, and customization help developers maintain control over their tools.
- The evolution of AI-powered coding, from autocomplete to intelligent code suggestions and beyond.
- Why the best developer experiences come from sensible defaults with room for deeper configuration.
- The future of LLM-based software engineering, where fine-tuning models on personal and team-level data could make AI coding assistants even more effective.
With companies increasingly integrating AI into development workflows, this conversation explores the real impact of these tools—and the importance of keeping developers in the driver's seat.
LINKS
The podcast livestream on YouTube (https://youtube.com/live/8QEgVCzm46U?feature=share)
Continue's website (https://www.continue.dev/)
Continue is hiring! (https://www.continue.dev/about-us)
amplified.dev: We believe in a future where developers are amplified, not automated (https://amplified.dev/)
Beyond Prompt and Pray, Building Reliable LLM-Powered Software in an Agentic World (https://www.oreilly.com/radar/beyond-prompt-and-pray/)
LLMOps Lessons Learned: Navigating the Wild West of Production LLMs 🚀 (https://www.zenml.io/blog/llmops-lessons-learned-navigating-the-wild-west-of-production-llms)
Building effective agents by Erik Schluntz and Barry Zhang, Anthropic (https://www.anthropic.com/research/building-effective-agents)
Ty on LinkedIn (https://www.linkedin.com/in/tylerjdunn/)
Hugo on twitter (https://x.com/hugobowne)
Vanishing Gradients on twitter (https://x.com/vanishingdata)
Vanishing Gradients on YouTube (https://www.youtube.com/channel/UC_NafIo-Ku2loOLrzm45ABA)
Vanishing Gradients on Twitter (https://x.com/vanishingdata)
Vanishing Gradients on Lu.ma (https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk) 
</description>
  <itunes:keywords>data science, machine learning, AI, LLMs</itunes:keywords>
  <content:encoded>
    <![CDATA[<p>AI coding assistants are reshaping how developers write, debug, and maintain code—but who’s really in control? In this episode, Hugo speaks with <strong>Tyler Dunn</strong>, CEO and co-founder of <strong>Continue</strong>, an open-source AI-powered code assistant that gives developers more customization and flexibility in their workflows.</p>

<p>In this episode, we dive into:</p>

<ul>
<li>The trade-offs between <strong>proprietary vs. open-source AI coding assistants</strong>—why open-source might be the future.</li>
<li>How structured workflows, modular AI, and customization help developers maintain <strong>control over their tools</strong>.</li>
<li>The evolution of AI-powered coding, from <strong>autocomplete to intelligent code suggestions</strong> and beyond.</li>
<li>Why the best developer experiences come from <strong>sensible defaults</strong> with room for deeper configuration.</li>
<li>The future of <strong>LLM-based software engineering</strong>, where fine-tuning models on personal and team-level data could make AI coding assistants even more effective.</li>
</ul>

<p>With companies increasingly integrating AI into development workflows, this conversation explores the real impact of these tools—and the importance of keeping developers in the driver&#39;s seat.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://youtube.com/live/8QEgVCzm46U?feature=share" rel="nofollow">The podcast livestream on YouTube</a></li>
<li><a href="https://www.continue.dev/" rel="nofollow">Continue&#39;s website</a></li>
<li><a href="https://www.continue.dev/about-us" rel="nofollow">Continue is hiring!</a></li>
<li><a href="https://amplified.dev/" rel="nofollow">amplified.dev: We believe in a future where developers are amplified, not automated</a></li>
<li><a href="https://www.oreilly.com/radar/beyond-prompt-and-pray/" rel="nofollow">Beyond Prompt and Pray, Building Reliable LLM-Powered Software in an Agentic World</a></li>
<li><a href="https://www.zenml.io/blog/llmops-lessons-learned-navigating-the-wild-west-of-production-llms" rel="nofollow">LLMOps Lessons Learned: Navigating the Wild West of Production LLMs 🚀</a></li>
<li><a href="https://www.anthropic.com/research/building-effective-agents" rel="nofollow">Building effective agents by Erik Schluntz and Barry Zhang, Anthropic</a></li>
<li><a href="https://www.linkedin.com/in/tylerjdunn/" rel="nofollow">Ty on LinkedIn</a></li>
<li><a href="https://x.com/hugobowne" rel="nofollow">Hugo on twitter</a></li>
<li><a href="https://x.com/vanishingdata" rel="nofollow">Vanishing Gradients on twitter</a></li>
<li><a href="https://www.youtube.com/channel/UC_NafIo-Ku2loOLrzm45ABA" rel="nofollow">Vanishing Gradients on YouTube</a></li>
<li><a href="https://x.com/vanishingdata" rel="nofollow">Vanishing Gradients on Twitter</a></li>
<li><a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">Vanishing Gradients on Lu.ma</a></li>
</ul>]]>
  </content:encoded>
  <itunes:summary>
    <![CDATA[<p>AI coding assistants are reshaping how developers write, debug, and maintain code—but who’s really in control? In this episode, Hugo speaks with <strong>Tyler Dunn</strong>, CEO and co-founder of <strong>Continue</strong>, an open-source AI-powered code assistant that gives developers more customization and flexibility in their workflows.</p>

<p>In this episode, we dive into:</p>

<ul>
<li>The trade-offs between <strong>proprietary vs. open-source AI coding assistants</strong>—why open-source might be the future.</li>
<li>How structured workflows, modular AI, and customization help developers maintain <strong>control over their tools</strong>.</li>
<li>The evolution of AI-powered coding, from <strong>autocomplete to intelligent code suggestions</strong> and beyond.</li>
<li>Why the best developer experiences come from <strong>sensible defaults</strong> with room for deeper configuration.</li>
<li>The future of <strong>LLM-based software engineering</strong>, where fine-tuning models on personal and team-level data could make AI coding assistants even more effective.</li>
</ul>

<p>With companies increasingly integrating AI into development workflows, this conversation explores the real impact of these tools—and the importance of keeping developers in the driver&#39;s seat.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://youtube.com/live/8QEgVCzm46U?feature=share" rel="nofollow">The podcast livestream on YouTube</a></li>
<li><a href="https://www.continue.dev/" rel="nofollow">Continue&#39;s website</a></li>
<li><a href="https://www.continue.dev/about-us" rel="nofollow">Continue is hiring!</a></li>
<li><a href="https://amplified.dev/" rel="nofollow">amplified.dev: We believe in a future where developers are amplified, not automated</a></li>
<li><a href="https://www.oreilly.com/radar/beyond-prompt-and-pray/" rel="nofollow">Beyond Prompt and Pray, Building Reliable LLM-Powered Software in an Agentic World</a></li>
<li><a href="https://www.zenml.io/blog/llmops-lessons-learned-navigating-the-wild-west-of-production-llms" rel="nofollow">LLMOps Lessons Learned: Navigating the Wild West of Production LLMs 🚀</a></li>
<li><a href="https://www.anthropic.com/research/building-effective-agents" rel="nofollow">Building effective agents by Erik Schluntz and Barry Zhang, Anthropic</a></li>
<li><a href="https://www.linkedin.com/in/tylerjdunn/" rel="nofollow">Ty on LinkedIn</a></li>
<li><a href="https://x.com/hugobowne" rel="nofollow">Hugo on twitter</a></li>
<li><a href="https://x.com/vanishingdata" rel="nofollow">Vanishing Gradients on twitter</a></li>
<li><a href="https://www.youtube.com/channel/UC_NafIo-Ku2loOLrzm45ABA" rel="nofollow">Vanishing Gradients on YouTube</a></li>
<li><a href="https://x.com/vanishingdata" rel="nofollow">Vanishing Gradients on Twitter</a></li>
<li><a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">Vanishing Gradients on Lu.ma</a></li>
</ul>]]>
  </itunes:summary>
</item>
<item>
  <title>Episode 43: Tales from 400+ LLM Deployments: Building Reliable AI Agents in Production</title>
  <link>https://vanishinggradients.fireside.fm/43</link>
  <guid isPermaLink="false">ff9906ad-8576-40c7-9e0f-26dff301e52c</guid>
  <pubDate>Fri, 17 Jan 2025 08:00:00 +1100</pubDate>
  <author>Hugo Bowne-Anderson</author>
  <enclosure url="https://aphid.fireside.fm/d/1437767933/140c3904-8258-4c39-a698-a112b7077bd7/ff9906ad-8576-40c7-9e0f-26dff301e52c.mp3" length="58615769" type="audio/mpeg"/>
  <itunes:episodeType>full</itunes:episodeType>
  <itunes:season>1</itunes:season>
  <itunes:author>Hugo Bowne-Anderson</itunes:author>
  <itunes:subtitle>Hugo speaks with Alex Strick van Linschoten, Machine Learning Engineer at ZenML and creator of a comprehensive LLMOps database documenting over 400 deployments. Alex's extensive research into real-world LLM implementations gives him unique insight into what actually works—and what doesn't—when deploying AI agents in production.</itunes:subtitle>
  <itunes:duration>1:01:03</itunes:duration>
  <itunes:explicit>no</itunes:explicit>
  <itunes:image href="https://media24.fireside.fm/file/fireside-images-2024/podcasts/images/1/140c3904-8258-4c39-a698-a112b7077bd7/cover.jpg?v=1"/>
  <description>Hugo speaks with Alex Strick van Linschoten, Machine Learning Engineer at ZenML and creator of a comprehensive LLMOps database documenting over 400 deployments. Alex's extensive research into real-world LLM implementations gives him unique insight into what actually works—and what doesn't—when deploying AI agents in production.
In this episode, we dive into:
- The current state of AI agents in production, from successes to common failure modes
- Practical lessons learned from analyzing hundreds of real-world LLM deployments
- How companies like Anthropic, Klarna, and Dropbox are using patterns like ReAct, RAG, and microservices to build reliable systems
- The evolution of LLM capabilities, from expanding context windows to multimodal applications
- Why most companies still prefer structured workflows over fully autonomous agents
We also explore real-world case studies of production hurdles, including cascading failures, API misfires, and hallucination challenges. Alex shares concrete strategies for integrating LLMs into your pipelines while maintaining reliability and control.
Whether you're scaling agents or building LLM-powered systems, this episode offers practical insights for navigating the complex landscape of LLMOps in 2025.
LINKS
The podcast livestream on YouTube (https://youtube.com/live/-8Gr9fVVX9g?feature=share)
The LLMOps database (https://www.zenml.io/llmops-database)
All blog posts about the database (https://www.zenml.io/category/llmops)
Anthropic's Building effective agents essay (https://www.anthropic.com/research/building-effective-agents)
Alex on LinkedIn (https://www.linkedin.com/in/strickvl/)
Hugo on twitter (https://x.com/hugobowne)
Vanishing Gradients on twitter (https://x.com/vanishingdata)
Vanishing Gradients on YouTube (https://www.youtube.com/channel/UC_NafIo-Ku2loOLrzm45ABA)
Vanishing Gradients on Twitter (https://x.com/vanishingdata)
Vanishing Gradients on Lu.ma (https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk)
</description>
  <itunes:keywords>data science, machine learning, AI, LLMs</itunes:keywords>
  <content:encoded>
    <![CDATA[<p>Hugo speaks with Alex Strick van Linschoten, Machine Learning Engineer at ZenML and creator of a comprehensive LLMOps database documenting over 400 deployments. Alex&#39;s extensive research into real-world LLM implementations gives him unique insight into what actually works—and what doesn&#39;t—when deploying AI agents in production.</p>

<p>In this episode, we dive into:</p>

<ul>
<li>The current state of AI agents in production, from successes to common failure modes</li>
<li>Practical lessons learned from analyzing hundreds of real-world LLM deployments</li>
<li>How companies like Anthropic, Klarna, and Dropbox are using patterns like ReAct, RAG, and microservices to build reliable systems</li>
<li>The evolution of LLM capabilities, from expanding context windows to multimodal applications</li>
<li>Why most companies still prefer structured workflows over fully autonomous agents</li>
</ul>

<p>We also explore real-world case studies of production hurdles, including cascading failures, API misfires, and hallucination challenges. Alex shares concrete strategies for integrating LLMs into your pipelines while maintaining reliability and control.</p>

<p>Whether you&#39;re scaling agents or building LLM-powered systems, this episode offers practical insights for navigating the complex landscape of LLMOps in 2025.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://youtube.com/live/-8Gr9fVVX9g?feature=share" rel="nofollow">The podcast livestream on YouTube</a></li>
<li><a href="https://www.zenml.io/llmops-database" rel="nofollow">The LLMOps database</a></li>
<li><a href="https://www.zenml.io/category/llmops" rel="nofollow">All blog posts about the database</a></li>
<li><a href="https://www.anthropic.com/research/building-effective-agents" rel="nofollow">Anthropic&#39;s Building effective agents essay</a></li>
<li><a href="https://www.linkedin.com/in/strickvl/" rel="nofollow">Alex on LinkedIn</a></li>
<li><a href="https://x.com/hugobowne" rel="nofollow">Hugo on twitter</a></li>
<li><a href="https://x.com/vanishingdata" rel="nofollow">Vanishing Gradients on twitter</a></li>
<li><a href="https://www.youtube.com/channel/UC_NafIo-Ku2loOLrzm45ABA" rel="nofollow">Vanishing Gradients on YouTube</a></li>
<li><a href="https://x.com/vanishingdata" rel="nofollow">Vanishing Gradients on Twitter</a></li>
<li><a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">Vanishing Gradients on Lu.ma</a></li>
</ul>]]>
  </content:encoded>
  <itunes:summary>
    <![CDATA[<p>Hugo speaks with Alex Strick van Linschoten, Machine Learning Engineer at ZenML and creator of a comprehensive LLMOps database documenting over 400 deployments. Alex&#39;s extensive research into real-world LLM implementations gives him unique insight into what actually works—and what doesn&#39;t—when deploying AI agents in production.</p>

<p>In this episode, we dive into:</p>

<ul>
<li>The current state of AI agents in production, from successes to common failure modes</li>
<li>Practical lessons learned from analyzing hundreds of real-world LLM deployments</li>
<li>How companies like Anthropic, Klarna, and Dropbox are using patterns like ReAct, RAG, and microservices to build reliable systems</li>
<li>The evolution of LLM capabilities, from expanding context windows to multimodal applications</li>
<li>Why most companies still prefer structured workflows over fully autonomous agents</li>
</ul>

<p>We also explore real-world case studies of production hurdles, including cascading failures, API misfires, and hallucination challenges. Alex shares concrete strategies for integrating LLMs into your pipelines while maintaining reliability and control.</p>

<p>Whether you&#39;re scaling agents or building LLM-powered systems, this episode offers practical insights for navigating the complex landscape of LLMOps in 2025.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://youtube.com/live/-8Gr9fVVX9g?feature=share" rel="nofollow">The podcast livestream on YouTube</a></li>
<li><a href="https://www.zenml.io/llmops-database" rel="nofollow">The LLMOps database</a></li>
<li><a href="https://www.zenml.io/category/llmops" rel="nofollow">All blog posts about the database</a></li>
<li><a href="https://www.anthropic.com/research/building-effective-agents" rel="nofollow">Anthropic&#39;s Building effective agents essay</a></li>
<li><a href="https://www.linkedin.com/in/strickvl/" rel="nofollow">Alex on LinkedIn</a></li>
<li><a href="https://x.com/hugobowne" rel="nofollow">Hugo on twitter</a></li>
<li><a href="https://x.com/vanishingdata" rel="nofollow">Vanishing Gradients on twitter</a></li>
<li><a href="https://www.youtube.com/channel/UC_NafIo-Ku2loOLrzm45ABA" rel="nofollow">Vanishing Gradients on YouTube</a></li>
<li><a href="https://x.com/vanishingdata" rel="nofollow">Vanishing Gradients on Twitter</a></li>
<li><a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">Vanishing Gradients on Lu.ma</a></li>
</ul>]]>
  </itunes:summary>
</item>
<item>
  <title>Episode 42: Learning, Teaching, and Building in the Age of AI</title>
  <link>https://vanishinggradients.fireside.fm/42</link>
  <guid isPermaLink="false">6af2e172-b72b-418b-baa6-369299f37b8b</guid>
  <pubDate>Sat, 04 Jan 2025 14:00:00 +1100</pubDate>
  <author>Hugo Bowne-Anderson</author>
  <enclosure url="https://aphid.fireside.fm/d/1437767933/140c3904-8258-4c39-a698-a112b7077bd7/6af2e172-b72b-418b-baa6-369299f37b8b.mp3" length="76860106" type="audio/mpeg"/>
  <itunes:episodeType>full</itunes:episodeType>
  <itunes:season>1</itunes:season>
  <itunes:author>Hugo Bowne-Anderson</itunes:author>
  <itunes:subtitle>The tables turn as Hugo sits down with Alex Andorra, host of Learning Bayesian Statistics. Hugo shares his journey from mathematics to AI, reflecting on how Bayesian inference shapes his approach to data science, teaching, and building AI-powered applications.</itunes:subtitle>
  <itunes:duration>1:20:03</itunes:duration>
  <itunes:explicit>no</itunes:explicit>
  <itunes:image href="https://media24.fireside.fm/file/fireside-images-2024/podcasts/images/1/140c3904-8258-4c39-a698-a112b7077bd7/cover.jpg?v=1"/>
  <description>In this episode of Vanishing Gradients, the tables turn as Hugo sits down with Alex Andorra, host of Learning Bayesian Statistics. Hugo shares his journey from mathematics to AI, reflecting on how Bayesian inference shapes his approach to data science, teaching, and building AI-powered applications.
They dive into the realities of deploying LLM applications, overcoming “proof-of-concept purgatory,” and why first principles and iteration are critical for success in AI. Whether you’re an educator, software engineer, or data scientist, this episode offers valuable insights into the intersection of AI, product development, and real-world deployment.
LINKS
The podcast on YouTube (https://www.youtube.com/watch?v=BRIYytbqtP0)
The original podcast episode (https://learnbayesstats.com/episode/122-learning-and-teaching-in-the-age-of-ai-hugo-bowne-anderson)
Alex Andorra on LinkedIn (https://www.linkedin.com/in/alex-andorra/)
Hugo on LinkedIn (https://www.linkedin.com/in/hugo-bowne-anderson-045939a5/)
Hugo on twitter (https://x.com/hugobowne)
Vanishing Gradients on twitter (https://x.com/vanishingdata)
Hugo's "Building LLM Applications for Data Scientists and Software Engineers" course (https://maven.com/s/course/d56067f338) 
</description>
  <itunes:keywords>data science, machine learning, AI, LLMs</itunes:keywords>
  <content:encoded>
    <![CDATA[<p>In this episode of Vanishing Gradients, the tables turn as Hugo sits down with Alex Andorra, host of Learning Bayesian Statistics. Hugo shares his journey from mathematics to AI, reflecting on how Bayesian inference shapes his approach to data science, teaching, and building AI-powered applications.</p>

<p>They dive into the realities of deploying LLM applications, overcoming “proof-of-concept purgatory,” and why first principles and iteration are critical for success in AI. Whether you’re an educator, software engineer, or data scientist, this episode offers valuable insights into the intersection of AI, product development, and real-world deployment.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://www.youtube.com/watch?v=BRIYytbqtP0" rel="nofollow">The podcast on YouTube</a></li>
<li><a href="https://learnbayesstats.com/episode/122-learning-and-teaching-in-the-age-of-ai-hugo-bowne-anderson" rel="nofollow">The original podcast episode</a></li>
<li><a href="https://www.linkedin.com/in/alex-andorra/" rel="nofollow">Alex Andorra on LinkedIn</a></li>
<li><a href="https://www.linkedin.com/in/hugo-bowne-anderson-045939a5/" rel="nofollow">Hugo on LinkedIn</a></li>
<li><a href="https://x.com/hugobowne" rel="nofollow">Hugo on twitter</a></li>
<li><a href="https://x.com/vanishingdata" rel="nofollow">Vanishing Gradients on twitter</a></li>
<li><a href="https://maven.com/s/course/d56067f338" rel="nofollow">Hugo&#39;s &quot;Building LLM Applications for Data Scientists and Software Engineers&quot; course</a></li>
</ul>]]>
  </content:encoded>
  <itunes:summary>
    <![CDATA[<p>In this episode of Vanishing Gradients, the tables turn as Hugo sits down with Alex Andorra, host of Learning Bayesian Statistics. Hugo shares his journey from mathematics to AI, reflecting on how Bayesian inference shapes his approach to data science, teaching, and building AI-powered applications.</p>

<p>They dive into the realities of deploying LLM applications, overcoming “proof-of-concept purgatory,” and why first principles and iteration are critical for success in AI. Whether you’re an educator, software engineer, or data scientist, this episode offers valuable insights into the intersection of AI, product development, and real-world deployment.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://www.youtube.com/watch?v=BRIYytbqtP0" rel="nofollow">The podcast on YouTube</a></li>
<li><a href="https://learnbayesstats.com/episode/122-learning-and-teaching-in-the-age-of-ai-hugo-bowne-anderson" rel="nofollow">The original podcast episode</a></li>
<li><a href="https://www.linkedin.com/in/alex-andorra/" rel="nofollow">Alex Andorra on LinkedIn</a></li>
<li><a href="https://www.linkedin.com/in/hugo-bowne-anderson-045939a5/" rel="nofollow">Hugo on LinkedIn</a></li>
<li><a href="https://x.com/hugobowne" rel="nofollow">Hugo on twitter</a></li>
<li><a href="https://x.com/vanishingdata" rel="nofollow">Vanishing Gradients on twitter</a></li>
<li><a href="https://maven.com/s/course/d56067f338" rel="nofollow">Hugo&#39;s &quot;Building LLM Applications for Data Scientists and Software Engineers&quot; course</a></li>
</ul>]]>
  </itunes:summary>
</item>
<item>
  <title>Episode 41: Beyond Prompt Engineering: Can AI Learn to Set Its Own Goals?</title>
  <link>https://vanishinggradients.fireside.fm/41</link>
  <guid isPermaLink="false">695d8cc9-b111-4f1d-9871-82962ae023f4</guid>
  <pubDate>Tue, 31 Dec 2024 10:00:00 +1100</pubDate>
  <author>Hugo Bowne-Anderson</author>
  <enclosure url="https://aphid.fireside.fm/d/1437767933/140c3904-8258-4c39-a698-a112b7077bd7/695d8cc9-b111-4f1d-9871-82962ae023f4.mp3" length="42114740" type="audio/mpeg"/>
  <itunes:episodeType>full</itunes:episodeType>
  <itunes:season>1</itunes:season>
  <itunes:author>Hugo Bowne-Anderson</itunes:author>
  <itunes:subtitle>Hugo Bowne-Anderson hosts a panel discussion from the MLOps World and Generative AI Summit in Austin, exploring the long-term growth of AI by distinguishing real problem-solving from trend-based solutions. If you're navigating the evolving landscape of generative AI, productionizing models, or questioning the hype, this episode dives into the tough questions shaping the field.</itunes:subtitle>
  <itunes:duration>43:51</itunes:duration>
  <itunes:explicit>yes</itunes:explicit>
  <itunes:image href="https://media24.fireside.fm/file/fireside-images-2024/podcasts/images/1/140c3904-8258-4c39-a698-a112b7077bd7/cover.jpg?v=1"/>
  <description>Hugo Bowne-Anderson hosts a panel discussion from the MLOps World and Generative AI Summit in Austin, exploring the long-term growth of AI by distinguishing real problem-solving from trend-based solutions. If you're navigating the evolving landscape of generative AI, productionizing models, or questioning the hype, this episode dives into the tough questions shaping the field.
The panel features:  
- Ben Taylor (Jepson) (https://www.linkedin.com/in/jepsontaylor/) – CEO and Founder at VEOX Inc., with experience in AI exploration, genetic programming, and deep learning.  
- Joe Reis (https://www.linkedin.com/in/josephreis/) – Co-founder of Ternary Data and author of Fundamentals of Data Engineering.  
- Juan Sequeda (https://www.linkedin.com/in/juansequeda/) – Principal Scientist and Head of AI Lab at Data.World, known for his expertise in knowledge graphs and the semantic web.  
The discussion unpacks essential topics such as:  
- The shift from prompt engineering to goal engineering—letting AI iterate toward well-defined objectives.  
- Whether generative AI is having an electricity moment or more of a blockchain trajectory.  
- The combinatorial power of AI to explore new solutions, drawing parallels to AlphaZero redefining strategy games.  
- The POC-to-production gap and why AI projects stall.  
- Failure modes, hallucinations, and governance risks—and how to mitigate them.  
- The disconnect between executive optimism and employee workload.  
Hugo also mentions his upcoming workshop on escaping Proof-of-Concept Purgatory, which has evolved into a Maven course "Building LLM Applications for Data Scientists and Software Engineers" launching in January (https://maven.com/hugo-stefan/building-llm-apps-ds-and-swe-from-first-principles?utm_campaign=8123d0&amp;amp;utm_medium=partner&amp;amp;utm_source=instructor). Vanishing Gradient listeners can get 25% off the course (use the code VG25), with $1,000 in Modal compute credits included.
A huge thanks to Dave Scharbach and the Toronto Machine Learning Society for organizing the conference and to the audience for their thoughtful questions.
As we head into the new year, this conversation offers a reality check amidst the growing AI agent hype.  
LINKS
Hugo on twitter (https://x.com/hugobowne)
Hugo on LinkedIn (https://www.linkedin.com/in/hugo-bowne-anderson-045939a5/)
Vanishing Gradients on twitter (https://x.com/vanishingdata)
"Building LLM Applications for Data Scientists and Software Engineers" course (https://maven.com/hugo-stefan/building-llm-apps-ds-and-swe-from-first-principles?utm_campaign=8123d0&amp;amp;utm_medium=partner&amp;amp;utm_source=instructor).
</description>
  <itunes:keywords>data science, machine learning, AI, LLMs</itunes:keywords>
  <content:encoded>
    <![CDATA[<p>Hugo Bowne-Anderson hosts a panel discussion from the MLOps World and Generative AI Summit in Austin, exploring the long-term growth of AI by distinguishing real problem-solving from trend-based solutions. If you&#39;re navigating the evolving landscape of generative AI, productionizing models, or questioning the hype, this episode dives into the tough questions shaping the field.</p>

<p>The panel features:  </p>

<ul>
<li><a href="https://www.linkedin.com/in/jepsontaylor/" rel="nofollow"><strong>Ben Taylor (Jepson)</strong></a> – CEO and Founder at VEOX Inc., with experience in AI exploration, genetic programming, and deep learning.<br></li>
<li><a href="https://www.linkedin.com/in/josephreis/" rel="nofollow"><strong>Joe Reis</strong></a> – Co-founder of Ternary Data and author of <em>Fundamentals of Data Engineering</em>.<br></li>
<li><a href="https://www.linkedin.com/in/juansequeda/" rel="nofollow"><strong>Juan Sequeda</strong></a> – Principal Scientist and Head of AI Lab at Data.World, known for his expertise in knowledge graphs and the semantic web.<br></li>
</ul>

<p>The discussion unpacks essential topics such as:  </p>

<ul>
<li>The shift from <strong>prompt engineering</strong> to <strong>goal engineering</strong>—letting AI iterate toward well-defined objectives.<br></li>
<li>Whether generative AI is having an <strong>electricity moment</strong> or more of a <strong>blockchain trajectory</strong>.<br></li>
<li>The <strong>combinatorial power of AI</strong> to explore new solutions, drawing parallels to AlphaZero redefining strategy games.<br></li>
<li>The <strong>POC-to-production gap</strong> and why AI projects stall.<br></li>
<li><strong>Failure modes, hallucinations, and governance risks</strong>—and how to mitigate them.<br></li>
<li>The disconnect between executive optimism and employee workload.<br></li>
</ul>

<p>Hugo also mentions his upcoming workshop on <strong>escaping Proof-of-Concept Purgatory</strong>, <a href="https://maven.com/hugo-stefan/building-llm-apps-ds-and-swe-from-first-principles?utm_campaign=8123d0&utm_medium=partner&utm_source=instructor" rel="nofollow">which has evolved into a <strong>Maven course &quot;Building LLM Applications for Data Scientists and Software Engineers&quot; launching in January</strong></a>. Vanishing Gradient listeners can get 25% off the course (use the code VG25), with $1,000 in Modal compute credits included.</p>

<p>A huge thanks to <strong>Dave Scharbach and the Toronto Machine Learning Society</strong> for organizing the conference and to the audience for their thoughtful questions.</p>

<p>As we head into the new year, this conversation offers a reality check amidst the growing AI agent hype.  </p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://x.com/hugobowne" rel="nofollow">Hugo on twitter</a></li>
<li><a href="https://www.linkedin.com/in/hugo-bowne-anderson-045939a5/" rel="nofollow">Hugo on LinkedIn</a></li>
<li><a href="https://x.com/vanishingdata" rel="nofollow">Vanishing Gradients on twitter</a></li>
<li><a href="https://maven.com/hugo-stefan/building-llm-apps-ds-and-swe-from-first-principles?utm_campaign=8123d0&utm_medium=partner&utm_source=instructor" rel="nofollow">&quot;Building LLM Applications for Data Scientists and Software Engineers&quot; course</a>.</li>
</ul>]]>
  </content:encoded>
  <itunes:summary>
    <![CDATA[<p>Hugo Bowne-Anderson hosts a panel discussion from the MLOps World and Generative AI Summit in Austin, exploring the long-term growth of AI by distinguishing real problem-solving from trend-based solutions. If you&#39;re navigating the evolving landscape of generative AI, productionizing models, or questioning the hype, this episode dives into the tough questions shaping the field.</p>

<p>The panel features:  </p>

<ul>
<li><a href="https://www.linkedin.com/in/jepsontaylor/" rel="nofollow"><strong>Ben Taylor (Jepson)</strong></a> – CEO and Founder at VEOX Inc., with experience in AI exploration, genetic programming, and deep learning.<br></li>
<li><a href="https://www.linkedin.com/in/josephreis/" rel="nofollow"><strong>Joe Reis</strong></a> – Co-founder of Ternary Data and author of <em>Fundamentals of Data Engineering</em>.<br></li>
<li><a href="https://www.linkedin.com/in/juansequeda/" rel="nofollow"><strong>Juan Sequeda</strong></a> – Principal Scientist and Head of AI Lab at Data.World, known for his expertise in knowledge graphs and the semantic web.<br></li>
</ul>

<p>The discussion unpacks essential topics such as:  </p>

<ul>
<li>The shift from <strong>prompt engineering</strong> to <strong>goal engineering</strong>—letting AI iterate toward well-defined objectives.<br></li>
<li>Whether generative AI is having an <strong>electricity moment</strong> or more of a <strong>blockchain trajectory</strong>.<br></li>
<li>The <strong>combinatorial power of AI</strong> to explore new solutions, drawing parallels to AlphaZero redefining strategy games.<br></li>
<li>The <strong>POC-to-production gap</strong> and why AI projects stall.<br></li>
<li><strong>Failure modes, hallucinations, and governance risks</strong>—and how to mitigate them.<br></li>
<li>The disconnect between executive optimism and employee workload.<br></li>
</ul>

<p>Hugo also mentions his upcoming workshop on <strong>escaping Proof-of-Concept Purgatory</strong>, <a href="https://maven.com/hugo-stefan/building-llm-apps-ds-and-swe-from-first-principles?utm_campaign=8123d0&utm_medium=partner&utm_source=instructor" rel="nofollow">which has evolved into a <strong>Maven course &quot;Building LLM Applications for Data Scientists and Software Engineers&quot; launching in January</strong></a>. Vanishing Gradient listeners can get 25% off the course (use the code VG25), with $1,000 in Modal compute credits included.</p>

<p>A huge thanks to <strong>Dave Scharbach and the Toronto Machine Learning Society</strong> for organizing the conference and to the audience for their thoughtful questions.</p>

<p>As we head into the new year, this conversation offers a reality check amidst the growing AI agent hype.  </p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://x.com/hugobowne" rel="nofollow">Hugo on twitter</a></li>
<li><a href="https://www.linkedin.com/in/hugo-bowne-anderson-045939a5/" rel="nofollow">Hugo on LinkedIn</a></li>
<li><a href="https://x.com/vanishingdata" rel="nofollow">Vanishing Gradients on twitter</a></li>
<li><a href="https://maven.com/hugo-stefan/building-llm-apps-ds-and-swe-from-first-principles?utm_campaign=8123d0&utm_medium=partner&utm_source=instructor" rel="nofollow">&quot;Building LLM Applications for Data Scientists and Software Engineers&quot; course</a>.</li>
</ul>]]>
  </itunes:summary>
</item>
<item>
  <title>Episode 40: What Every LLM Developer Needs to Know About GPUs</title>
  <link>https://vanishinggradients.fireside.fm/40</link>
  <guid isPermaLink="false">b1b66484-5fd0-4bcb-91cb-8bf7201a5ded</guid>
  <pubDate>Tue, 24 Dec 2024 15:00:00 +1100</pubDate>
  <author>Hugo Bowne-Anderson</author>
  <enclosure url="https://aphid.fireside.fm/d/1437767933/140c3904-8258-4c39-a698-a112b7077bd7/b1b66484-5fd0-4bcb-91cb-8bf7201a5ded.mp3" length="99441605" type="audio/mpeg"/>
  <itunes:episodeType>full</itunes:episodeType>
  <itunes:season>1</itunes:season>
  <itunes:author>Hugo Bowne-Anderson</itunes:author>
  <itunes:subtitle>Hugo speaks with **Charles Frye**, Developer Advocate at Modal and someone who really knows GPUs inside and out. If you’re a data scientist, machine learning engineer, AI researcher, or just someone trying to make sense of **hardware for LLMs and AI workflows**, this episode is for you.  

Charles and Hugo dive into the **practical side of GPUs**—from **running inference** on large models, to **fine-tuning** and even **training from scratch.** </itunes:subtitle>
  <itunes:duration>1:43:34</itunes:duration>
  <itunes:explicit>no</itunes:explicit>
  <itunes:image href="https://media24.fireside.fm/file/fireside-images-2024/podcasts/images/1/140c3904-8258-4c39-a698-a112b7077bd7/cover.jpg?v=1"/>
  <description>Hugo speaks with Charles Frye, Developer Advocate at Modal and someone who really knows GPUs inside and out. If you’re a data scientist, machine learning engineer, AI researcher, or just someone trying to make sense of hardware for LLMs and AI workflows, this episode is for you.  
Charles and Hugo dive into the practical side of GPUs—from running inference on large models, to fine-tuning and even training from scratch. They unpack the real pain points developers face, like figuring out:  
- How much VRAM you actually need.  
- Why memory—not compute—ends up being the bottleneck.  
- How to make quick, back-of-the-envelope calculations to size up hardware for your tasks.  
- And where things like fine-tuning, quantization, and retrieval-augmented generation (RAG) fit into the mix.  
One thing Hugo really appreciate is that Charles and the Modal team recently put together the GPU Glossary—a resource that breaks down GPU internals in a way that’s actually useful for developers. We reference it a few times throughout the episode, so check it out in the show notes below.  
🔧 Charles also does a demo during the episode—some of it is visual, but we talk through the key points so you’ll still get value from the audio. If you’d like to see the demo in action, check out the livestream linked below.
This is the "Building LLM Applications for Data Scientists and Software Engineers" course that Hugo is teaching with Stefan Krawczyk (ex-StitchFix) in January (https://maven.com/s/course/d56067f338). Charles is giving a guest lecture at on hardware for LLMs, and Modal is giving all students $1K worth of compute credits (use the code VG25 for $200 off).
LINKS
The livestream on YouTube (https://www.youtube.com/live/INryb8Hjk3c?si=0cbb0-Nxem1P987d)
The GPU Glossary (https://modal.com/gpu-glossary) by the Modal team
What We’ve Learned From A Year of Building with LLMs (https://applied-llms.org/) by Charles and friends
Charles on twitter (https://x.com/charles_irl)
Hugo on twitter (https://x.com/hugobowne)
Vanishing Gradients on twitter (https://x.com/vanishingdata)
</description>
  <itunes:keywords>data science, machine learning, AI, LLMs</itunes:keywords>
  <content:encoded>
    <![CDATA[<p>Hugo speaks with <strong>Charles Frye</strong>, Developer Advocate at Modal and someone who really knows GPUs inside and out. If you’re a data scientist, machine learning engineer, AI researcher, or just someone trying to make sense of <strong>hardware for LLMs and AI workflows</strong>, this episode is for you.  </p>

<p>Charles and Hugo dive into the <strong>practical side of GPUs</strong>—from <strong>running inference</strong> on large models, to <strong>fine-tuning</strong> and even <strong>training from scratch.</strong> They unpack the <strong>real pain points</strong> developers face, like figuring out:  </p>

<ul>
<li>How much VRAM you actually need.<br></li>
<li>Why memory—not compute—ends up being the bottleneck.<br></li>
<li>How to make quick, <strong>back-of-the-envelope calculations</strong> to size up hardware for your tasks.<br></li>
<li>And where things like <strong>fine-tuning, quantization, and retrieval-augmented generation (RAG)</strong> fit into the mix.<br></li>
</ul>

<p>One thing Hugo really appreciate is that Charles and the Modal team recently put together the <strong>GPU Glossary</strong>—a resource that breaks down GPU internals in a way that’s actually useful for developers. We reference it a few times throughout the episode, so check it out in the show notes below.  </p>

<p>🔧 <strong>Charles also does a demo during the episode</strong>—some of it is visual, but we talk through the key points so you’ll still get value from the audio. If you’d like to see the demo in action, check out the livestream linked below.</p>

<p><a href="https://maven.com/s/course/d56067f338" rel="nofollow">This is the &quot;Building LLM Applications for Data Scientists and Software Engineers&quot; course that Hugo is teaching with Stefan Krawczyk (ex-StitchFix) in January</a>. Charles is giving a guest lecture at on hardware for LLMs, and Modal is giving all students $1K worth of compute credits (use the code VG25 for $200 off).</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://www.youtube.com/live/INryb8Hjk3c?si=0cbb0-Nxem1P987d" rel="nofollow">The livestream on YouTube</a></li>
<li><a href="https://modal.com/gpu-glossary" rel="nofollow">The GPU Glossary</a> by the Modal team</li>
<li><a href="https://applied-llms.org/" rel="nofollow">What We’ve Learned From A Year of Building with LLMs</a> by Charles and friends</li>
<li><a href="https://x.com/charles_irl" rel="nofollow">Charles on twitter</a></li>
<li><a href="https://x.com/hugobowne" rel="nofollow">Hugo on twitter</a></li>
<li><a href="https://x.com/vanishingdata" rel="nofollow">Vanishing Gradients on twitter</a></li>
</ul>]]>
  </content:encoded>
  <itunes:summary>
    <![CDATA[<p>Hugo speaks with <strong>Charles Frye</strong>, Developer Advocate at Modal and someone who really knows GPUs inside and out. If you’re a data scientist, machine learning engineer, AI researcher, or just someone trying to make sense of <strong>hardware for LLMs and AI workflows</strong>, this episode is for you.  </p>

<p>Charles and Hugo dive into the <strong>practical side of GPUs</strong>—from <strong>running inference</strong> on large models, to <strong>fine-tuning</strong> and even <strong>training from scratch.</strong> They unpack the <strong>real pain points</strong> developers face, like figuring out:  </p>

<ul>
<li>How much VRAM you actually need.<br></li>
<li>Why memory—not compute—ends up being the bottleneck.<br></li>
<li>How to make quick, <strong>back-of-the-envelope calculations</strong> to size up hardware for your tasks.<br></li>
<li>And where things like <strong>fine-tuning, quantization, and retrieval-augmented generation (RAG)</strong> fit into the mix.<br></li>
</ul>

<p>One thing Hugo really appreciate is that Charles and the Modal team recently put together the <strong>GPU Glossary</strong>—a resource that breaks down GPU internals in a way that’s actually useful for developers. We reference it a few times throughout the episode, so check it out in the show notes below.  </p>

<p>🔧 <strong>Charles also does a demo during the episode</strong>—some of it is visual, but we talk through the key points so you’ll still get value from the audio. If you’d like to see the demo in action, check out the livestream linked below.</p>

<p><a href="https://maven.com/s/course/d56067f338" rel="nofollow">This is the &quot;Building LLM Applications for Data Scientists and Software Engineers&quot; course that Hugo is teaching with Stefan Krawczyk (ex-StitchFix) in January</a>. Charles is giving a guest lecture at on hardware for LLMs, and Modal is giving all students $1K worth of compute credits (use the code VG25 for $200 off).</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://www.youtube.com/live/INryb8Hjk3c?si=0cbb0-Nxem1P987d" rel="nofollow">The livestream on YouTube</a></li>
<li><a href="https://modal.com/gpu-glossary" rel="nofollow">The GPU Glossary</a> by the Modal team</li>
<li><a href="https://applied-llms.org/" rel="nofollow">What We’ve Learned From A Year of Building with LLMs</a> by Charles and friends</li>
<li><a href="https://x.com/charles_irl" rel="nofollow">Charles on twitter</a></li>
<li><a href="https://x.com/hugobowne" rel="nofollow">Hugo on twitter</a></li>
<li><a href="https://x.com/vanishingdata" rel="nofollow">Vanishing Gradients on twitter</a></li>
</ul>]]>
  </itunes:summary>
</item>
<item>
  <title>Episode 39: From Models to Products: Bridging Research and Practice in Generative AI at Google Labs</title>
  <link>https://vanishinggradients.fireside.fm/39</link>
  <guid isPermaLink="false">bf5453c0-4aa2-4abb-b323-20334f787512</guid>
  <pubDate>Tue, 26 Nov 2024 03:00:00 +1100</pubDate>
  <author>Hugo Bowne-Anderson</author>
  <enclosure url="https://aphid.fireside.fm/d/1437767933/140c3904-8258-4c39-a698-a112b7077bd7/bf5453c0-4aa2-4abb-b323-20334f787512.mp3" length="99346310" type="audio/mpeg"/>
  <itunes:episodeType>full</itunes:episodeType>
  <itunes:season>1</itunes:season>
  <itunes:author>Hugo Bowne-Anderson</itunes:author>
  <itunes:subtitle>From building rockets at SpaceX to advancing generative AI at Google Labs, Ravin Kumar has carved a unique path through the world of technology. In this episode, we explore how to build scalable, reliable AI systems, the skills needed to work across the AI/ML pipeline, and the real-world impact of tools like open-weight models such as Gemma. Ravin also shares insights into designing AI tools like Notebook LM with the user journey at the forefront.</itunes:subtitle>
  <itunes:duration>1:43:28</itunes:duration>
  <itunes:explicit>no</itunes:explicit>
  <itunes:image href="https://media24.fireside.fm/file/fireside-images-2024/podcasts/images/1/140c3904-8258-4c39-a698-a112b7077bd7/cover.jpg?v=1"/>
  <description>Hugo speaks with Ravin Kumar,*Senior Research Data Scientist at Google Labs. Ravin’s career has taken him from building rockets at SpaceX to driving data science and technology at Sweetgreen, and now to advancing generative AI research and applications at Google Labs and DeepMind. His multidisciplinary experience gives him a rare perspective on building AI systems that combine technical rigor with practical utility.
In this episode, we dive into:
    • Ravin’s fascinating career path, including the skills and mindsets needed to work effectively with AI and machine learning models at different stages of the pipeline.
    • How to build generative AI systems that are scalable, reliable, and aligned with user needs.
    • Real-world applications of generative AI, such as using open weight models such as Gemma to help a bakery streamline operations—an example of delivering tangible business value through AI.
    • The critical role of UX in AI adoption, and how Ravin approaches designing tools like Notebook LM with the user journey in mind.
We also include a live demo where Ravin uses Notebook LM to analyze my website, extract insights, and even generate a podcast-style conversation about me. While some of the demo is visual, much can be appreciated through audio, and we’ve added a link to the video in the show notes for those who want to see it in action. We’ve also included the generated segment at the end of the episode for you to enjoy.
LINKS
The livestream on YouTube (https://www.youtube.com/live/ffS6NWqoo_k)
Google Labs (https://labs.google/)
Ravin's GenAI Handbook (https://ravinkumar.com/GenAiGuidebook/book_intro.html)
Breadboard: A library for prototyping generative AI applications (https://breadboard-ai.github.io/breadboard/)
As mentioned in the episode, Hugo is teaching a four-week course, Building LLM Applications for Data Scientists and SWEs, co-led with Stefan Krawczyk (Dagworks, ex-StitchFix). The course focuses on building scalable, production-grade generative AI systems, with hands-on sessions, $1,000+ in cloud credits, live Q&amp;amp;As, and guest lectures from industry experts.
Listeners of Vanishing Gradients can get 25% off the course using this special link (https://maven.com/hugo-stefan/building-llm-apps-ds-and-swe-from-first-principles?promoCode=VG25) or by applying the code VG25 at checkout.
</description>
  <itunes:keywords>data science, machine learning, AI, LLMs</itunes:keywords>
  <content:encoded>
    <![CDATA[<p>Hugo speaks with Ravin Kumar,*Senior Research Data Scientist at Google Labs. Ravin’s career has taken him from building rockets at SpaceX to driving data science and technology at Sweetgreen, and now to advancing generative AI research and applications at Google Labs and DeepMind. His multidisciplinary experience gives him a rare perspective on building AI systems that combine technical rigor with practical utility.</p>

<p>In this episode, we dive into:<br>
    • Ravin’s fascinating career path, including the skills and mindsets needed to work effectively with AI and machine learning models at different stages of the pipeline.<br>
    • How to build generative AI systems that are scalable, reliable, and aligned with user needs.<br>
    • Real-world applications of generative AI, such as using open weight models such as Gemma to help a bakery streamline operations—an example of delivering tangible business value through AI.<br>
    • The critical role of UX in AI adoption, and how Ravin approaches designing tools like Notebook LM with the user journey in mind.</p>

<p>We also include a live demo where Ravin uses Notebook LM to analyze my website, extract insights, and even generate a podcast-style conversation about me. While some of the demo is visual, much can be appreciated through audio, and we’ve added a link to the video in the show notes for those who want to see it in action. We’ve also included the generated segment at the end of the episode for you to enjoy.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://www.youtube.com/live/ffS6NWqoo_k" rel="nofollow">The livestream on YouTube</a></li>
<li><a href="https://labs.google/" rel="nofollow">Google Labs</a></li>
<li><a href="https://ravinkumar.com/GenAiGuidebook/book_intro.html" rel="nofollow">Ravin&#39;s GenAI Handbook</a></li>
<li><a href="https://breadboard-ai.github.io/breadboard/" rel="nofollow">Breadboard: A library for prototyping generative AI applications</a></li>
</ul>

<p>As mentioned in the episode, Hugo is teaching a four-week course, <strong>Building LLM Applications for Data Scientists and SWEs</strong>, co-led with Stefan Krawczyk (Dagworks, ex-StitchFix). The course focuses on building scalable, production-grade generative AI systems, with hands-on sessions, $1,000+ in cloud credits, live Q&amp;As, and guest lectures from industry experts.</p>

<p>Listeners of Vanishing Gradients can get 25% off the course using <a href="https://maven.com/hugo-stefan/building-llm-apps-ds-and-swe-from-first-principles?promoCode=VG25" rel="nofollow">this special link</a> or by applying the code VG25 at checkout.</p>]]>
  </content:encoded>
  <itunes:summary>
    <![CDATA[<p>Hugo speaks with Ravin Kumar,*Senior Research Data Scientist at Google Labs. Ravin’s career has taken him from building rockets at SpaceX to driving data science and technology at Sweetgreen, and now to advancing generative AI research and applications at Google Labs and DeepMind. His multidisciplinary experience gives him a rare perspective on building AI systems that combine technical rigor with practical utility.</p>

<p>In this episode, we dive into:<br>
    • Ravin’s fascinating career path, including the skills and mindsets needed to work effectively with AI and machine learning models at different stages of the pipeline.<br>
    • How to build generative AI systems that are scalable, reliable, and aligned with user needs.<br>
    • Real-world applications of generative AI, such as using open weight models such as Gemma to help a bakery streamline operations—an example of delivering tangible business value through AI.<br>
    • The critical role of UX in AI adoption, and how Ravin approaches designing tools like Notebook LM with the user journey in mind.</p>

<p>We also include a live demo where Ravin uses Notebook LM to analyze my website, extract insights, and even generate a podcast-style conversation about me. While some of the demo is visual, much can be appreciated through audio, and we’ve added a link to the video in the show notes for those who want to see it in action. We’ve also included the generated segment at the end of the episode for you to enjoy.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://www.youtube.com/live/ffS6NWqoo_k" rel="nofollow">The livestream on YouTube</a></li>
<li><a href="https://labs.google/" rel="nofollow">Google Labs</a></li>
<li><a href="https://ravinkumar.com/GenAiGuidebook/book_intro.html" rel="nofollow">Ravin&#39;s GenAI Handbook</a></li>
<li><a href="https://breadboard-ai.github.io/breadboard/" rel="nofollow">Breadboard: A library for prototyping generative AI applications</a></li>
</ul>

<p>As mentioned in the episode, Hugo is teaching a four-week course, <strong>Building LLM Applications for Data Scientists and SWEs</strong>, co-led with Stefan Krawczyk (Dagworks, ex-StitchFix). The course focuses on building scalable, production-grade generative AI systems, with hands-on sessions, $1,000+ in cloud credits, live Q&amp;As, and guest lectures from industry experts.</p>

<p>Listeners of Vanishing Gradients can get 25% off the course using <a href="https://maven.com/hugo-stefan/building-llm-apps-ds-and-swe-from-first-principles?promoCode=VG25" rel="nofollow">this special link</a> or by applying the code VG25 at checkout.</p>]]>
  </itunes:summary>
</item>
<item>
  <title>Episode 38: The Art of Freelance AI Consulting and Products: Data, Dollars, and Deliverables</title>
  <link>https://vanishinggradients.fireside.fm/38</link>
  <guid isPermaLink="false">c1a5c8d1-777a-41b7-a123-6b06861dbc35</guid>
  <pubDate>Tue, 05 Nov 2024 10:00:00 +1100</pubDate>
  <author>Hugo Bowne-Anderson</author>
  <enclosure url="https://aphid.fireside.fm/d/1437767933/140c3904-8258-4c39-a698-a112b7077bd7/c1a5c8d1-777a-41b7-a123-6b06861dbc35.mp3" length="80443270" type="audio/mpeg"/>
  <itunes:episodeType>full</itunes:episodeType>
  <itunes:season>1</itunes:season>
  <itunes:author>Hugo Bowne-Anderson</itunes:author>
  <itunes:subtitle>Hugo speaks with Jason Liu, an independent AI consultant with experience at Meta and Stitch Fix. At Stitch Fix, Jason developed impactful AI systems, like a $50 million product similarity search and the widely adopted Flight recommendation framework. Now, he helps startups and enterprises design and deploy production-level AI applications, with a focus on retrieval-augmented generation (RAG) and scalable solutions.</itunes:subtitle>
  <itunes:duration>1:23:47</itunes:duration>
  <itunes:explicit>no</itunes:explicit>
  <itunes:image href="https://media24.fireside.fm/file/fireside-images-2024/podcasts/images/1/140c3904-8258-4c39-a698-a112b7077bd7/cover.jpg?v=1"/>
  <description>Hugo speaks with Jason Liu, an independent AI consultant with experience at Meta and Stitch Fix. At Stitch Fix, Jason developed impactful AI systems, like a $50 million product similarity search and the widely adopted Flight recommendation framework. Now, he helps startups and enterprises design and deploy production-level AI applications, with a focus on retrieval-augmented generation (RAG) and scalable solutions.
This episode is a bit of an experiment. Instead of our usual technical deep dives, we’re focusing on the world of AI consulting and freelancing. We explore Jason’s consulting playbook, covering how he structures contracts to maximize value, strategies for moving from hourly billing to securing larger deals, and the mindset shift needed to align incentives with clients. We’ll also discuss the challenges of moving from deterministic software to probabilistic AI systems and even do a live role-playing session where Jason coaches me on client engagement and pricing pitfalls.
LINKS
The livestream on YouTube (https://youtube.com/live/9CFs06UDbGI?feature=share)
Jason's Upcoming course: AI Consultant Accelerator: From Expert to High-Demand Business (https://maven.com/indie-consulting/ai-consultant-accelerator?utm_campaign=9532cc&amp;amp;utm_medium=partner&amp;amp;utm_source=instructor)
Hugo's upcoming course: Building LLM Applications for Data Scientists and Software Engineers (https://maven.com/s/course/d56067f338)
Jason's website (https://jxnl.co/)
Jason's indie consulting newsletter (https://indieconsulting.podia.com/)
Your AI Product Needs Evals by Hamel Husain (https://hamel.dev/blog/posts/evals/)
What We’ve Learned From A Year of Building with LLMs (https://applied-llms.org/)
Dear Future AI Consultant by Jason (https://jxnl.co/writing/#dear-future-ai-consultant)
Alex Hormozi's books (https://www.acquisition.com/books)
The Burnout Society by Byung-Chul Han (https://www.sup.org/books/theory-and-philosophy/burnout-society)
Jason on Twitter (https://x.com/jxnlco)
Vanishing Gradients on Twitter (https://twitter.com/vanishingdata)
Hugo on Twitter (https://twitter.com/hugobowne)
Vanishing Gradients' lu.ma calendar (https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk)
Vanishing Gradients on YouTube (https://www.youtube.com/@vanishinggradients) 
</description>
  <itunes:keywords>AI, LLMs, machine learning, data science, GenAI, consulting</itunes:keywords>
  <content:encoded>
    <![CDATA[<p>Hugo speaks with Jason Liu, an independent AI consultant with experience at Meta and Stitch Fix. At Stitch Fix, Jason developed impactful AI systems, like a $50 million product similarity search and the widely adopted Flight recommendation framework. Now, he helps startups and enterprises design and deploy production-level AI applications, with a focus on retrieval-augmented generation (RAG) and scalable solutions.</p>

<p>This episode is a bit of an experiment. Instead of our usual technical deep dives, we’re focusing on the world of AI consulting and freelancing. We explore Jason’s consulting playbook, covering how he structures contracts to maximize value, strategies for moving from hourly billing to securing larger deals, and the mindset shift needed to align incentives with clients. We’ll also discuss the challenges of moving from deterministic software to probabilistic AI systems and even do a live role-playing session where Jason coaches me on client engagement and pricing pitfalls.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://youtube.com/live/9CFs06UDbGI?feature=share" rel="nofollow">The livestream on YouTube</a></li>
<li><a href="https://maven.com/indie-consulting/ai-consultant-accelerator?utm_campaign=9532cc&utm_medium=partner&utm_source=instructor" rel="nofollow">Jason&#39;s Upcoming course: AI Consultant Accelerator: From Expert to High-Demand Business</a></li>
<li><a href="https://maven.com/s/course/d56067f338" rel="nofollow">Hugo&#39;s upcoming course: Building LLM Applications for Data Scientists and Software Engineers</a></li>
<li><a href="https://jxnl.co/" rel="nofollow">Jason&#39;s website</a></li>
<li><a href="https://indieconsulting.podia.com/" rel="nofollow">Jason&#39;s indie consulting newsletter</a></li>
<li><a href="https://hamel.dev/blog/posts/evals/" rel="nofollow">Your AI Product Needs Evals by Hamel Husain</a></li>
<li><a href="https://applied-llms.org/" rel="nofollow">What We’ve Learned From A Year of Building with LLMs</a></li>
<li><a href="https://jxnl.co/writing/#dear-future-ai-consultant" rel="nofollow">Dear Future AI Consultant by Jason</a></li>
<li><a href="https://www.acquisition.com/books" rel="nofollow">Alex Hormozi&#39;s books</a></li>
<li><a href="https://www.sup.org/books/theory-and-philosophy/burnout-society" rel="nofollow">The Burnout Society by Byung-Chul Han</a></li>
<li><a href="https://x.com/jxnlco" rel="nofollow">Jason on Twitter</a></li>
<li><a href="https://twitter.com/vanishingdata" rel="nofollow">Vanishing Gradients on Twitter</a></li>
<li><a href="https://twitter.com/hugobowne" rel="nofollow">Hugo on Twitter</a></li>
<li><a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">Vanishing Gradients&#39; lu.ma calendar</a></li>
<li><a href="https://www.youtube.com/@vanishinggradients" rel="nofollow">Vanishing Gradients on YouTube</a></li>
</ul>]]>
  </content:encoded>
  <itunes:summary>
    <![CDATA[<p>Hugo speaks with Jason Liu, an independent AI consultant with experience at Meta and Stitch Fix. At Stitch Fix, Jason developed impactful AI systems, like a $50 million product similarity search and the widely adopted Flight recommendation framework. Now, he helps startups and enterprises design and deploy production-level AI applications, with a focus on retrieval-augmented generation (RAG) and scalable solutions.</p>

<p>This episode is a bit of an experiment. Instead of our usual technical deep dives, we’re focusing on the world of AI consulting and freelancing. We explore Jason’s consulting playbook, covering how he structures contracts to maximize value, strategies for moving from hourly billing to securing larger deals, and the mindset shift needed to align incentives with clients. We’ll also discuss the challenges of moving from deterministic software to probabilistic AI systems and even do a live role-playing session where Jason coaches me on client engagement and pricing pitfalls.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://youtube.com/live/9CFs06UDbGI?feature=share" rel="nofollow">The livestream on YouTube</a></li>
<li><a href="https://maven.com/indie-consulting/ai-consultant-accelerator?utm_campaign=9532cc&utm_medium=partner&utm_source=instructor" rel="nofollow">Jason&#39;s Upcoming course: AI Consultant Accelerator: From Expert to High-Demand Business</a></li>
<li><a href="https://maven.com/s/course/d56067f338" rel="nofollow">Hugo&#39;s upcoming course: Building LLM Applications for Data Scientists and Software Engineers</a></li>
<li><a href="https://jxnl.co/" rel="nofollow">Jason&#39;s website</a></li>
<li><a href="https://indieconsulting.podia.com/" rel="nofollow">Jason&#39;s indie consulting newsletter</a></li>
<li><a href="https://hamel.dev/blog/posts/evals/" rel="nofollow">Your AI Product Needs Evals by Hamel Husain</a></li>
<li><a href="https://applied-llms.org/" rel="nofollow">What We’ve Learned From A Year of Building with LLMs</a></li>
<li><a href="https://jxnl.co/writing/#dear-future-ai-consultant" rel="nofollow">Dear Future AI Consultant by Jason</a></li>
<li><a href="https://www.acquisition.com/books" rel="nofollow">Alex Hormozi&#39;s books</a></li>
<li><a href="https://www.sup.org/books/theory-and-philosophy/burnout-society" rel="nofollow">The Burnout Society by Byung-Chul Han</a></li>
<li><a href="https://x.com/jxnlco" rel="nofollow">Jason on Twitter</a></li>
<li><a href="https://twitter.com/vanishingdata" rel="nofollow">Vanishing Gradients on Twitter</a></li>
<li><a href="https://twitter.com/hugobowne" rel="nofollow">Hugo on Twitter</a></li>
<li><a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">Vanishing Gradients&#39; lu.ma calendar</a></li>
<li><a href="https://www.youtube.com/@vanishinggradients" rel="nofollow">Vanishing Gradients on YouTube</a></li>
</ul>]]>
  </itunes:summary>
</item>
<item>
  <title>Episode 37: Prompt Engineering, Security in Generative AI, and the Future of AI Research Part 2</title>
  <link>https://vanishinggradients.fireside.fm/37</link>
  <guid isPermaLink="false">eadec2c4-f8f9-45b0-ae7e-5867f7201801</guid>
  <pubDate>Tue, 08 Oct 2024 17:00:00 +1100</pubDate>
  <author>Hugo Bowne-Anderson</author>
  <enclosure url="https://aphid.fireside.fm/d/1437767933/140c3904-8258-4c39-a698-a112b7077bd7/eadec2c4-f8f9-45b0-ae7e-5867f7201801.mp3" length="48585166" type="audio/mpeg"/>
  <itunes:episodeType>full</itunes:episodeType>
  <itunes:season>1</itunes:season>
  <itunes:author>Hugo Bowne-Anderson</itunes:author>
  <itunes:subtitle>Hugo speaks with three leading figures from the world of AI research: Sander Schulhoff, a recent University of Maryland graduate and lead contributor to the Learn Prompting initiative; Philip Resnik, professor at the University of Maryland, known for his pioneering work in computational linguistics; and Dennis Peskoff, a researcher from Princeton specializing in prompt engineering and its applications in the social sciences.</itunes:subtitle>
  <itunes:duration>50:36</itunes:duration>
  <itunes:explicit>no</itunes:explicit>
  <itunes:image href="https://media24.fireside.fm/file/fireside-images-2024/podcasts/images/1/140c3904-8258-4c39-a698-a112b7077bd7/cover.jpg?v=1"/>
  <description>Hugo speaks with three leading figures from the world of AI research: Sander Schulhoff, a recent University of Maryland graduate and lead contributor to the Learn Prompting initiative; Philip Resnik, professor at the University of Maryland, known for his pioneering work in computational linguistics; and Dennis Peskoff, a researcher from Princeton specializing in prompt engineering and its applications in the social sciences.
This is Part 2 of a special two-part episode, prompted—no pun intended—by these guys being part of a team, led by Sander, that wrote a 76-page survey analyzing prompting techniques, agents, and generative AI. The survey included contributors from OpenAI, Microsoft, the University of Maryland, Princeton, and more.
In this episode, we cover:
The Prompt Report: A comprehensive survey on prompting techniques, agents, and generative AI, including advanced evaluation methods for assessing these techniques.
Security Risks and Prompt Hacking: A detailed exploration of the security concerns surrounding prompt engineering, including Sander’s thoughts on its potential applications in cybersecurity and military contexts.
AI’s Impact Across Fields: A discussion on how generative AI is reshaping various domains, including the social sciences and security.
Multimodal AI: Updates on how large language models (LLMs) are expanding to interact with images, code, and music.
Case Study - Detecting Suicide Risk: A careful examination of how prompting techniques are being used in important areas like detecting suicide risk, showcasing the critical potential of AI in addressing sensitive, real-world challenges.
The episode concludes with a reflection on the evolving landscape of LLMs and multimodal AI, and what might be on the horizon.
If you haven’t yet, make sure to check out Part 1, where we discuss the history of NLP, prompt engineering techniques, and Sander’s development of the Learn Prompting initiative.
LINKS
The livestream on YouTube (https://youtube.com/live/FreXovgG-9A?feature=share)
The Prompt Report: A Systematic Survey of Prompting Techniques (https://arxiv.org/abs/2406.06608)
Learn Prompting: Your Guide to Communicating with AI (https://learnprompting.org/)
Vanishing Gradients on Twitter (https://twitter.com/vanishingdata)
Hugo on Twitter (https://twitter.com/hugobowne)
Vanishing Gradients' lu.ma calendar (https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk)
Vanishing Gradients on YouTube (https://www.youtube.com/@vanishinggradients)
</description>
  <itunes:keywords>AI, LLMs, machine learning, data science, GenAI, NLP</itunes:keywords>
  <content:encoded>
    <![CDATA[<p>Hugo speaks with three leading figures from the world of AI research: Sander Schulhoff, a recent University of Maryland graduate and lead contributor to the Learn Prompting initiative; Philip Resnik, professor at the University of Maryland, known for his pioneering work in computational linguistics; and Dennis Peskoff, a researcher from Princeton specializing in prompt engineering and its applications in the social sciences.</p>

<p>This is Part 2 of a special two-part episode, prompted—no pun intended—by these guys being part of a team, led by Sander, that wrote a 76-page survey analyzing prompting techniques, agents, and generative AI. The survey included contributors from OpenAI, Microsoft, the University of Maryland, Princeton, and more.</p>

<p>In this episode, we cover:</p>

<ul>
<li><p><strong>The Prompt Report:</strong> A comprehensive survey on prompting techniques, agents, and generative AI, including advanced evaluation methods for assessing these techniques.</p></li>
<li><p><strong>Security Risks and Prompt Hacking:</strong> A detailed exploration of the security concerns surrounding prompt engineering, including Sander’s thoughts on its potential applications in cybersecurity and military contexts.</p></li>
<li><p><strong>AI’s Impact Across Fields:</strong> A discussion on how generative AI is reshaping various domains, including the social sciences and security.</p></li>
<li><p><strong>Multimodal AI:</strong> Updates on how large language models (LLMs) are expanding to interact with images, code, and music.</p></li>
<li><p><strong>Case Study - Detecting Suicide Risk:</strong> A careful examination of how prompting techniques are being used in important areas like detecting suicide risk, showcasing the critical potential of AI in addressing sensitive, real-world challenges.</p></li>
</ul>

<p>The episode concludes with a reflection on the evolving landscape of <strong>LLMs</strong> and multimodal AI, and what might be on the horizon.</p>

<p>If you haven’t yet, make sure to check out <strong>Part 1</strong>, where we discuss the history of NLP, prompt engineering techniques, and Sander’s development of the Learn Prompting initiative.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://youtube.com/live/FreXovgG-9A?feature=share" rel="nofollow">The livestream on YouTube</a></li>
<li><a href="https://arxiv.org/abs/2406.06608" rel="nofollow">The Prompt Report: A Systematic Survey of Prompting Techniques</a></li>
<li><a href="https://learnprompting.org/" rel="nofollow">Learn Prompting: Your Guide to Communicating with AI</a></li>
<li><a href="https://twitter.com/vanishingdata" rel="nofollow">Vanishing Gradients on Twitter</a></li>
<li><a href="https://twitter.com/hugobowne" rel="nofollow">Hugo on Twitter</a></li>
<li><a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">Vanishing Gradients&#39; lu.ma calendar</a></li>
<li><a href="https://www.youtube.com/@vanishinggradients" rel="nofollow">Vanishing Gradients on YouTube</a></li>
</ul>]]>
  </content:encoded>
  <itunes:summary>
    <![CDATA[<p>Hugo speaks with three leading figures from the world of AI research: Sander Schulhoff, a recent University of Maryland graduate and lead contributor to the Learn Prompting initiative; Philip Resnik, professor at the University of Maryland, known for his pioneering work in computational linguistics; and Dennis Peskoff, a researcher from Princeton specializing in prompt engineering and its applications in the social sciences.</p>

<p>This is Part 2 of a special two-part episode, prompted—no pun intended—by these guys being part of a team, led by Sander, that wrote a 76-page survey analyzing prompting techniques, agents, and generative AI. The survey included contributors from OpenAI, Microsoft, the University of Maryland, Princeton, and more.</p>

<p>In this episode, we cover:</p>

<ul>
<li><p><strong>The Prompt Report:</strong> A comprehensive survey on prompting techniques, agents, and generative AI, including advanced evaluation methods for assessing these techniques.</p></li>
<li><p><strong>Security Risks and Prompt Hacking:</strong> A detailed exploration of the security concerns surrounding prompt engineering, including Sander’s thoughts on its potential applications in cybersecurity and military contexts.</p></li>
<li><p><strong>AI’s Impact Across Fields:</strong> A discussion on how generative AI is reshaping various domains, including the social sciences and security.</p></li>
<li><p><strong>Multimodal AI:</strong> Updates on how large language models (LLMs) are expanding to interact with images, code, and music.</p></li>
<li><p><strong>Case Study - Detecting Suicide Risk:</strong> A careful examination of how prompting techniques are being used in important areas like detecting suicide risk, showcasing the critical potential of AI in addressing sensitive, real-world challenges.</p></li>
</ul>

<p>The episode concludes with a reflection on the evolving landscape of <strong>LLMs</strong> and multimodal AI, and what might be on the horizon.</p>

<p>If you haven’t yet, make sure to check out <strong>Part 1</strong>, where we discuss the history of NLP, prompt engineering techniques, and Sander’s development of the Learn Prompting initiative.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://youtube.com/live/FreXovgG-9A?feature=share" rel="nofollow">The livestream on YouTube</a></li>
<li><a href="https://arxiv.org/abs/2406.06608" rel="nofollow">The Prompt Report: A Systematic Survey of Prompting Techniques</a></li>
<li><a href="https://learnprompting.org/" rel="nofollow">Learn Prompting: Your Guide to Communicating with AI</a></li>
<li><a href="https://twitter.com/vanishingdata" rel="nofollow">Vanishing Gradients on Twitter</a></li>
<li><a href="https://twitter.com/hugobowne" rel="nofollow">Hugo on Twitter</a></li>
<li><a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">Vanishing Gradients&#39; lu.ma calendar</a></li>
<li><a href="https://www.youtube.com/@vanishinggradients" rel="nofollow">Vanishing Gradients on YouTube</a></li>
</ul>]]>
  </itunes:summary>
</item>
<item>
  <title>Episode 36: Prompt Engineering, Security in Generative AI, and the Future of AI Research Part 1</title>
  <link>https://vanishinggradients.fireside.fm/36</link>
  <guid isPermaLink="false">acd8aaec-1788-459d-a4e9-10feae67a19a</guid>
  <pubDate>Mon, 30 Sep 2024 18:00:00 +1000</pubDate>
  <author>Hugo Bowne-Anderson</author>
  <enclosure url="https://aphid.fireside.fm/d/1437767933/140c3904-8258-4c39-a698-a112b7077bd7/acd8aaec-1788-459d-a4e9-10feae67a19a.mp3" length="61232193" type="audio/mpeg"/>
  <itunes:episodeType>full</itunes:episodeType>
  <itunes:season>1</itunes:season>
  <itunes:author>Hugo Bowne-Anderson</itunes:author>
  <itunes:subtitle>Hugo speaks with three leading figures from the world of AI research: Sander Schulhoff, a recent University of Maryland graduate and lead contributor to the Learn Prompting initiative; Philip Resnik, professor at the University of Maryland, known for his pioneering work in computational linguistics; and Dennis Peskoff, a researcher from Princeton specializing in prompt engineering and its applications in the social sciences.</itunes:subtitle>
  <itunes:duration>1:03:46</itunes:duration>
  <itunes:explicit>no</itunes:explicit>
  <itunes:image href="https://media24.fireside.fm/file/fireside-images-2024/podcasts/images/1/140c3904-8258-4c39-a698-a112b7077bd7/cover.jpg?v=1"/>
  <description>Hugo speaks with three leading figures from the world of AI research: Sander Schulhoff, a recent University of Maryland graduate and lead contributor to the Learn Prompting initiative; Philip Resnik, professor at the University of Maryland, known for his pioneering work in computational linguistics; and Dennis Peskoff, a researcher from Princeton specializing in prompt engineering and its applications in the social sciences.
This is Part 1 of a special two-part episode, prompted—no pun intended—by these guys being part of a team, led by Sander, that wrote a 76-page survey analyzing prompting techniques, agents, and generative AI. The survey included contributors from OpenAI, Microsoft, the University of Maryland, Princeton, and more.
In this first part, 
* we’ll explore the critical role of prompt engineering, 
* &amp;amp; diving into adversarial techniques like prompt hacking and 
* the challenges of evaluating these techniques. 
* we’ll examine the impact of few-shot learning and 
* the groundbreaking taxonomy of prompting techniques from the Prompt Report.
Along the way, 
* we’ll uncover the rich history of natural language processing (NLP) and AI, showing how modern prompting techniques evolved from early rule-based systems and statistical methods. 
* we’ll also hear how Sander’s experimentation with GPT-3 for diplomatic tasks led him to develop Learn Prompting, and 
* how Dennis highlights the accessibility of AI through prompting, which allows non-technical users to interact with AI without needing to code.
Finally, we’ll explore the future of multimodal AI, where LLMs interact with images, code, and even music creation. Make sure to tune in to Part 2, where we dive deeper into security risks, prompt hacking, and more.
LINKS
The livestream on YouTube (https://youtube.com/live/FreXovgG-9A?feature=share)
The Prompt Report: A Systematic Survey of Prompting Techniques (https://arxiv.org/abs/2406.06608)
Learn Prompting: Your Guide to Communicating with AI (https://learnprompting.org/)
Vanishing Gradients on Twitter (https://twitter.com/vanishingdata)
Hugo on Twitter (https://twitter.com/hugobowne)
Vanishing Gradients' lu.ma calendar (https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk)
Vanishing Gradients on YouTube (https://www.youtube.com/@vanishinggradients)
</description>
  <itunes:keywords>AI, LLMs, damachine learning, data science, GenAI, prompt engineering</itunes:keywords>
  <content:encoded>
    <![CDATA[<p>Hugo speaks with three leading figures from the world of AI research: Sander Schulhoff, a recent University of Maryland graduate and lead contributor to the Learn Prompting initiative; Philip Resnik, professor at the University of Maryland, known for his pioneering work in computational linguistics; and Dennis Peskoff, a researcher from Princeton specializing in prompt engineering and its applications in the social sciences.</p>

<p>This is Part 1 of a special two-part episode, prompted—no pun intended—by these guys being part of a team, led by Sander, that wrote a 76-page survey analyzing prompting techniques, agents, and generative AI. The survey included contributors from OpenAI, Microsoft, the University of Maryland, Princeton, and more.</p>

<p>In this first part, </p>

<ul>
<li>we’ll explore the critical role of prompt engineering, </li>
<li>&amp; diving into adversarial techniques like prompt hacking and </li>
<li>the challenges of evaluating these techniques. </li>
<li>we’ll examine the impact of few-shot learning and </li>
<li>the groundbreaking taxonomy of prompting techniques from the Prompt Report.</li>
</ul>

<p>Along the way, </p>

<ul>
<li>we’ll uncover the rich history of natural language processing (NLP) and AI, showing how modern prompting techniques evolved from early rule-based systems and statistical methods. </li>
<li>we’ll also hear how Sander’s experimentation with GPT-3 for diplomatic tasks led him to develop Learn Prompting, and </li>
<li>how Dennis highlights the accessibility of AI through prompting, which allows non-technical users to interact with AI without needing to code.</li>
</ul>

<p>Finally, we’ll explore the future of multimodal AI, where LLMs interact with images, code, and even music creation. Make sure to tune in to Part 2, where we dive deeper into security risks, prompt hacking, and more.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://youtube.com/live/FreXovgG-9A?feature=share" rel="nofollow">The livestream on YouTube</a></li>
<li><a href="https://arxiv.org/abs/2406.06608" rel="nofollow">The Prompt Report: A Systematic Survey of Prompting Techniques</a></li>
<li><a href="https://learnprompting.org/" rel="nofollow">Learn Prompting: Your Guide to Communicating with AI</a></li>
<li><a href="https://twitter.com/vanishingdata" rel="nofollow">Vanishing Gradients on Twitter</a></li>
<li><a href="https://twitter.com/hugobowne" rel="nofollow">Hugo on Twitter</a></li>
<li><a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">Vanishing Gradients&#39; lu.ma calendar</a></li>
<li><a href="https://www.youtube.com/@vanishinggradients" rel="nofollow">Vanishing Gradients on YouTube</a></li>
</ul>]]>
  </content:encoded>
  <itunes:summary>
    <![CDATA[<p>Hugo speaks with three leading figures from the world of AI research: Sander Schulhoff, a recent University of Maryland graduate and lead contributor to the Learn Prompting initiative; Philip Resnik, professor at the University of Maryland, known for his pioneering work in computational linguistics; and Dennis Peskoff, a researcher from Princeton specializing in prompt engineering and its applications in the social sciences.</p>

<p>This is Part 1 of a special two-part episode, prompted—no pun intended—by these guys being part of a team, led by Sander, that wrote a 76-page survey analyzing prompting techniques, agents, and generative AI. The survey included contributors from OpenAI, Microsoft, the University of Maryland, Princeton, and more.</p>

<p>In this first part, </p>

<ul>
<li>we’ll explore the critical role of prompt engineering, </li>
<li>&amp; diving into adversarial techniques like prompt hacking and </li>
<li>the challenges of evaluating these techniques. </li>
<li>we’ll examine the impact of few-shot learning and </li>
<li>the groundbreaking taxonomy of prompting techniques from the Prompt Report.</li>
</ul>

<p>Along the way, </p>

<ul>
<li>we’ll uncover the rich history of natural language processing (NLP) and AI, showing how modern prompting techniques evolved from early rule-based systems and statistical methods. </li>
<li>we’ll also hear how Sander’s experimentation with GPT-3 for diplomatic tasks led him to develop Learn Prompting, and </li>
<li>how Dennis highlights the accessibility of AI through prompting, which allows non-technical users to interact with AI without needing to code.</li>
</ul>

<p>Finally, we’ll explore the future of multimodal AI, where LLMs interact with images, code, and even music creation. Make sure to tune in to Part 2, where we dive deeper into security risks, prompt hacking, and more.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://youtube.com/live/FreXovgG-9A?feature=share" rel="nofollow">The livestream on YouTube</a></li>
<li><a href="https://arxiv.org/abs/2406.06608" rel="nofollow">The Prompt Report: A Systematic Survey of Prompting Techniques</a></li>
<li><a href="https://learnprompting.org/" rel="nofollow">Learn Prompting: Your Guide to Communicating with AI</a></li>
<li><a href="https://twitter.com/vanishingdata" rel="nofollow">Vanishing Gradients on Twitter</a></li>
<li><a href="https://twitter.com/hugobowne" rel="nofollow">Hugo on Twitter</a></li>
<li><a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">Vanishing Gradients&#39; lu.ma calendar</a></li>
<li><a href="https://www.youtube.com/@vanishinggradients" rel="nofollow">Vanishing Gradients on YouTube</a></li>
</ul>]]>
  </itunes:summary>
</item>
<item>
  <title>Episode 35: Open Science at NASA -- Measuring Impact and the Future of AI</title>
  <link>https://vanishinggradients.fireside.fm/35</link>
  <guid isPermaLink="false">feeeecc8-a170-48c7-ae4c-8dd64484c64c</guid>
  <pubDate>Thu, 19 Sep 2024 17:00:00 +1000</pubDate>
  <author>Hugo Bowne-Anderson</author>
  <enclosure url="https://aphid.fireside.fm/d/1437767933/140c3904-8258-4c39-a698-a112b7077bd7/feeeecc8-a170-48c7-ae4c-8dd64484c64c.mp3" length="55905303" type="audio/mpeg"/>
  <itunes:episodeType>full</itunes:episodeType>
  <itunes:season>1</itunes:season>
  <itunes:author>Hugo Bowne-Anderson</itunes:author>
  <itunes:subtitle>Hugo speaks with Dr. Chelle Gentemann, Open Science Program Scientist for NASA’s Office of the Chief Science Data Officer, about NASA’s ambitious efforts to integrate AI across the research lifecycle. In this episode, we’ll dive deeper into how AI is transforming NASA’s approach to science, making data more accessible and advancing open science practices.</itunes:subtitle>
  <itunes:duration>58:13</itunes:duration>
  <itunes:explicit>no</itunes:explicit>
  <itunes:image href="https://media24.fireside.fm/file/fireside-images-2024/podcasts/images/1/140c3904-8258-4c39-a698-a112b7077bd7/cover.jpg?v=1"/>
  <description>Hugo speaks with Dr. Chelle Gentemann, Open Science Program Scientist for NASA’s Office of the Chief Science Data Officer, about NASA’s ambitious efforts to integrate AI across the research lifecycle. In this episode, we’ll dive deeper into how AI is transforming NASA’s approach to science, making data more accessible and advancing open science practices. We explore
Measuring the Impact of Open Science: How NASA is developing new metrics to evaluate the effectiveness of open science, moving beyond traditional publication-based assessments.
The Process of Scientific Discovery: Insights into the collaborative nature of research and how breakthroughs are achieved at NASA.
** AI Applications in NASA’s Science:** From rats in space to exploring the origins of the universe, we cover how AI is being applied across NASA’s divisions to improve data accessibility and analysis.
Addressing Challenges in Open Science: The complexities of implementing open science within government agencies and research environments.
Reforming Incentive Systems: How NASA is reconsidering traditional metrics like publications and citations, and starting to recognize contributions such as software development and data sharing.
The Future of Open Science: How open science is shaping the future of research, fostering interdisciplinary collaboration, and increasing accessibility.
This conversation offers valuable insights for researchers, data scientists, and those interested in the practical applications of AI and open science. Join us as we discuss how NASA is working to make science more collaborative, reproducible, and impactful.
LINKS
The livestream on YouTube (https://youtube.com/live/VJDg3ZbkNOE?feature=share)
NASA's Open Science 101 course &amp;lt;-- do it to learn and also to get NASA Swag! (https://openscience101.org/)
Science Cast (https://sciencecast.org/)
NASA and IBM Openly Release Geospatial AI Foundation Model for NASA Earth Observation Data (https://www.earthdata.nasa.gov/news/impact-ibm-hls-foundation-model)
Jake VanderPlas' daily conundrum tweet from 2013 (https://x.com/jakevdp/status/408678764705378304)
Replit, "an AI-powered software development &amp;amp; deployment platform for building, sharing, and shipping software fast." (https://replit.com/) 
</description>
  <itunes:keywords>AI, LLMs, damachine learning, data science, GenAI</itunes:keywords>
  <content:encoded>
    <![CDATA[<p>Hugo speaks with Dr. Chelle Gentemann, Open Science Program Scientist for NASA’s Office of the Chief Science Data Officer, about NASA’s ambitious efforts to integrate AI across the research lifecycle. In this episode, we’ll dive deeper into how AI is transforming NASA’s approach to science, making data more accessible and advancing open science practices. We explore</p>

<ul>
<li><strong>Measuring the Impact of Open Science:</strong> How NASA is developing new metrics to evaluate the effectiveness of open science, moving beyond traditional publication-based assessments.</li>
<li><strong>The Process of Scientific Discovery:</strong> Insights into the collaborative nature of research and how breakthroughs are achieved at NASA.</li>
<li>** AI Applications in NASA’s Science:** From rats in space to exploring the origins of the universe, we cover how AI is being applied across NASA’s divisions to improve data accessibility and analysis.</li>
<li><strong>Addressing Challenges in Open Science:</strong> The complexities of implementing open science within government agencies and research environments.</li>
<li><strong>Reforming Incentive Systems:</strong> How NASA is reconsidering traditional metrics like publications and citations, and starting to recognize contributions such as software development and data sharing.</li>
<li><strong>The Future of Open Science:</strong> How open science is shaping the future of research, fostering interdisciplinary collaboration, and increasing accessibility.</li>
</ul>

<p>This conversation offers valuable insights for researchers, data scientists, and those interested in the practical applications of AI and open science. Join us as we discuss how NASA is working to make science more collaborative, reproducible, and impactful.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://youtube.com/live/VJDg3ZbkNOE?feature=share" rel="nofollow">The livestream on YouTube</a></li>
<li><a href="https://openscience101.org/" rel="nofollow">NASA&#39;s Open Science 101 course &lt;-- do it to learn and also to get NASA Swag!</a></li>
<li><a href="https://sciencecast.org/" rel="nofollow">Science Cast</a></li>
<li><a href="https://www.earthdata.nasa.gov/news/impact-ibm-hls-foundation-model" rel="nofollow">NASA and IBM Openly Release Geospatial AI Foundation Model for NASA Earth Observation Data</a></li>
<li><a href="https://x.com/jakevdp/status/408678764705378304" rel="nofollow">Jake VanderPlas&#39; daily conundrum tweet from 2013</a></li>
<li><a href="https://replit.com/" rel="nofollow">Replit, &quot;an AI-powered software development &amp; deployment platform for building, sharing, and shipping software fast.&quot;</a></li>
</ul>]]>
  </content:encoded>
  <itunes:summary>
    <![CDATA[<p>Hugo speaks with Dr. Chelle Gentemann, Open Science Program Scientist for NASA’s Office of the Chief Science Data Officer, about NASA’s ambitious efforts to integrate AI across the research lifecycle. In this episode, we’ll dive deeper into how AI is transforming NASA’s approach to science, making data more accessible and advancing open science practices. We explore</p>

<ul>
<li><strong>Measuring the Impact of Open Science:</strong> How NASA is developing new metrics to evaluate the effectiveness of open science, moving beyond traditional publication-based assessments.</li>
<li><strong>The Process of Scientific Discovery:</strong> Insights into the collaborative nature of research and how breakthroughs are achieved at NASA.</li>
<li>** AI Applications in NASA’s Science:** From rats in space to exploring the origins of the universe, we cover how AI is being applied across NASA’s divisions to improve data accessibility and analysis.</li>
<li><strong>Addressing Challenges in Open Science:</strong> The complexities of implementing open science within government agencies and research environments.</li>
<li><strong>Reforming Incentive Systems:</strong> How NASA is reconsidering traditional metrics like publications and citations, and starting to recognize contributions such as software development and data sharing.</li>
<li><strong>The Future of Open Science:</strong> How open science is shaping the future of research, fostering interdisciplinary collaboration, and increasing accessibility.</li>
</ul>

<p>This conversation offers valuable insights for researchers, data scientists, and those interested in the practical applications of AI and open science. Join us as we discuss how NASA is working to make science more collaborative, reproducible, and impactful.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://youtube.com/live/VJDg3ZbkNOE?feature=share" rel="nofollow">The livestream on YouTube</a></li>
<li><a href="https://openscience101.org/" rel="nofollow">NASA&#39;s Open Science 101 course &lt;-- do it to learn and also to get NASA Swag!</a></li>
<li><a href="https://sciencecast.org/" rel="nofollow">Science Cast</a></li>
<li><a href="https://www.earthdata.nasa.gov/news/impact-ibm-hls-foundation-model" rel="nofollow">NASA and IBM Openly Release Geospatial AI Foundation Model for NASA Earth Observation Data</a></li>
<li><a href="https://x.com/jakevdp/status/408678764705378304" rel="nofollow">Jake VanderPlas&#39; daily conundrum tweet from 2013</a></li>
<li><a href="https://replit.com/" rel="nofollow">Replit, &quot;an AI-powered software development &amp; deployment platform for building, sharing, and shipping software fast.&quot;</a></li>
</ul>]]>
  </itunes:summary>
</item>
<item>
  <title>Episode 34: The AI Revolution Will Not Be Monopolized</title>
  <link>https://vanishinggradients.fireside.fm/34</link>
  <guid isPermaLink="false">8c18d59e-9b79-4682-8e3c-ba682daf1c1c</guid>
  <pubDate>Thu, 22 Aug 2024 17:00:00 +1000</pubDate>
  <author>Hugo Bowne-Anderson</author>
  <enclosure url="https://aphid.fireside.fm/d/1437767933/140c3904-8258-4c39-a698-a112b7077bd7/8c18d59e-9b79-4682-8e3c-ba682daf1c1c.mp3" length="98751972" type="audio/mpeg"/>
  <itunes:episodeType>full</itunes:episodeType>
  <itunes:season>1</itunes:season>
  <itunes:author>Hugo Bowne-Anderson</itunes:author>
  <itunes:subtitle>Hugo speaks with Ines Montani and Matthew Honnibal, the creators of spaCy and founders of Explosion AI. Collectively, they've had a huge impact on the fields of industrial natural language processing (NLP), ML, and AI through their widely-used open-source library spaCy and their innovative annotation tool Prodigy.</itunes:subtitle>
  <itunes:duration>1:42:51</itunes:duration>
  <itunes:explicit>yes</itunes:explicit>
  <itunes:image href="https://media24.fireside.fm/file/fireside-images-2024/podcasts/images/1/140c3904-8258-4c39-a698-a112b7077bd7/cover.jpg?v=1"/>
  <description>Hugo speaks with Ines Montani and Matthew Honnibal, the creators of spaCy and founders of Explosion AI. Collectively, they've had a huge impact on the fields of industrial natural language processing (NLP), ML, and AI through their widely-used open-source library spaCy and their innovative annotation tool Prodigy. These tools have become essential for many data scientists and NLP practitioners in industry and academia alike.
In this wide-ranging discussion, we dive into:
• The evolution of applied NLP and its role in industry
• The balance between large language models and smaller, specialized models
• Human-in-the-loop distillation for creating faster, more data-private AI systems
• The challenges and opportunities in NLP, including modularity, transparency, and privacy
• The future of AI and software development
• The potential impact of AI regulation on innovation and competition
We also touch on their recent transition back to a smaller, more independent-minded company structure and the lessons learned from their journey in the AI startup world.
Ines and Matt offer invaluable insights for data scientists, machine learning practitioners, and anyone interested in the practical applications of AI. They share their thoughts on how to approach NLP projects, the importance of data quality, and the role of open-source in advancing the field.
Whether you're a seasoned NLP practitioner or just getting started with AI, this episode offers a wealth of knowledge from two of the field's most respected figures. Join us for a discussion that explores the current landscape of AI development, with insights that bridge the gap between cutting-edge research and real-world applications.
LINKS
The livestream on YouTube (https://youtube.com/live/-6o5-3cP0ik?feature=share)
How S&amp;amp;P Global is making markets more transparent with NLP, spaCy and Prodigy (https://explosion.ai/blog/sp-global-commodities)
A practical guide to human-in-the-loop distillation (https://explosion.ai/blog/human-in-the-loop-distillation)
Laws of Tech: Commoditize Your Complement (https://gwern.net/complement)
spaCy: Industrial-Strength Natural Language Processing (https://spacy.io/)
LLMs with spaCy (https://spacy.io/usage/large-language-models)
Explosion, building developer tools for AI, Machine Learning and Natural Language Processing (https://explosion.ai/)
Back to our roots: Company update and future plans, by Matt and Ines (https://explosion.ai/blog/back-to-our-roots-company-update)
Matt's detailed blog post: back to our roots (https://honnibal.dev/blog/back-to-our-roots)
Ines on twitter (https://x.com/_inesmontani)
Matt on twitter (https://x.com/honnibal)
Vanishing Gradients on Twitter (https://twitter.com/vanishingdata)
Hugo on Twitter (https://twitter.com/hugobowne)
Check out and subcribe to our lu.ma calendar (https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk) for upcoming livestreams!
</description>
  <itunes:keywords>AI, LLMs, machine learning, data science, GenAI, NLP</itunes:keywords>
  <content:encoded>
    <![CDATA[<p>Hugo speaks with Ines Montani and Matthew Honnibal, the creators of spaCy and founders of Explosion AI. Collectively, they&#39;ve had a huge impact on the fields of industrial natural language processing (NLP), ML, and AI through their widely-used open-source library spaCy and their innovative annotation tool Prodigy. These tools have become essential for many data scientists and NLP practitioners in industry and academia alike.</p>

<p>In this wide-ranging discussion, we dive into:</p>

<p>• The evolution of applied NLP and its role in industry<br>
• The balance between large language models and smaller, specialized models<br>
• Human-in-the-loop distillation for creating faster, more data-private AI systems<br>
• The challenges and opportunities in NLP, including modularity, transparency, and privacy<br>
• The future of AI and software development<br>
• The potential impact of AI regulation on innovation and competition</p>

<p>We also touch on their recent transition back to a smaller, more independent-minded company structure and the lessons learned from their journey in the AI startup world.</p>

<p>Ines and Matt offer invaluable insights for data scientists, machine learning practitioners, and anyone interested in the practical applications of AI. They share their thoughts on how to approach NLP projects, the importance of data quality, and the role of open-source in advancing the field.</p>

<p>Whether you&#39;re a seasoned NLP practitioner or just getting started with AI, this episode offers a wealth of knowledge from two of the field&#39;s most respected figures. Join us for a discussion that explores the current landscape of AI development, with insights that bridge the gap between cutting-edge research and real-world applications.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://youtube.com/live/-6o5-3cP0ik?feature=share" rel="nofollow">The livestream on YouTube</a></li>
<li><a href="https://explosion.ai/blog/sp-global-commodities" rel="nofollow">How S&amp;P Global is making markets more transparent with NLP, spaCy and Prodigy</a></li>
<li><a href="https://explosion.ai/blog/human-in-the-loop-distillation" rel="nofollow">A practical guide to human-in-the-loop distillation</a></li>
<li><a href="https://gwern.net/complement" rel="nofollow">Laws of Tech: Commoditize Your Complement</a></li>
<li><a href="https://spacy.io/" rel="nofollow">spaCy: Industrial-Strength Natural Language Processing</a></li>
<li><a href="https://spacy.io/usage/large-language-models" rel="nofollow">LLMs with spaCy</a></li>
<li><a href="https://explosion.ai/" rel="nofollow">Explosion, building developer tools for AI, Machine Learning and Natural Language Processing</a></li>
<li><a href="https://explosion.ai/blog/back-to-our-roots-company-update" rel="nofollow">Back to our roots: Company update and future plans, by Matt and Ines</a></li>
<li><a href="https://honnibal.dev/blog/back-to-our-roots" rel="nofollow">Matt&#39;s detailed blog post: back to our roots</a></li>
<li><a href="https://x.com/_inesmontani" rel="nofollow">Ines on twitter</a></li>
<li><a href="https://x.com/honnibal" rel="nofollow">Matt on twitter</a></li>
<li><a href="https://twitter.com/vanishingdata" rel="nofollow">Vanishing Gradients on Twitter</a></li>
<li><a href="https://twitter.com/hugobowne" rel="nofollow">Hugo on Twitter</a></li>
</ul>

<p>Check out and subcribe to our <a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">lu.ma calendar</a> for upcoming livestreams!</p>]]>
  </content:encoded>
  <itunes:summary>
    <![CDATA[<p>Hugo speaks with Ines Montani and Matthew Honnibal, the creators of spaCy and founders of Explosion AI. Collectively, they&#39;ve had a huge impact on the fields of industrial natural language processing (NLP), ML, and AI through their widely-used open-source library spaCy and their innovative annotation tool Prodigy. These tools have become essential for many data scientists and NLP practitioners in industry and academia alike.</p>

<p>In this wide-ranging discussion, we dive into:</p>

<p>• The evolution of applied NLP and its role in industry<br>
• The balance between large language models and smaller, specialized models<br>
• Human-in-the-loop distillation for creating faster, more data-private AI systems<br>
• The challenges and opportunities in NLP, including modularity, transparency, and privacy<br>
• The future of AI and software development<br>
• The potential impact of AI regulation on innovation and competition</p>

<p>We also touch on their recent transition back to a smaller, more independent-minded company structure and the lessons learned from their journey in the AI startup world.</p>

<p>Ines and Matt offer invaluable insights for data scientists, machine learning practitioners, and anyone interested in the practical applications of AI. They share their thoughts on how to approach NLP projects, the importance of data quality, and the role of open-source in advancing the field.</p>

<p>Whether you&#39;re a seasoned NLP practitioner or just getting started with AI, this episode offers a wealth of knowledge from two of the field&#39;s most respected figures. Join us for a discussion that explores the current landscape of AI development, with insights that bridge the gap between cutting-edge research and real-world applications.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://youtube.com/live/-6o5-3cP0ik?feature=share" rel="nofollow">The livestream on YouTube</a></li>
<li><a href="https://explosion.ai/blog/sp-global-commodities" rel="nofollow">How S&amp;P Global is making markets more transparent with NLP, spaCy and Prodigy</a></li>
<li><a href="https://explosion.ai/blog/human-in-the-loop-distillation" rel="nofollow">A practical guide to human-in-the-loop distillation</a></li>
<li><a href="https://gwern.net/complement" rel="nofollow">Laws of Tech: Commoditize Your Complement</a></li>
<li><a href="https://spacy.io/" rel="nofollow">spaCy: Industrial-Strength Natural Language Processing</a></li>
<li><a href="https://spacy.io/usage/large-language-models" rel="nofollow">LLMs with spaCy</a></li>
<li><a href="https://explosion.ai/" rel="nofollow">Explosion, building developer tools for AI, Machine Learning and Natural Language Processing</a></li>
<li><a href="https://explosion.ai/blog/back-to-our-roots-company-update" rel="nofollow">Back to our roots: Company update and future plans, by Matt and Ines</a></li>
<li><a href="https://honnibal.dev/blog/back-to-our-roots" rel="nofollow">Matt&#39;s detailed blog post: back to our roots</a></li>
<li><a href="https://x.com/_inesmontani" rel="nofollow">Ines on twitter</a></li>
<li><a href="https://x.com/honnibal" rel="nofollow">Matt on twitter</a></li>
<li><a href="https://twitter.com/vanishingdata" rel="nofollow">Vanishing Gradients on Twitter</a></li>
<li><a href="https://twitter.com/hugobowne" rel="nofollow">Hugo on Twitter</a></li>
</ul>

<p>Check out and subcribe to our <a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">lu.ma calendar</a> for upcoming livestreams!</p>]]>
  </itunes:summary>
</item>
<item>
  <title>Episode 32: Building Reliable and Robust ML/AI Pipelines</title>
  <link>https://vanishinggradients.fireside.fm/32</link>
  <guid isPermaLink="false">3aa4ba58-30aa-4a85-a139-e9057629171c</guid>
  <pubDate>Sat, 27 Jul 2024 13:00:00 +1000</pubDate>
  <author>Hugo Bowne-Anderson</author>
  <enclosure url="https://aphid.fireside.fm/d/1437767933/140c3904-8258-4c39-a698-a112b7077bd7/3aa4ba58-30aa-4a85-a139-e9057629171c.mp3" length="72173111" type="audio/mpeg"/>
  <itunes:episodeType>full</itunes:episodeType>
  <itunes:season>1</itunes:season>
  <itunes:author>Hugo Bowne-Anderson</itunes:author>
  <itunes:subtitle>Hugo speaks with Shreya Shankar, a researcher at UC Berkeley focusing on data management systems with a human-centered approach. Shreya's work is at the cutting edge of human-computer interaction (HCI) and AI, particularly in the realm of large language models (LLMs). Her impressive background includes being the first ML engineer at Viaduct, doing research engineering at Google Brain, and software engineering at Facebook.</itunes:subtitle>
  <itunes:duration>1:15:10</itunes:duration>
  <itunes:explicit>yes</itunes:explicit>
  <itunes:image href="https://media24.fireside.fm/file/fireside-images-2024/podcasts/images/1/140c3904-8258-4c39-a698-a112b7077bd7/cover.jpg?v=1"/>
  <description>Hugo speaks with Shreya Shankar, a researcher at UC Berkeley focusing on data management systems with a human-centered approach. Shreya's work is at the cutting edge of human-computer interaction (HCI) and AI, particularly in the realm of large language models (LLMs). Her impressive background includes being the first ML engineer at Viaduct, doing research engineering at Google Brain, and software engineering at Facebook.
In this episode, we dive deep into the world of LLMs and the critical challenges of building reliable AI pipelines. We'll explore:
The fascinating journey from classic machine learning to the current LLM revolution
Why Shreya believes most ML problems are actually data management issues
The concept of "data flywheels" for LLM applications and how to implement them
The intriguing world of evaluating AI systems - who validates the validators?
Shreya's work on SPADE and EvalGen, innovative tools for synthesizing data quality assertions and aligning LLM evaluations with human preferences
The importance of human-in-the-loop processes in AI development
The future of low-code and no-code tools in the AI landscape
We'll also touch on the potential pitfalls of over-relying on LLMs, the concept of "Habsburg AI," and how to avoid disappearing up our own proverbial arseholes in the world of recursive AI processes.
Whether you're a seasoned AI practitioner, a curious data scientist, or someone interested in the human side of AI development, this conversation offers valuable insights into building more robust, reliable, and human-centered AI systems.
LINKS
The livestream on YouTube (https://youtube.com/live/hKV6xSJZkB0?feature=share)
Shreya's website (https://www.sh-reya.com/)
Shreya on Twitter (https://x.com/sh_reya)
Data Flywheels for LLM Applications (https://www.sh-reya.com/blog/ai-engineering-flywheel/)
SPADE: Synthesizing Data Quality Assertions for Large Language Model Pipelines (https://arxiv.org/abs/2401.03038)
What We’ve Learned From A Year of Building with LLMs (https://applied-llms.org/)
Who Validates the Validators? Aligning LLM-Assisted Evaluation of LLM Outputs with Human Preferences (https://arxiv.org/abs/2404.12272)
Operationalizing Machine Learning: An Interview Study (https://arxiv.org/abs/2209.09125)
Vanishing Gradients on Twitter (https://twitter.com/vanishingdata)
Hugo on Twitter (https://twitter.com/hugobowne)
In the podcast, Hugo also mentioned that this was the 5th time he and Shreya chatted publicly. which is wild!
If you want to dive deep into Shreya's work and related topics through their chats, you can check them all out here:
Outerbounds' Fireside Chat: Operationalizing ML -- Patterns and Pain Points from MLOps Practitioners (https://www.youtube.com/watch?v=7zB6ESFto_U)
The Past, Present, and Future of Generative AI (https://youtu.be/q0A9CdGWXqc?si=XmaUnQmZiXL2eagS)
LLMs, OpenAI Dev Day, and the Existential Crisis for Machine Learning Engineering (https://www.youtube.com/live/MTJHvgJtynU?si=Ncjqn5YuFBemvOJ0)
Lessons from a Year of Building with LLMs (https://youtube.com/live/c0gcsprsFig?feature=share)
Check out and subcribe to our lu.ma calendar (https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk) for upcoming livestreams! 
</description>
  <itunes:keywords>AI, LLMs, machine learning, data science, GenAI</itunes:keywords>
  <content:encoded>
    <![CDATA[<p>Hugo speaks with Shreya Shankar, a researcher at UC Berkeley focusing on data management systems with a human-centered approach. Shreya&#39;s work is at the cutting edge of human-computer interaction (HCI) and AI, particularly in the realm of large language models (LLMs). Her impressive background includes being the first ML engineer at Viaduct, doing research engineering at Google Brain, and software engineering at Facebook.</p>

<p>In this episode, we dive deep into the world of LLMs and the critical challenges of building reliable AI pipelines. We&#39;ll explore:</p>

<ul>
<li>The fascinating journey from classic machine learning to the current LLM revolution</li>
<li>Why Shreya believes most ML problems are actually data management issues</li>
<li>The concept of &quot;data flywheels&quot; for LLM applications and how to implement them</li>
<li>The intriguing world of evaluating AI systems - who validates the validators?</li>
<li>Shreya&#39;s work on SPADE and EvalGen, innovative tools for synthesizing data quality assertions and aligning LLM evaluations with human preferences</li>
<li>The importance of human-in-the-loop processes in AI development</li>
<li>The future of low-code and no-code tools in the AI landscape</li>
</ul>

<p>We&#39;ll also touch on the potential pitfalls of over-relying on LLMs, the concept of &quot;Habsburg AI,&quot; and how to avoid disappearing up our own proverbial arseholes in the world of recursive AI processes.</p>

<p>Whether you&#39;re a seasoned AI practitioner, a curious data scientist, or someone interested in the human side of AI development, this conversation offers valuable insights into building more robust, reliable, and human-centered AI systems.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://youtube.com/live/hKV6xSJZkB0?feature=share" rel="nofollow">The livestream on YouTube</a></li>
<li><a href="https://www.sh-reya.com/" rel="nofollow">Shreya&#39;s website</a></li>
<li><a href="https://x.com/sh_reya" rel="nofollow">Shreya on Twitter</a></li>
<li><a href="https://www.sh-reya.com/blog/ai-engineering-flywheel/" rel="nofollow">Data Flywheels for LLM Applications</a></li>
<li><a href="https://arxiv.org/abs/2401.03038" rel="nofollow">SPADE: Synthesizing Data Quality Assertions for Large Language Model Pipelines</a></li>
<li><a href="https://applied-llms.org/" rel="nofollow">What We’ve Learned From A Year of Building with LLMs</a></li>
<li><a href="https://arxiv.org/abs/2404.12272" rel="nofollow">Who Validates the Validators? Aligning LLM-Assisted Evaluation of LLM Outputs with Human Preferences</a></li>
<li><a href="https://arxiv.org/abs/2209.09125" rel="nofollow">Operationalizing Machine Learning: An Interview Study</a></li>
<li><a href="https://twitter.com/vanishingdata" rel="nofollow">Vanishing Gradients on Twitter</a></li>
<li><a href="https://twitter.com/hugobowne" rel="nofollow">Hugo on Twitter</a></li>
</ul>

<p>In the podcast, Hugo also mentioned that this was the 5th time he and Shreya chatted publicly. which is wild!</p>

<p>If you want to dive deep into Shreya&#39;s work and related topics through their chats, you can check them all out here:</p>

<ol>
<li><a href="https://www.youtube.com/watch?v=7zB6ESFto_U" rel="nofollow">Outerbounds&#39; Fireside Chat: Operationalizing ML -- Patterns and Pain Points from MLOps Practitioners</a></li>
<li><a href="https://youtu.be/q0A9CdGWXqc?si=XmaUnQmZiXL2eagS" rel="nofollow">The Past, Present, and Future of Generative AI</a></li>
<li><a href="https://www.youtube.com/live/MTJHvgJtynU?si=Ncjqn5YuFBemvOJ0" rel="nofollow">LLMs, OpenAI Dev Day, and the Existential Crisis for Machine Learning Engineering</a></li>
<li><a href="https://youtube.com/live/c0gcsprsFig?feature=share" rel="nofollow">Lessons from a Year of Building with LLMs</a></li>
</ol>

<p>Check out and subcribe to our <a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">lu.ma calendar</a> for upcoming livestreams!</p>]]>
  </content:encoded>
  <itunes:summary>
    <![CDATA[<p>Hugo speaks with Shreya Shankar, a researcher at UC Berkeley focusing on data management systems with a human-centered approach. Shreya&#39;s work is at the cutting edge of human-computer interaction (HCI) and AI, particularly in the realm of large language models (LLMs). Her impressive background includes being the first ML engineer at Viaduct, doing research engineering at Google Brain, and software engineering at Facebook.</p>

<p>In this episode, we dive deep into the world of LLMs and the critical challenges of building reliable AI pipelines. We&#39;ll explore:</p>

<ul>
<li>The fascinating journey from classic machine learning to the current LLM revolution</li>
<li>Why Shreya believes most ML problems are actually data management issues</li>
<li>The concept of &quot;data flywheels&quot; for LLM applications and how to implement them</li>
<li>The intriguing world of evaluating AI systems - who validates the validators?</li>
<li>Shreya&#39;s work on SPADE and EvalGen, innovative tools for synthesizing data quality assertions and aligning LLM evaluations with human preferences</li>
<li>The importance of human-in-the-loop processes in AI development</li>
<li>The future of low-code and no-code tools in the AI landscape</li>
</ul>

<p>We&#39;ll also touch on the potential pitfalls of over-relying on LLMs, the concept of &quot;Habsburg AI,&quot; and how to avoid disappearing up our own proverbial arseholes in the world of recursive AI processes.</p>

<p>Whether you&#39;re a seasoned AI practitioner, a curious data scientist, or someone interested in the human side of AI development, this conversation offers valuable insights into building more robust, reliable, and human-centered AI systems.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://youtube.com/live/hKV6xSJZkB0?feature=share" rel="nofollow">The livestream on YouTube</a></li>
<li><a href="https://www.sh-reya.com/" rel="nofollow">Shreya&#39;s website</a></li>
<li><a href="https://x.com/sh_reya" rel="nofollow">Shreya on Twitter</a></li>
<li><a href="https://www.sh-reya.com/blog/ai-engineering-flywheel/" rel="nofollow">Data Flywheels for LLM Applications</a></li>
<li><a href="https://arxiv.org/abs/2401.03038" rel="nofollow">SPADE: Synthesizing Data Quality Assertions for Large Language Model Pipelines</a></li>
<li><a href="https://applied-llms.org/" rel="nofollow">What We’ve Learned From A Year of Building with LLMs</a></li>
<li><a href="https://arxiv.org/abs/2404.12272" rel="nofollow">Who Validates the Validators? Aligning LLM-Assisted Evaluation of LLM Outputs with Human Preferences</a></li>
<li><a href="https://arxiv.org/abs/2209.09125" rel="nofollow">Operationalizing Machine Learning: An Interview Study</a></li>
<li><a href="https://twitter.com/vanishingdata" rel="nofollow">Vanishing Gradients on Twitter</a></li>
<li><a href="https://twitter.com/hugobowne" rel="nofollow">Hugo on Twitter</a></li>
</ul>

<p>In the podcast, Hugo also mentioned that this was the 5th time he and Shreya chatted publicly. which is wild!</p>

<p>If you want to dive deep into Shreya&#39;s work and related topics through their chats, you can check them all out here:</p>

<ol>
<li><a href="https://www.youtube.com/watch?v=7zB6ESFto_U" rel="nofollow">Outerbounds&#39; Fireside Chat: Operationalizing ML -- Patterns and Pain Points from MLOps Practitioners</a></li>
<li><a href="https://youtu.be/q0A9CdGWXqc?si=XmaUnQmZiXL2eagS" rel="nofollow">The Past, Present, and Future of Generative AI</a></li>
<li><a href="https://www.youtube.com/live/MTJHvgJtynU?si=Ncjqn5YuFBemvOJ0" rel="nofollow">LLMs, OpenAI Dev Day, and the Existential Crisis for Machine Learning Engineering</a></li>
<li><a href="https://youtube.com/live/c0gcsprsFig?feature=share" rel="nofollow">Lessons from a Year of Building with LLMs</a></li>
</ol>

<p>Check out and subcribe to our <a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">lu.ma calendar</a> for upcoming livestreams!</p>]]>
  </itunes:summary>
</item>
<item>
  <title>Episode 31: Rethinking Data Science, Machine Learning, and AI</title>
  <link>https://vanishinggradients.fireside.fm/31</link>
  <guid isPermaLink="false">455d1587-7ba6-4850-920e-360d8cbe33d3</guid>
  <pubDate>Tue, 09 Jul 2024 19:00:00 +1000</pubDate>
  <author>Hugo Bowne-Anderson</author>
  <enclosure url="https://aphid.fireside.fm/d/1437767933/140c3904-8258-4c39-a698-a112b7077bd7/455d1587-7ba6-4850-920e-360d8cbe33d3.mp3" length="92236825" type="audio/mpeg"/>
  <itunes:episodeType>full</itunes:episodeType>
  <itunes:season>1</itunes:season>
  <itunes:author>Hugo Bowne-Anderson</itunes:author>
  <itunes:subtitle>Hugo speaks with Vincent Warmerdam, a senior data professional and machine learning engineer at :probabl, the exclusive brand operator of scikit-learn. Vincent is known for challenging common assumptions and exploring innovative approaches in data science and machine learning.</itunes:subtitle>
  <itunes:duration>1:36:04</itunes:duration>
  <itunes:explicit>no</itunes:explicit>
  <itunes:image href="https://media24.fireside.fm/file/fireside-images-2024/podcasts/images/1/140c3904-8258-4c39-a698-a112b7077bd7/cover.jpg?v=1"/>
  <description>Hugo speaks with Vincent Warmerdam, a senior data professional and machine learning engineer at :probabl, the exclusive brand operator of scikit-learn. Vincent is known for challenging common assumptions and exploring innovative approaches in data science and machine learning.
In this episode, they dive deep into rethinking established methods in data science, machine learning, and AI. We explore Vincent's principled approach to the field, including:
The critical importance of exposing yourself to real-world problems before applying ML solutions
Framing problems correctly and understanding the data generating process
The power of visualization and human intuition in data analysis
Questioning whether algorithms truly meet the actual problem at hand
The value of simple, interpretable models and when to consider more complex approaches
The importance of UI and user experience in data science tools
Strategies for preventing algorithmic failures by rethinking evaluation metrics and data quality
The potential and limitations of LLMs in the current data science landscape
The benefits of open-source collaboration and knowledge sharing in the community
Throughout the conversation, Vincent illustrates these principles with vivid, real-world examples from his extensive experience in the field. They also discuss Vincent's thoughts on the future of data science and his call to action for more knowledge sharing in the community through blogging and open dialogue.
LINKS
The livestream on YouTube (https://youtube.com/live/-CD66CI1pEo?feature=share)
Vincent's blog (https://koaning.io/)
CalmCode (https://calmcode.io/)
scikit-lego (https://koaning.github.io/scikit-lego/)
Vincent's book Data Science Fiction (WIP) (https://calmcode.io/book)
The Deon Checklist, an ethics checklist for data scientists (https://deon.drivendata.org/)
Of oaths and checklists, by DJ Patil, Hilary Mason and Mike Loukides (https://www.oreilly.com/radar/of-oaths-and-checklists/)
Vincent's Getting Started with NLP and spaCy Course course on Talk Python (https://training.talkpython.fm/courses/getting-started-with-spacy)
Vincent on twitter (https://x.com/fishnets88)
:probabl. on twitter (https://x.com/probabl_ai)
Vincent's PyData Amsterdam Keynote "Natural Intelligence is All You Need [tm]" (https://www.youtube.com/watch?v=C9p7suS-NGk)
Vincent's PyData Amsterdam 2019 talk: The profession of solving (the wrong problem)  (https://www.youtube.com/watch?v=kYMfE9u-lMo)
Vanishing Gradients on Twitter (https://twitter.com/vanishingdata)
Hugo on Twitter (https://twitter.com/hugobowne)
Check out and subcribe to our lu.ma calendar (https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk) for upcoming livestreams! 
</description>
  <itunes:keywords>AI, LLMs, machine learning, data science, GenAI</itunes:keywords>
  <content:encoded>
    <![CDATA[<p>Hugo speaks with Vincent Warmerdam, a senior data professional and machine learning engineer at :probabl, the exclusive brand operator of scikit-learn. Vincent is known for challenging common assumptions and exploring innovative approaches in data science and machine learning.</p>

<p>In this episode, they dive deep into rethinking established methods in data science, machine learning, and AI. We explore Vincent&#39;s principled approach to the field, including:</p>

<ul>
<li>The critical importance of exposing yourself to real-world problems before applying ML solutions</li>
<li>Framing problems correctly and understanding the data generating process</li>
<li>The power of visualization and human intuition in data analysis</li>
<li>Questioning whether algorithms truly meet the actual problem at hand</li>
<li>The value of simple, interpretable models and when to consider more complex approaches</li>
<li>The importance of UI and user experience in data science tools</li>
<li>Strategies for preventing algorithmic failures by rethinking evaluation metrics and data quality</li>
<li>The potential and limitations of LLMs in the current data science landscape</li>
<li>The benefits of open-source collaboration and knowledge sharing in the community</li>
</ul>

<p>Throughout the conversation, Vincent illustrates these principles with vivid, real-world examples from his extensive experience in the field. They also discuss Vincent&#39;s thoughts on the future of data science and his call to action for more knowledge sharing in the community through blogging and open dialogue.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://youtube.com/live/-CD66CI1pEo?feature=share" rel="nofollow">The livestream on YouTube</a></li>
<li><a href="https://koaning.io/" rel="nofollow">Vincent&#39;s blog</a></li>
<li><a href="https://calmcode.io/" rel="nofollow">CalmCode</a></li>
<li><a href="https://koaning.github.io/scikit-lego/" rel="nofollow">scikit-lego</a></li>
<li><a href="https://calmcode.io/book" rel="nofollow">Vincent&#39;s book Data Science Fiction (WIP)</a></li>
<li><a href="https://deon.drivendata.org/" rel="nofollow">The Deon Checklist, an ethics checklist for data scientists</a></li>
<li><a href="https://www.oreilly.com/radar/of-oaths-and-checklists/" rel="nofollow">Of oaths and checklists, by DJ Patil, Hilary Mason and Mike Loukides</a></li>
<li><a href="https://training.talkpython.fm/courses/getting-started-with-spacy" rel="nofollow">Vincent&#39;s Getting Started with NLP and spaCy Course course on Talk Python</a></li>
<li><a href="https://x.com/fishnets88" rel="nofollow">Vincent on twitter</a></li>
<li><a href="https://x.com/probabl_ai" rel="nofollow">:probabl. on twitter</a></li>
<li><a href="https://www.youtube.com/watch?v=C9p7suS-NGk" rel="nofollow">Vincent&#39;s PyData Amsterdam Keynote &quot;Natural Intelligence is All You Need [tm]&quot;</a></li>
<li><a href="https://www.youtube.com/watch?v=kYMfE9u-lMo" rel="nofollow">Vincent&#39;s PyData Amsterdam 2019 talk: The profession of solving (the wrong problem) </a></li>
<li><a href="https://twitter.com/vanishingdata" rel="nofollow">Vanishing Gradients on Twitter</a></li>
<li><a href="https://twitter.com/hugobowne" rel="nofollow">Hugo on Twitter</a></li>
</ul>

<p>Check out and subcribe to our <a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">lu.ma calendar</a> for upcoming livestreams!</p>]]>
  </content:encoded>
  <itunes:summary>
    <![CDATA[<p>Hugo speaks with Vincent Warmerdam, a senior data professional and machine learning engineer at :probabl, the exclusive brand operator of scikit-learn. Vincent is known for challenging common assumptions and exploring innovative approaches in data science and machine learning.</p>

<p>In this episode, they dive deep into rethinking established methods in data science, machine learning, and AI. We explore Vincent&#39;s principled approach to the field, including:</p>

<ul>
<li>The critical importance of exposing yourself to real-world problems before applying ML solutions</li>
<li>Framing problems correctly and understanding the data generating process</li>
<li>The power of visualization and human intuition in data analysis</li>
<li>Questioning whether algorithms truly meet the actual problem at hand</li>
<li>The value of simple, interpretable models and when to consider more complex approaches</li>
<li>The importance of UI and user experience in data science tools</li>
<li>Strategies for preventing algorithmic failures by rethinking evaluation metrics and data quality</li>
<li>The potential and limitations of LLMs in the current data science landscape</li>
<li>The benefits of open-source collaboration and knowledge sharing in the community</li>
</ul>

<p>Throughout the conversation, Vincent illustrates these principles with vivid, real-world examples from his extensive experience in the field. They also discuss Vincent&#39;s thoughts on the future of data science and his call to action for more knowledge sharing in the community through blogging and open dialogue.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://youtube.com/live/-CD66CI1pEo?feature=share" rel="nofollow">The livestream on YouTube</a></li>
<li><a href="https://koaning.io/" rel="nofollow">Vincent&#39;s blog</a></li>
<li><a href="https://calmcode.io/" rel="nofollow">CalmCode</a></li>
<li><a href="https://koaning.github.io/scikit-lego/" rel="nofollow">scikit-lego</a></li>
<li><a href="https://calmcode.io/book" rel="nofollow">Vincent&#39;s book Data Science Fiction (WIP)</a></li>
<li><a href="https://deon.drivendata.org/" rel="nofollow">The Deon Checklist, an ethics checklist for data scientists</a></li>
<li><a href="https://www.oreilly.com/radar/of-oaths-and-checklists/" rel="nofollow">Of oaths and checklists, by DJ Patil, Hilary Mason and Mike Loukides</a></li>
<li><a href="https://training.talkpython.fm/courses/getting-started-with-spacy" rel="nofollow">Vincent&#39;s Getting Started with NLP and spaCy Course course on Talk Python</a></li>
<li><a href="https://x.com/fishnets88" rel="nofollow">Vincent on twitter</a></li>
<li><a href="https://x.com/probabl_ai" rel="nofollow">:probabl. on twitter</a></li>
<li><a href="https://www.youtube.com/watch?v=C9p7suS-NGk" rel="nofollow">Vincent&#39;s PyData Amsterdam Keynote &quot;Natural Intelligence is All You Need [tm]&quot;</a></li>
<li><a href="https://www.youtube.com/watch?v=kYMfE9u-lMo" rel="nofollow">Vincent&#39;s PyData Amsterdam 2019 talk: The profession of solving (the wrong problem) </a></li>
<li><a href="https://twitter.com/vanishingdata" rel="nofollow">Vanishing Gradients on Twitter</a></li>
<li><a href="https://twitter.com/hugobowne" rel="nofollow">Hugo on Twitter</a></li>
</ul>

<p>Check out and subcribe to our <a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">lu.ma calendar</a> for upcoming livestreams!</p>]]>
  </itunes:summary>
</item>
<item>
  <title>Episode 28: Beyond Supervised Learning: The Rise of In-Context Learning with LLMs</title>
  <link>https://vanishinggradients.fireside.fm/28</link>
  <guid isPermaLink="false">b268a89e-4fc9-4f9f-a2a5-c7636b3fbd70</guid>
  <pubDate>Mon, 10 Jun 2024 08:00:00 +1000</pubDate>
  <author>Hugo Bowne-Anderson</author>
  <enclosure url="https://aphid.fireside.fm/d/1437767933/140c3904-8258-4c39-a698-a112b7077bd7/b268a89e-4fc9-4f9f-a2a5-c7636b3fbd70.mp3" length="63014789" type="audio/mpeg"/>
  <itunes:episodeType>full</itunes:episodeType>
  <itunes:season>1</itunes:season>
  <itunes:author>Hugo Bowne-Anderson</itunes:author>
  <itunes:subtitle>Hugo speaks with Alan Nichol, co-founder and CTO of Rasa, where they build software to enable developers to create enterprise-grade conversational AI and chatbot systems across industries like telcos, healthcare, fintech, and government.</itunes:subtitle>
  <itunes:duration>1:05:38</itunes:duration>
  <itunes:explicit>no</itunes:explicit>
  <itunes:image href="https://media24.fireside.fm/file/fireside-images-2024/podcasts/images/1/140c3904-8258-4c39-a698-a112b7077bd7/cover.jpg?v=1"/>
  <description>Hugo speaks with Alan Nichol, co-founder and CTO of Rasa, where they build software to enable developers to create enterprise-grade conversational AI and chatbot systems across industries like telcos, healthcare, fintech, and government.
What's super cool is that Alan and the Rasa team have been doing this type of thing for over a decade, giving them a wealth of wisdom on how to effectively incorporate LLMs into chatbots - and how not to. For example, if you want a chatbot that takes specific and important actions like transferring money, do you want to fully entrust the conversation to one big LLM like ChatGPT, or secure what the LLMs can do inside key business logic?
In this episode, they also dive into the history of conversational AI and explore how the advent of LLMs is reshaping the field. Alan shares his perspective on how supervised learning has failed us in some ways and discusses what he sees as the most overrated and underrated aspects of LLMs.
Alan offers advice for those looking to work with LLMs and conversational AI, emphasizing the importance of not sleeping on proven techniques and looking beyond the latest hype. In a live demo, he showcases Rasa's Calm (Conversational AI with Language Models), which allows developers to define business logic declaratively and separate it from the LLM, enabling reliable execution of conversational flows.
LINKS
The livestream on YouTube (https://www.youtube.com/live/kMFBYC2pB30?si=yV5sGq1iuC47LBSi)
Alan's Rasa CALM Demo: Building Conversational AI with LLMs  (https://youtu.be/4UnxaJ-GcT0?si=6uLY3GD5DkOmWiBW)
Alan on twitter.com (https://x.com/alanmnichol)
Rasa (https://rasa.com/)
CALM, an LLM-native approach to building reliable conversational AI (https://rasa.com/docs/rasa-pro/calm/)
Task-Oriented Dialogue with In-Context Learning (https://arxiv.org/abs/2402.12234)
'We don’t know how to build conversational software yet' by Alan Nicol (https://medium.com/rasa-blog/we-don-t-know-how-to-build-conversational-software-yet-a18301db0e4b)
Vanishing Gradients on Twitter (https://twitter.com/vanishingdata)
Hugo on Twitter (https://twitter.com/hugobowne)
Upcoming Livestreams
Lessons from a Year of Building with LLMs (https://lu.ma/e8huz3s6?utm_source=vgan)
VALIDATING THE VALIDATORS with Shreya Shanker (https://lu.ma/zz3qic45?utm_source=vgan)
</description>
  <itunes:keywords>AI, LLMs, machine learning, data science</itunes:keywords>
  <content:encoded>
    <![CDATA[<p>Hugo speaks with Alan Nichol, co-founder and CTO of Rasa, where they build software to enable developers to create enterprise-grade conversational AI and chatbot systems across industries like telcos, healthcare, fintech, and government.</p>

<p>What&#39;s super cool is that Alan and the Rasa team have been doing this type of thing for over a decade, giving them a wealth of wisdom on how to effectively incorporate LLMs into chatbots - and how not to. For example, if you want a chatbot that takes specific and important actions like transferring money, do you want to fully entrust the conversation to one big LLM like ChatGPT, or secure what the LLMs can do inside key business logic?</p>

<p>In this episode, they also dive into the history of conversational AI and explore how the advent of LLMs is reshaping the field. Alan shares his perspective on how supervised learning has failed us in some ways and discusses what he sees as the most overrated and underrated aspects of LLMs.</p>

<p>Alan offers advice for those looking to work with LLMs and conversational AI, emphasizing the importance of not sleeping on proven techniques and looking beyond the latest hype. In a live demo, he showcases Rasa&#39;s Calm (Conversational AI with Language Models), which allows developers to define business logic declaratively and separate it from the LLM, enabling reliable execution of conversational flows.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://www.youtube.com/live/kMFBYC2pB30?si=yV5sGq1iuC47LBSi" rel="nofollow">The livestream on YouTube</a></li>
<li><a href="https://youtu.be/4UnxaJ-GcT0?si=6uLY3GD5DkOmWiBW" rel="nofollow">Alan&#39;s Rasa CALM Demo: Building Conversational AI with LLMs </a></li>
<li><a href="https://x.com/alanmnichol" rel="nofollow">Alan on twitter.com</a></li>
<li><a href="https://rasa.com/" rel="nofollow">Rasa</a></li>
<li><a href="https://rasa.com/docs/rasa-pro/calm/" rel="nofollow">CALM, an LLM-native approach to building reliable conversational AI</a></li>
<li><a href="https://arxiv.org/abs/2402.12234" rel="nofollow">Task-Oriented Dialogue with In-Context Learning</a></li>
<li><a href="https://medium.com/rasa-blog/we-don-t-know-how-to-build-conversational-software-yet-a18301db0e4b" rel="nofollow">&#39;We don’t know how to build conversational software yet&#39; by Alan Nicol</a></li>
<li><a href="https://twitter.com/vanishingdata" rel="nofollow">Vanishing Gradients on Twitter</a></li>
<li><a href="https://twitter.com/hugobowne" rel="nofollow">Hugo on Twitter</a></li>
</ul>

<p><strong>Upcoming Livestreams</strong></p>

<ul>
<li><a href="https://lu.ma/e8huz3s6?utm_source=vgan" rel="nofollow">Lessons from a Year of Building with LLMs</a></li>
<li><a href="https://lu.ma/zz3qic45?utm_source=vgan" rel="nofollow">VALIDATING THE VALIDATORS with Shreya Shanker</a></li>
</ul>]]>
  </content:encoded>
  <itunes:summary>
    <![CDATA[<p>Hugo speaks with Alan Nichol, co-founder and CTO of Rasa, where they build software to enable developers to create enterprise-grade conversational AI and chatbot systems across industries like telcos, healthcare, fintech, and government.</p>

<p>What&#39;s super cool is that Alan and the Rasa team have been doing this type of thing for over a decade, giving them a wealth of wisdom on how to effectively incorporate LLMs into chatbots - and how not to. For example, if you want a chatbot that takes specific and important actions like transferring money, do you want to fully entrust the conversation to one big LLM like ChatGPT, or secure what the LLMs can do inside key business logic?</p>

<p>In this episode, they also dive into the history of conversational AI and explore how the advent of LLMs is reshaping the field. Alan shares his perspective on how supervised learning has failed us in some ways and discusses what he sees as the most overrated and underrated aspects of LLMs.</p>

<p>Alan offers advice for those looking to work with LLMs and conversational AI, emphasizing the importance of not sleeping on proven techniques and looking beyond the latest hype. In a live demo, he showcases Rasa&#39;s Calm (Conversational AI with Language Models), which allows developers to define business logic declaratively and separate it from the LLM, enabling reliable execution of conversational flows.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://www.youtube.com/live/kMFBYC2pB30?si=yV5sGq1iuC47LBSi" rel="nofollow">The livestream on YouTube</a></li>
<li><a href="https://youtu.be/4UnxaJ-GcT0?si=6uLY3GD5DkOmWiBW" rel="nofollow">Alan&#39;s Rasa CALM Demo: Building Conversational AI with LLMs </a></li>
<li><a href="https://x.com/alanmnichol" rel="nofollow">Alan on twitter.com</a></li>
<li><a href="https://rasa.com/" rel="nofollow">Rasa</a></li>
<li><a href="https://rasa.com/docs/rasa-pro/calm/" rel="nofollow">CALM, an LLM-native approach to building reliable conversational AI</a></li>
<li><a href="https://arxiv.org/abs/2402.12234" rel="nofollow">Task-Oriented Dialogue with In-Context Learning</a></li>
<li><a href="https://medium.com/rasa-blog/we-don-t-know-how-to-build-conversational-software-yet-a18301db0e4b" rel="nofollow">&#39;We don’t know how to build conversational software yet&#39; by Alan Nicol</a></li>
<li><a href="https://twitter.com/vanishingdata" rel="nofollow">Vanishing Gradients on Twitter</a></li>
<li><a href="https://twitter.com/hugobowne" rel="nofollow">Hugo on Twitter</a></li>
</ul>

<p><strong>Upcoming Livestreams</strong></p>

<ul>
<li><a href="https://lu.ma/e8huz3s6?utm_source=vgan" rel="nofollow">Lessons from a Year of Building with LLMs</a></li>
<li><a href="https://lu.ma/zz3qic45?utm_source=vgan" rel="nofollow">VALIDATING THE VALIDATORS with Shreya Shanker</a></li>
</ul>]]>
  </itunes:summary>
</item>
<item>
  <title>Episode 27: How to Build Terrible AI Systems</title>
  <link>https://vanishinggradients.fireside.fm/27</link>
  <guid isPermaLink="false">d42a2479-a220-4f72-bf48-946c4a393efa</guid>
  <pubDate>Fri, 31 May 2024 10:00:00 +1000</pubDate>
  <author>Hugo Bowne-Anderson</author>
  <enclosure url="https://aphid.fireside.fm/d/1437767933/140c3904-8258-4c39-a698-a112b7077bd7/d42a2479-a220-4f72-bf48-946c4a393efa.mp3" length="88718026" type="audio/mpeg"/>
  <itunes:episodeType>full</itunes:episodeType>
  <itunes:season>1</itunes:season>
  <itunes:author>Hugo Bowne-Anderson</itunes:author>
  <itunes:subtitle>Hugo speaks with Jason Liu, an independent consultant who uses his expertise in recommendation systems to help fast-growing startups build out their RAG applications. He was previously at Meta and Stitch Fix is also the creator of Instructor, Flight, and an ML and data science educator.</itunes:subtitle>
  <itunes:duration>1:32:24</itunes:duration>
  <itunes:explicit>no</itunes:explicit>
  <itunes:image href="https://media24.fireside.fm/file/fireside-images-2024/podcasts/images/1/140c3904-8258-4c39-a698-a112b7077bd7/cover.jpg?v=1"/>
  <description>Hugo speaks with Jason Liu, an independent consultant who uses his expertise in recommendation systems to help fast-growing startups build out their RAG applications. He was previously at Meta and Stitch Fix is also the creator of Instructor, Flight, and an ML and data science educator.
They talk about how Jason approaches consulting companies across many industries, including construction and sales, in building production LLM apps, his playbook for getting ML and AI up and running to build and maintain such apps, and the future of tooling to do so.
They take an inverted thinking approach, envisaging all the failure modes that would result in building terrible AI systems, and then figure out how to avoid such pitfalls.
LINKS
The livestream on YouTube (https://youtube.com/live/USTG6sQlB6s?feature=share)
Jason's website (https://jxnl.co/)
PyDdantic is all you need, Jason's Keynote at AI Engineer Summit, 2023 (https://youtu.be/yj-wSRJwrrc?si=JIGhN0mx0i50dUR9)
How to build a terrible RAG system by Jason (https://jxnl.co/writing/2024/01/07/inverted-thinking-rag/)
To express interest in Jason's Systematically improving RAG Applications course (https://q7gjsgfstrp.typeform.com/ragcourse?typeform-source=vg)
Vanishing Gradients on Twitter (https://twitter.com/vanishingdata)
Hugo on Twitter (https://twitter.com/hugobowne)
Upcoming Livestreams
Good Riddance to Supervised Learning with Alan Nichol (CTO and co-founder, Rasa) (https://lu.ma/gphzzyyn?utm_source=vgj)
Lessons from a Year of Building with LLMs (https://lu.ma/e8huz3s6?utm_source=vgj)
</description>
  <itunes:keywords>AI, LLMs, machine learning, data science</itunes:keywords>
  <content:encoded>
    <![CDATA[<p>Hugo speaks with Jason Liu, an independent consultant who uses his expertise in recommendation systems to help fast-growing startups build out their RAG applications. He was previously at Meta and Stitch Fix is also the creator of Instructor, Flight, and an ML and data science educator.</p>

<p>They talk about how Jason approaches consulting companies across many industries, including construction and sales, in building production LLM apps, his playbook for getting ML and AI up and running to build and maintain such apps, and the future of tooling to do so.</p>

<p>They take an inverted thinking approach, envisaging all the failure modes that would result in building terrible AI systems, and then figure out how to avoid such pitfalls.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://youtube.com/live/USTG6sQlB6s?feature=share" rel="nofollow">The livestream on YouTube</a></li>
<li><a href="https://jxnl.co/" rel="nofollow">Jason&#39;s website</a></li>
<li><a href="https://youtu.be/yj-wSRJwrrc?si=JIGhN0mx0i50dUR9" rel="nofollow">PyDdantic is all you need, Jason&#39;s Keynote at AI Engineer Summit, 2023</a></li>
<li><a href="https://jxnl.co/writing/2024/01/07/inverted-thinking-rag/" rel="nofollow">How to build a terrible RAG system by Jason</a></li>
<li><a href="https://q7gjsgfstrp.typeform.com/ragcourse?typeform-source=vg" rel="nofollow">To express interest in Jason&#39;s <em>Systematically improving RAG Applications</em> course</a></li>
<li><a href="https://twitter.com/vanishingdata" rel="nofollow">Vanishing Gradients on Twitter</a></li>
<li><a href="https://twitter.com/hugobowne" rel="nofollow">Hugo on Twitter</a></li>
</ul>

<p><strong>Upcoming Livestreams</strong></p>

<ul>
<li><a href="https://lu.ma/gphzzyyn?utm_source=vgj" rel="nofollow">Good Riddance to Supervised Learning with Alan Nichol (CTO and co-founder, Rasa)</a></li>
<li><a href="https://lu.ma/e8huz3s6?utm_source=vgj" rel="nofollow">Lessons from a Year of Building with LLMs</a></li>
</ul>]]>
  </content:encoded>
  <itunes:summary>
    <![CDATA[<p>Hugo speaks with Jason Liu, an independent consultant who uses his expertise in recommendation systems to help fast-growing startups build out their RAG applications. He was previously at Meta and Stitch Fix is also the creator of Instructor, Flight, and an ML and data science educator.</p>

<p>They talk about how Jason approaches consulting companies across many industries, including construction and sales, in building production LLM apps, his playbook for getting ML and AI up and running to build and maintain such apps, and the future of tooling to do so.</p>

<p>They take an inverted thinking approach, envisaging all the failure modes that would result in building terrible AI systems, and then figure out how to avoid such pitfalls.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://youtube.com/live/USTG6sQlB6s?feature=share" rel="nofollow">The livestream on YouTube</a></li>
<li><a href="https://jxnl.co/" rel="nofollow">Jason&#39;s website</a></li>
<li><a href="https://youtu.be/yj-wSRJwrrc?si=JIGhN0mx0i50dUR9" rel="nofollow">PyDdantic is all you need, Jason&#39;s Keynote at AI Engineer Summit, 2023</a></li>
<li><a href="https://jxnl.co/writing/2024/01/07/inverted-thinking-rag/" rel="nofollow">How to build a terrible RAG system by Jason</a></li>
<li><a href="https://q7gjsgfstrp.typeform.com/ragcourse?typeform-source=vg" rel="nofollow">To express interest in Jason&#39;s <em>Systematically improving RAG Applications</em> course</a></li>
<li><a href="https://twitter.com/vanishingdata" rel="nofollow">Vanishing Gradients on Twitter</a></li>
<li><a href="https://twitter.com/hugobowne" rel="nofollow">Hugo on Twitter</a></li>
</ul>

<p><strong>Upcoming Livestreams</strong></p>

<ul>
<li><a href="https://lu.ma/gphzzyyn?utm_source=vgj" rel="nofollow">Good Riddance to Supervised Learning with Alan Nichol (CTO and co-founder, Rasa)</a></li>
<li><a href="https://lu.ma/e8huz3s6?utm_source=vgj" rel="nofollow">Lessons from a Year of Building with LLMs</a></li>
</ul>]]>
  </itunes:summary>
</item>
  </channel>
</rss>
