<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" encoding="UTF-8" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:sy="http://purl.org/rss/1.0/modules/syndication/" xmlns:admin="http://webns.net/mvcb/" xmlns:atom="http://www.w3.org/2005/Atom/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:fireside="http://fireside.fm/modules/rss/fireside">
  <channel>
    <fireside:hostname>web02.fireside.fm</fireside:hostname>
    <fireside:genDate>Sat, 04 Apr 2026 16:39:36 -0500</fireside:genDate>
    <generator>Fireside (https://fireside.fm)</generator>
    <title>Vanishing Gradients - Episodes Tagged with “Genai”</title>
    <link>https://vanishinggradients.fireside.fm/tags/genai</link>
    <pubDate>Sat, 22 Nov 2025 18:30:00 +1100</pubDate>
    <description>A podcast about all things data, brought to you by data scientist Hugo Bowne-Anderson.
It's time for more critical conversations about the challenges in our industry in order to build better compasses for the solution space! To this end, this podcast will consist of long-format conversations between Hugo and other people who work broadly in the data science, machine learning, and AI spaces. We'll dive deep into all the moving parts of the data world, so if you're new to the space, you'll have an opportunity to learn from the experts. And if you've been around for a while, you'll find out what's happening in many other parts of the data world.
</description>
    <language>en-us</language>
    <itunes:type>episodic</itunes:type>
    <itunes:subtitle>a data podcast with hugo bowne-anderson</itunes:subtitle>
    <itunes:author>Hugo Bowne-Anderson</itunes:author>
    <itunes:summary>A podcast about all things data, brought to you by data scientist Hugo Bowne-Anderson.
It's time for more critical conversations about the challenges in our industry in order to build better compasses for the solution space! To this end, this podcast will consist of long-format conversations between Hugo and other people who work broadly in the data science, machine learning, and AI spaces. We'll dive deep into all the moving parts of the data world, so if you're new to the space, you'll have an opportunity to learn from the experts. And if you've been around for a while, you'll find out what's happening in many other parts of the data world.
</itunes:summary>
    <itunes:image href="https://media24.fireside.fm/file/fireside-images-2024/podcasts/images/1/140c3904-8258-4c39-a698-a112b7077bd7/cover.jpg?v=1"/>
    <itunes:explicit>no</itunes:explicit>
    <itunes:keywords>data science, machine learning, AI</itunes:keywords>
    <itunes:owner>
      <itunes:name>Hugo Bowne-Anderson</itunes:name>
      <itunes:email>hugobowne@hey.com</itunes:email>
    </itunes:owner>
<itunes:category text="Technology"/>
<item>
  <title>Episode 63: Why Gemini 3 Will Change How You Build AI Agents with Ravin Kumar (Google DeepMind)</title>
  <link>https://vanishinggradients.fireside.fm/63</link>
  <guid isPermaLink="false">cc45813e-f7ec-434d-a816-e7a5dfba8946</guid>
  <pubDate>Sat, 22 Nov 2025 18:30:00 +1100</pubDate>
  <author>Hugo Bowne-Anderson</author>
  <enclosure url="https://aphid.fireside.fm/d/1437767933/140c3904-8258-4c39-a698-a112b7077bd7/cc45813e-f7ec-434d-a816-e7a5dfba8946.mp3" length="117720514" type="audio/mpeg"/>
  <itunes:episodeType>full</itunes:episodeType>
  <itunes:author>Hugo Bowne-Anderson</itunes:author>
  <itunes:subtitle>Gemini 3 is a few days old and the massive leap in performance and model reasoning has big implications for builders: as models begin to self-heal, builders are literally tearing out the functionality their built just months ago: ripping out the defensive coding and reshipping their agent harnesses entirely.

Ravin Kumar (Google DeepMind) joins Hugo to breaks down exactly why the rapid evolution of models like Gemini 3 is changing how we build software. They detail the shift from simple tool calling to building reliable "Agent Harnesses", explore the architectural tradeoffs between deterministic workflows and high-agency systems, the nuance of preventing context rot in massive windows, and why proper evaluation infrastructure is the only way to manage the chaos of autonomous loops.</itunes:subtitle>
  <itunes:duration>1:00:12</itunes:duration>
  <itunes:explicit>no</itunes:explicit>
  <itunes:image href="https://media24.fireside.fm/file/fireside-images-2024/podcasts/images/1/140c3904-8258-4c39-a698-a112b7077bd7/cover.jpg?v=1"/>
  <description>Gemini 3 is a few days old and the massive leap in performance and model reasoning has big implications for builders: as models begin to self-heal, builders are literally tearing out the functionality they built just months ago... ripping out the defensive coding and reshipping their agent harnesses entirely.
Ravin Kumar (Google DeepMind) joins Hugo to breaks down exactly why the rapid evolution of models like Gemini 3 is changing how we build software. They detail the shift from simple tool calling to building reliable "Agent Harnesses", explore the architectural tradeoffs between deterministic workflows and high-agency systems, the nuance of preventing context rot in massive windows, and why proper evaluation infrastructure is the only way to manage the chaos of autonomous loops.
They talk through:
- The implications of models that can "self-heal" and fix their own code
- The two cultures of agents: LLM workflows with a few tools versus when you should unleash high-agency, autonomous systems.
- Inside NotebookLM: moving from prototypes to viral production features like Audio Overviews
- Why Needle in a Haystack benchmarks often fail to predict real-world performance
- How to build agent harnesses that turn model capabilities into product velocity
- The shift from measuring latency to managing time-to-compute for reasoning tasks
LINKS
From Context Engineering to AI Agent Harnesses: The New Software Discipline, a podcast Hugo did with Lance Martin, LangChain (https://high-signal.delphina.ai/episode/context-engineering-to-ai-agent-harnesses-the-new-software-discipline)
Context Rot: How Increasing Input Tokens Impacts LLM Performance (https://research.trychroma.com/context-rot)
Effective context engineering for AI agents by Anthropic (https://www.anthropic.com/engineering/effective-context-engineering-for-ai-agents)
Upcoming Events on Luma (https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk)
Watch the podcast video on YouTube (https://youtu.be/CloimQsQuJM)
Join the final cohort of our Building AI Applications course starting March 10, 2026 (25% off for listeners) (https://maven.com/hugo-stefan/building-ai-apps-ds-and-swe-from-first-principles?promoCode=vgfs): https://maven.com/hugo-stefan/building-ai-apps-ds-and-swe-from-first-principles?promoCode=vgfs 
</description>
  <itunes:keywords>ai, genai, machine learning</itunes:keywords>
  <content:encoded>
    <![CDATA[<p>Gemini 3 is a few days old and the massive leap in performance and model reasoning has big implications for builders: as models begin to self-heal, builders are literally tearing out the functionality they built just months ago... ripping out the defensive coding and reshipping their agent harnesses entirely.</p>

<p>Ravin Kumar (Google DeepMind) joins Hugo to breaks down exactly why the rapid evolution of models like Gemini 3 is changing how we build software. They detail the shift from simple tool calling to building reliable &quot;Agent Harnesses&quot;, explore the architectural tradeoffs between deterministic workflows and high-agency systems, the nuance of preventing context rot in massive windows, and why proper evaluation infrastructure is the only way to manage the chaos of autonomous loops.</p>

<p>They talk through:</p>

<ul>
<li>The implications of models that can &quot;self-heal&quot; and fix their own code</li>
<li>The two cultures of agents: LLM workflows with a few tools versus when you should unleash high-agency, autonomous systems.</li>
<li>Inside NotebookLM: moving from prototypes to viral production features like Audio Overviews</li>
<li>Why Needle in a Haystack benchmarks often fail to predict real-world performance</li>
<li>How to build agent harnesses that turn model capabilities into product velocity</li>
<li>The shift from measuring latency to managing time-to-compute for reasoning tasks</li>
</ul>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://high-signal.delphina.ai/episode/context-engineering-to-ai-agent-harnesses-the-new-software-discipline" rel="nofollow">From Context Engineering to AI Agent Harnesses: The New Software Discipline, a podcast Hugo did with Lance Martin, LangChain</a></li>
<li><a href="https://research.trychroma.com/context-rot" rel="nofollow">Context Rot: How Increasing Input Tokens Impacts LLM Performance</a></li>
<li><a href="https://www.anthropic.com/engineering/effective-context-engineering-for-ai-agents" rel="nofollow">Effective context engineering for AI agents by Anthropic</a></li>
<li><a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">Upcoming Events on Luma</a></li>
<li><a href="https://youtu.be/CloimQsQuJM" rel="nofollow">Watch the podcast video on YouTube</a></li>
</ul>

<p><a href="https://maven.com/hugo-stefan/building-ai-apps-ds-and-swe-from-first-principles?promoCode=vgfs" rel="nofollow">Join the final cohort of our Building AI Applications course starting March 10, 2026 (25% off for listeners)</a>: <a href="https://maven.com/hugo-stefan/building-ai-apps-ds-and-swe-from-first-principles?promoCode=vgfs" rel="nofollow">https://maven.com/hugo-stefan/building-ai-apps-ds-and-swe-from-first-principles?promoCode=vgfs</a></p>]]>
  </content:encoded>
  <itunes:summary>
    <![CDATA[<p>Gemini 3 is a few days old and the massive leap in performance and model reasoning has big implications for builders: as models begin to self-heal, builders are literally tearing out the functionality they built just months ago... ripping out the defensive coding and reshipping their agent harnesses entirely.</p>

<p>Ravin Kumar (Google DeepMind) joins Hugo to breaks down exactly why the rapid evolution of models like Gemini 3 is changing how we build software. They detail the shift from simple tool calling to building reliable &quot;Agent Harnesses&quot;, explore the architectural tradeoffs between deterministic workflows and high-agency systems, the nuance of preventing context rot in massive windows, and why proper evaluation infrastructure is the only way to manage the chaos of autonomous loops.</p>

<p>They talk through:</p>

<ul>
<li>The implications of models that can &quot;self-heal&quot; and fix their own code</li>
<li>The two cultures of agents: LLM workflows with a few tools versus when you should unleash high-agency, autonomous systems.</li>
<li>Inside NotebookLM: moving from prototypes to viral production features like Audio Overviews</li>
<li>Why Needle in a Haystack benchmarks often fail to predict real-world performance</li>
<li>How to build agent harnesses that turn model capabilities into product velocity</li>
<li>The shift from measuring latency to managing time-to-compute for reasoning tasks</li>
</ul>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://high-signal.delphina.ai/episode/context-engineering-to-ai-agent-harnesses-the-new-software-discipline" rel="nofollow">From Context Engineering to AI Agent Harnesses: The New Software Discipline, a podcast Hugo did with Lance Martin, LangChain</a></li>
<li><a href="https://research.trychroma.com/context-rot" rel="nofollow">Context Rot: How Increasing Input Tokens Impacts LLM Performance</a></li>
<li><a href="https://www.anthropic.com/engineering/effective-context-engineering-for-ai-agents" rel="nofollow">Effective context engineering for AI agents by Anthropic</a></li>
<li><a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">Upcoming Events on Luma</a></li>
<li><a href="https://youtu.be/CloimQsQuJM" rel="nofollow">Watch the podcast video on YouTube</a></li>
</ul>

<p><a href="https://maven.com/hugo-stefan/building-ai-apps-ds-and-swe-from-first-principles?promoCode=vgfs" rel="nofollow">Join the final cohort of our Building AI Applications course starting March 10, 2026 (25% off for listeners)</a>: <a href="https://maven.com/hugo-stefan/building-ai-apps-ds-and-swe-from-first-principles?promoCode=vgfs" rel="nofollow">https://maven.com/hugo-stefan/building-ai-apps-ds-and-swe-from-first-principles?promoCode=vgfs</a></p>]]>
  </itunes:summary>
</item>
<item>
  <title>Episode 60: 10 Things I Hate About AI Evals with Hamel Husain</title>
  <link>https://vanishinggradients.fireside.fm/60</link>
  <guid isPermaLink="false">0fbc2a65-3bfc-4f8a-83ac-d370f1a30e13</guid>
  <pubDate>Tue, 30 Sep 2025 17:30:00 +1000</pubDate>
  <author>Hugo Bowne-Anderson</author>
  <enclosure url="https://aphid.fireside.fm/d/1437767933/140c3904-8258-4c39-a698-a112b7077bd7/0fbc2a65-3bfc-4f8a-83ac-d370f1a30e13.mp3" length="105505355" type="audio/mpeg"/>
  <itunes:episodeType>full</itunes:episodeType>
  <itunes:author>Hugo Bowne-Anderson</itunes:author>
  <itunes:subtitle>Most AI teams find "evals" frustrating, but ML Engineer Hamel Husain argues they’re just using the wrong playbook. In this episode, he lays out a data-centric approach to systematically measure and improve AI, turning unreliable prototypes into robust, production-ready systems.
</itunes:subtitle>
  <itunes:duration>1:13:15</itunes:duration>
  <itunes:explicit>no</itunes:explicit>
  <itunes:image href="https://media24.fireside.fm/file/fireside-images-2024/podcasts/images/1/140c3904-8258-4c39-a698-a112b7077bd7/cover.jpg?v=1"/>
  <description>Most AI teams find "evals" frustrating, but ML Engineer Hamel Husain argues they’re just using the wrong playbook. In this episode, he lays out a data-centric approach to systematically measure and improve AI, turning unreliable prototypes into robust, production-ready systems.
Drawing from his experience getting countless teams unstuck, Hamel explains why the solution requires a "revenge of the data scientists." He details the essential mindset shifts, error analysis techniques, and practical steps needed to move beyond guesswork and build AI products you can actually trust.
We talk through:
  The 10(+1) critical mistakes that cause teams to waste time on evals
  Why "hallucination scores" are a waste of time (and what to measure instead)
  The manual review process that finds major issues in hours, not weeks
  A step-by-step method for building LLM judges you can actually trust
  How to use domain experts without getting stuck in endless review committees
  Guest Bryan Bischof's "Failure as a Funnel" for debugging complex AI agents
If you're tired of ambiguous "vibe checks" and want a clear process that delivers real improvement, this episode provides the definitive roadmap.
LINKS
Hamel's website and blog (https://hamel.dev/)
Hugo speaks with Philip Carter (Honeycomb) about aligning your LLM-as-a-judge with your domain expertise (https://vanishinggradients.fireside.fm/51)
Hamel Husain on Lenny's pocast, which includes a live demo of error analysis (https://www.lennysnewsletter.com/p/why-ai-evals-are-the-hottest-new-skill)
The episode of VG in which Hamel and Hugo talk about Hamel's "data consulting in Vegas" era (https://vanishinggradients.fireside.fm/9)
Upcoming Events on Luma (https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk)
Watch the podcast video on YouTube (https://youtube.com/live/QEk-XwrkqhI?feature=share)
Hamel's AI evals course, which he teaches with Shreya Shankar (UC Berkeley): starts Oct 6 and this link gives 35% off! (https://maven.com/parlance-labs/evals?promoCode=GOHUGORGOHOME) https://maven.com/parlance-labs/evals?promoCode=GOHUGORGOHOME
🎓 Learn more:
Hugo's course: Building LLM Applications for Data Scientists and Software Engineers (https://maven.com/s/course/d56067f338) — https://maven.com/s/course/d56067f338  
</description>
  <itunes:keywords>AI, GenAI, LLMs, data science, machine learning, evals</itunes:keywords>
  <content:encoded>
    <![CDATA[<p>Most AI teams find &quot;evals&quot; frustrating, but ML Engineer Hamel Husain argues they’re just using the wrong playbook. In this episode, he lays out a data-centric approach to systematically measure and improve AI, turning unreliable prototypes into robust, production-ready systems.</p>

<p>Drawing from his experience getting countless teams unstuck, Hamel explains why the solution requires a &quot;revenge of the data scientists.&quot; He details the essential mindset shifts, error analysis techniques, and practical steps needed to move beyond guesswork and build AI products you can actually trust.</p>

<p>We talk through:</p>

<ul>
<li>  The 10(+1) critical mistakes that cause teams to waste time on evals</li>
<li>  Why &quot;hallucination scores&quot; are a waste of time (and what to measure instead)</li>
<li>  The manual review process that finds major issues in hours, not weeks</li>
<li>  A step-by-step method for building LLM judges you can actually trust</li>
<li>  How to use domain experts without getting stuck in endless review committees</li>
<li>  Guest Bryan Bischof&#39;s &quot;Failure as a Funnel&quot; for debugging complex AI agents</li>
</ul>

<p>If you&#39;re tired of ambiguous &quot;vibe checks&quot; and want a clear process that delivers real improvement, this episode provides the definitive roadmap.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://hamel.dev/" rel="nofollow">Hamel&#39;s website and blog</a></li>
<li><a href="https://vanishinggradients.fireside.fm/51" rel="nofollow">Hugo speaks with Philip Carter (Honeycomb) about aligning your LLM-as-a-judge with your domain expertise</a></li>
<li><a href="https://www.lennysnewsletter.com/p/why-ai-evals-are-the-hottest-new-skill" rel="nofollow">Hamel Husain on Lenny&#39;s pocast, which includes a live demo of error analysis</a></li>
<li><a href="https://vanishinggradients.fireside.fm/9" rel="nofollow">The episode of VG in which Hamel and Hugo talk about Hamel&#39;s &quot;data consulting in Vegas&quot; era</a></li>
<li><a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">Upcoming Events on Luma</a></li>
<li><a href="https://youtube.com/live/QEk-XwrkqhI?feature=share" rel="nofollow">Watch the podcast video on YouTube</a></li>
<li><a href="https://maven.com/parlance-labs/evals?promoCode=GOHUGORGOHOME" rel="nofollow">Hamel&#39;s AI evals course, which he teaches with Shreya Shankar (UC Berkeley): starts Oct 6 and this link gives 35% off!</a> <a href="https://maven.com/parlance-labs/evals?promoCode=GOHUGORGOHOME" rel="nofollow">https://maven.com/parlance-labs/evals?promoCode=GOHUGORGOHOME</a></li>
</ul>

<p>🎓 Learn more:</p>

<ul>
<li><strong>Hugo&#39;s course:</strong> <a href="https://maven.com/s/course/d56067f338" rel="nofollow">Building LLM Applications for Data Scientists and Software Engineers</a> — <a href="https://maven.com/s/course/d56067f338" rel="nofollow">https://maven.com/s/course/d56067f338</a> </li>
</ul>]]>
  </content:encoded>
  <itunes:summary>
    <![CDATA[<p>Most AI teams find &quot;evals&quot; frustrating, but ML Engineer Hamel Husain argues they’re just using the wrong playbook. In this episode, he lays out a data-centric approach to systematically measure and improve AI, turning unreliable prototypes into robust, production-ready systems.</p>

<p>Drawing from his experience getting countless teams unstuck, Hamel explains why the solution requires a &quot;revenge of the data scientists.&quot; He details the essential mindset shifts, error analysis techniques, and practical steps needed to move beyond guesswork and build AI products you can actually trust.</p>

<p>We talk through:</p>

<ul>
<li>  The 10(+1) critical mistakes that cause teams to waste time on evals</li>
<li>  Why &quot;hallucination scores&quot; are a waste of time (and what to measure instead)</li>
<li>  The manual review process that finds major issues in hours, not weeks</li>
<li>  A step-by-step method for building LLM judges you can actually trust</li>
<li>  How to use domain experts without getting stuck in endless review committees</li>
<li>  Guest Bryan Bischof&#39;s &quot;Failure as a Funnel&quot; for debugging complex AI agents</li>
</ul>

<p>If you&#39;re tired of ambiguous &quot;vibe checks&quot; and want a clear process that delivers real improvement, this episode provides the definitive roadmap.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://hamel.dev/" rel="nofollow">Hamel&#39;s website and blog</a></li>
<li><a href="https://vanishinggradients.fireside.fm/51" rel="nofollow">Hugo speaks with Philip Carter (Honeycomb) about aligning your LLM-as-a-judge with your domain expertise</a></li>
<li><a href="https://www.lennysnewsletter.com/p/why-ai-evals-are-the-hottest-new-skill" rel="nofollow">Hamel Husain on Lenny&#39;s pocast, which includes a live demo of error analysis</a></li>
<li><a href="https://vanishinggradients.fireside.fm/9" rel="nofollow">The episode of VG in which Hamel and Hugo talk about Hamel&#39;s &quot;data consulting in Vegas&quot; era</a></li>
<li><a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">Upcoming Events on Luma</a></li>
<li><a href="https://youtube.com/live/QEk-XwrkqhI?feature=share" rel="nofollow">Watch the podcast video on YouTube</a></li>
<li><a href="https://maven.com/parlance-labs/evals?promoCode=GOHUGORGOHOME" rel="nofollow">Hamel&#39;s AI evals course, which he teaches with Shreya Shankar (UC Berkeley): starts Oct 6 and this link gives 35% off!</a> <a href="https://maven.com/parlance-labs/evals?promoCode=GOHUGORGOHOME" rel="nofollow">https://maven.com/parlance-labs/evals?promoCode=GOHUGORGOHOME</a></li>
</ul>

<p>🎓 Learn more:</p>

<ul>
<li><strong>Hugo&#39;s course:</strong> <a href="https://maven.com/s/course/d56067f338" rel="nofollow">Building LLM Applications for Data Scientists and Software Engineers</a> — <a href="https://maven.com/s/course/d56067f338" rel="nofollow">https://maven.com/s/course/d56067f338</a> </li>
</ul>]]>
  </itunes:summary>
</item>
<item>
  <title>Episode 56: DeepMind Just Dropped Gemma 270M... And Here’s Why It Matters</title>
  <link>https://vanishinggradients.fireside.fm/56</link>
  <guid isPermaLink="false">4f0a10fa-b411-458f-91b4-b68784e2d557</guid>
  <pubDate>Fri, 15 Aug 2025 02:00:00 +1000</pubDate>
  <author>Hugo Bowne-Anderson</author>
  <enclosure url="https://aphid.fireside.fm/d/1437767933/140c3904-8258-4c39-a698-a112b7077bd7/4f0a10fa-b411-458f-91b4-b68784e2d557.mp3" length="88573733" type="audio/mpeg"/>
  <itunes:episodeType>full</itunes:episodeType>
  <itunes:author>Hugo Bowne-Anderson</itunes:author>
  <itunes:subtitle>While much of the AI world chases ever-larger models, Ravin Kumar (Google DeepMind) and his team build across the size spectrum, from billions of parameters down to this week’s release: **Gemma 270M**, the smallest member yet of the Gemma 3 open-weight family. At just 270 million parameters, a quarter the size of Gemma 1B, it’s designed for speed, efficiency, and fine-tuning.  

We explore what makes 270M special, where it fits alongside its billion-parameter siblings, and why you might reach for it in production even if you think “small” means “just for experiments.”  </itunes:subtitle>
  <itunes:duration>45:40</itunes:duration>
  <itunes:explicit>no</itunes:explicit>
  <itunes:image href="https://media24.fireside.fm/file/fireside-images-2024/podcasts/images/1/140c3904-8258-4c39-a698-a112b7077bd7/cover.jpg?v=1"/>
  <description>While much of the AI world chases ever-larger models, Ravin Kumar (Google DeepMind) and his team build across the size spectrum, from billions of parameters down to this week’s release: Gemma 270M, the smallest member yet of the Gemma 3 open-weight family. At just 270 million parameters, a quarter the size of Gemma 1B,  it’s designed for speed, efficiency, and fine-tuning.  
We explore what makes 270M special, where it fits alongside its billion-parameter siblings, and why you might reach for it in production even if you think “small” means “just for experiments.”  
We talk through:  
- Where 270M fits into the Gemma 3 lineup — and why it exists  
- On-device use cases where latency, privacy, and efficiency matter  
- How smaller models open up rapid, targeted fine-tuning  
- Running multiple models in parallel without heavyweight hardware  
- Why “small” models might drive the next big wave of AI adoption  
If you’ve ever wondered what you’d do with a model this size (or how to squeeze the most out of it) this episode will show you how small can punch far above its weight.
LINKS
Introducing Gemma 3 270M: The compact model for hyper-efficient AI (Google Developer Blog) (https://developers.googleblog.com/en/introducing-gemma-3-270m/)
Full Model Fine-Tune Guide using Hugging Face Transformers (https://ai.google.dev/gemma/docs/core/huggingface_text_full_finetune)
The Gemma 270M model on HuggingFace (https://huggingface.co/google/gemma-3-270m)
The Gemma 270M model on Ollama (https://ollama.com/library/gemma3:270m)
Building AI Agents with Gemma 3, a workshop with Ravin and Hugo (https://www.youtube.com/live/-IWstEStqok) (Code here (https://github.com/canyon289/ai_agent_basics))
From Images to Agents: Building and Evaluating Multimodal AI Workflows, a workshop with Ravin and Hugo (https://www.youtube.com/live/FNlM7lSt8Uk)(Code here (https://github.com/canyon289/ai_image_agent))
Evaluating AI Agents: From Demos to Dependability, an upcoming workshop with Ravin and Hugo (https://lu.ma/ezgny3dl)
Upcoming Events on Luma (https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk)
Watch the podcast video on YouTube (https://youtu.be/VZDw6C2A_8E)
🎓 Learn more:
Hugo's course: Building LLM Applications for Data Scientists and Software Engineers (https://maven.com/s/course/d56067f338) — https://maven.com/s/course/d56067f338 ($600 off early bird discount for November cohort availiable until August 16) 
</description>
  <itunes:keywords>LLMs, GenAI, machine learning</itunes:keywords>
  <content:encoded>
    <![CDATA[<p>While much of the AI world chases ever-larger models, Ravin Kumar (Google DeepMind) and his team build across the size spectrum, from billions of parameters down to this week’s release: <strong>Gemma 270M</strong>, the smallest member yet of the Gemma 3 open-weight family. At just 270 million parameters, a quarter the size of Gemma 1B,  it’s designed for speed, efficiency, and fine-tuning.  </p>

<p>We explore what makes 270M special, where it fits alongside its billion-parameter siblings, and why you might reach for it in production even if you think “small” means “just for experiments.”  </p>

<p><strong>We talk through:</strong>  </p>

<ul>
<li>Where 270M fits into the Gemma 3 lineup — and why it exists<br></li>
<li>On-device use cases where latency, privacy, and efficiency matter<br></li>
<li>How smaller models open up rapid, targeted fine-tuning<br></li>
<li>Running multiple models in parallel without heavyweight hardware<br></li>
<li>Why “small” models might drive the next big wave of AI adoption<br></li>
</ul>

<p>If you’ve ever wondered what you’d do with a model this size (or how to squeeze the most out of it) this episode will show you how small can punch far above its weight.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://developers.googleblog.com/en/introducing-gemma-3-270m/" rel="nofollow">Introducing Gemma 3 270M: The compact model for hyper-efficient AI (Google Developer Blog)</a></li>
<li><a href="https://ai.google.dev/gemma/docs/core/huggingface_text_full_finetune" rel="nofollow">Full Model Fine-Tune Guide using Hugging Face Transformers</a></li>
<li><a href="https://huggingface.co/google/gemma-3-270m" rel="nofollow">The Gemma 270M model on HuggingFace</a></li>
<li><a href="https://ollama.com/library/gemma3:270m" rel="nofollow">The Gemma 270M model on Ollama</a></li>
<li><a href="https://www.youtube.com/live/-IWstEStqok" rel="nofollow">Building AI Agents with Gemma 3, a workshop with Ravin and Hugo</a> (<a href="https://github.com/canyon289/ai_agent_basics" rel="nofollow">Code here</a>)</li>
<li><a href="https://www.youtube.com/live/FNlM7lSt8Uk" rel="nofollow">From Images to Agents: Building and Evaluating Multimodal AI Workflows, a workshop with Ravin and Hugo</a>(<a href="https://github.com/canyon289/ai_image_agent" rel="nofollow">Code here</a>)</li>
<li><a href="https://lu.ma/ezgny3dl" rel="nofollow">Evaluating AI Agents: From Demos to Dependability, an upcoming workshop with Ravin and Hugo</a></li>
<li><a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">Upcoming Events on Luma</a></li>
<li><a href="https://youtu.be/VZDw6C2A_8E" rel="nofollow">Watch the podcast video on YouTube</a></li>
</ul>

<p>🎓 Learn more:</p>

<ul>
<li><strong>Hugo&#39;s course:</strong> <a href="https://maven.com/s/course/d56067f338" rel="nofollow">Building LLM Applications for Data Scientists and Software Engineers</a> — <a href="https://maven.com/s/course/d56067f338" rel="nofollow">https://maven.com/s/course/d56067f338</a> ($600 off early bird discount for November cohort availiable until August 16)</li>
</ul>]]>
  </content:encoded>
  <itunes:summary>
    <![CDATA[<p>While much of the AI world chases ever-larger models, Ravin Kumar (Google DeepMind) and his team build across the size spectrum, from billions of parameters down to this week’s release: <strong>Gemma 270M</strong>, the smallest member yet of the Gemma 3 open-weight family. At just 270 million parameters, a quarter the size of Gemma 1B,  it’s designed for speed, efficiency, and fine-tuning.  </p>

<p>We explore what makes 270M special, where it fits alongside its billion-parameter siblings, and why you might reach for it in production even if you think “small” means “just for experiments.”  </p>

<p><strong>We talk through:</strong>  </p>

<ul>
<li>Where 270M fits into the Gemma 3 lineup — and why it exists<br></li>
<li>On-device use cases where latency, privacy, and efficiency matter<br></li>
<li>How smaller models open up rapid, targeted fine-tuning<br></li>
<li>Running multiple models in parallel without heavyweight hardware<br></li>
<li>Why “small” models might drive the next big wave of AI adoption<br></li>
</ul>

<p>If you’ve ever wondered what you’d do with a model this size (or how to squeeze the most out of it) this episode will show you how small can punch far above its weight.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://developers.googleblog.com/en/introducing-gemma-3-270m/" rel="nofollow">Introducing Gemma 3 270M: The compact model for hyper-efficient AI (Google Developer Blog)</a></li>
<li><a href="https://ai.google.dev/gemma/docs/core/huggingface_text_full_finetune" rel="nofollow">Full Model Fine-Tune Guide using Hugging Face Transformers</a></li>
<li><a href="https://huggingface.co/google/gemma-3-270m" rel="nofollow">The Gemma 270M model on HuggingFace</a></li>
<li><a href="https://ollama.com/library/gemma3:270m" rel="nofollow">The Gemma 270M model on Ollama</a></li>
<li><a href="https://www.youtube.com/live/-IWstEStqok" rel="nofollow">Building AI Agents with Gemma 3, a workshop with Ravin and Hugo</a> (<a href="https://github.com/canyon289/ai_agent_basics" rel="nofollow">Code here</a>)</li>
<li><a href="https://www.youtube.com/live/FNlM7lSt8Uk" rel="nofollow">From Images to Agents: Building and Evaluating Multimodal AI Workflows, a workshop with Ravin and Hugo</a>(<a href="https://github.com/canyon289/ai_image_agent" rel="nofollow">Code here</a>)</li>
<li><a href="https://lu.ma/ezgny3dl" rel="nofollow">Evaluating AI Agents: From Demos to Dependability, an upcoming workshop with Ravin and Hugo</a></li>
<li><a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">Upcoming Events on Luma</a></li>
<li><a href="https://youtu.be/VZDw6C2A_8E" rel="nofollow">Watch the podcast video on YouTube</a></li>
</ul>

<p>🎓 Learn more:</p>

<ul>
<li><strong>Hugo&#39;s course:</strong> <a href="https://maven.com/s/course/d56067f338" rel="nofollow">Building LLM Applications for Data Scientists and Software Engineers</a> — <a href="https://maven.com/s/course/d56067f338" rel="nofollow">https://maven.com/s/course/d56067f338</a> ($600 off early bird discount for November cohort availiable until August 16)</li>
</ul>]]>
  </itunes:summary>
</item>
<item>
  <title>Episode 54: Scaling AI: From Colab to Clusters — A Practitioner’s Guide to Distributed Training and Inference</title>
  <link>https://vanishinggradients.fireside.fm/54</link>
  <guid isPermaLink="false">151b5251-bd41-4528-87bf-763165b8ccc7</guid>
  <pubDate>Sat, 19 Jul 2025 02:00:00 +1000</pubDate>
  <author>Hugo Bowne-Anderson</author>
  <enclosure url="https://aphid.fireside.fm/d/1437767933/140c3904-8258-4c39-a698-a112b7077bd7/151b5251-bd41-4528-87bf-763165b8ccc7.mp3" length="59469240" type="audio/mpeg"/>
  <itunes:episodeType>full</itunes:episodeType>
  <itunes:season>1</itunes:season>
  <itunes:author>Hugo Bowne-Anderson</itunes:author>
  <itunes:subtitle>Colab is cozy. But production won’t fit on a single GPU. Zach Mueller leads Accelerate at Hugging Face and spends his days helping people go from solo scripts to scalable systems. In this episode, he joins me to demystify distributed training and inference — not just for research labs, but for any ML engineer trying to ship real software.</itunes:subtitle>
  <itunes:duration>41:17</itunes:duration>
  <itunes:explicit>no</itunes:explicit>
  <itunes:image href="https://media24.fireside.fm/file/fireside-images-2024/podcasts/images/1/140c3904-8258-4c39-a698-a112b7077bd7/cover.jpg?v=1"/>
  <description>Colab is cozy. But production won’t fit on a single GPU.
Zach Mueller leads Accelerate at Hugging Face and spends his days helping people go from solo scripts to scalable systems. In this episode, he joins me to demystify distributed training and inference — not just for research labs, but for any ML engineer trying to ship real software.
We talk through:
    • From Colab to clusters: why scaling isn’t just about training massive models, but serving agents, handling load, and speeding up iteration
    • Zero-to-two GPUs: how to get started without Kubernetes, Slurm, or a PhD in networking
    • Scaling tradeoffs: when to care about interconnects, which infra bottlenecks actually matter, and how to avoid chasing performance ghosts
    • The GPU middle class: strategies for training and serving on a shoestring, with just a few cards or modest credits
    • Local experiments, global impact: why learning distributed systems—even just a little—can set you apart as an engineer
If you’ve ever stared at a Hugging Face training script and wondered how to run it on something more than your laptop: this one’s for you.
LINKS
Zach on LinkedIn (https://www.linkedin.com/in/zachary-mueller-135257118/)
Hugo's blog post on Stop Buliding AI Agents (https://www.linkedin.com/posts/hugo-bowne-anderson-045939a5_yesterday-i-posted-about-stop-building-ai-activity-7346942036752613376-b8-t/)
Upcoming Events on Luma (https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk)
Hugo's recent newsletter about upcoming events and more! (https://hugobowne.substack.com/p/stop-building-agents)
🎓 Learn more:
Hugo's course: Building LLM Applications for Data Scientists and Software Engineers (https://maven.com/s/course/d56067f338) — https://maven.com/s/course/d56067f338
Zach's course (45% off for VG listeners!): Scratch to Scale: Large-Scale Training in the Modern World (https://maven.com/walk-with-code/scratch-to-scale?promoCode=hugo39) -- https://maven.com/walk-with-code/scratch-to-scale?promoCode=hugo39
📺 Watch the video version on YouTube: YouTube link (https://youtube.com/live/76NAtzWZ25s?feature=share) 
</description>
  <itunes:keywords>AI, LLM, compute, GenAI</itunes:keywords>
  <content:encoded>
    <![CDATA[<p><strong>Colab is cozy. But production won’t fit on a single GPU.</strong><br>
Zach Mueller leads Accelerate at Hugging Face and spends his days helping people go from solo scripts to scalable systems. In this episode, he joins me to demystify distributed training and inference — not just for research labs, but for any ML engineer trying to ship real software.</p>

<p>We talk through:<br>
    • From Colab to clusters: why scaling isn’t just about training massive models, but serving agents, handling load, and speeding up iteration<br>
    • Zero-to-two GPUs: how to get started without Kubernetes, Slurm, or a PhD in networking<br>
    • Scaling tradeoffs: when to care about interconnects, which infra bottlenecks actually matter, and how to avoid chasing performance ghosts<br>
    • The GPU middle class: strategies for training and serving on a shoestring, with just a few cards or modest credits<br>
    • Local experiments, global impact: why learning distributed systems—even just a little—can set you apart as an engineer</p>

<p>If you’ve ever stared at a Hugging Face training script and wondered how to run it on something more than your laptop: this one’s for you.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://www.linkedin.com/in/zachary-mueller-135257118/" rel="nofollow">Zach on LinkedIn</a></li>
<li><a href="https://www.linkedin.com/posts/hugo-bowne-anderson-045939a5_yesterday-i-posted-about-stop-building-ai-activity-7346942036752613376-b8-t/" rel="nofollow">Hugo&#39;s blog post on Stop Buliding AI Agents</a></li>
<li><a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">Upcoming Events on Luma</a></li>
<li><a href="https://hugobowne.substack.com/p/stop-building-agents" rel="nofollow">Hugo&#39;s recent newsletter about upcoming events and more!</a></li>
</ul>

<p>🎓 Learn more:</p>

<ul>
<li><strong>Hugo&#39;s course:</strong> <a href="https://maven.com/s/course/d56067f338" rel="nofollow">Building LLM Applications for Data Scientists and Software Engineers</a> — <a href="https://maven.com/s/course/d56067f338" rel="nofollow">https://maven.com/s/course/d56067f338</a></li>
<li><strong>Zach&#39;s course (45% off for VG listeners!):</strong> <a href="https://maven.com/walk-with-code/scratch-to-scale?promoCode=hugo39" rel="nofollow">Scratch to Scale: Large-Scale Training in the Modern World</a> -- <a href="https://maven.com/walk-with-code/scratch-to-scale?promoCode=hugo39" rel="nofollow">https://maven.com/walk-with-code/scratch-to-scale?promoCode=hugo39</a></li>
</ul>

<p>📺 <strong>Watch the video version on YouTube:</strong> <a href="https://youtube.com/live/76NAtzWZ25s?feature=share" rel="nofollow">YouTube link</a></p>]]>
  </content:encoded>
  <itunes:summary>
    <![CDATA[<p><strong>Colab is cozy. But production won’t fit on a single GPU.</strong><br>
Zach Mueller leads Accelerate at Hugging Face and spends his days helping people go from solo scripts to scalable systems. In this episode, he joins me to demystify distributed training and inference — not just for research labs, but for any ML engineer trying to ship real software.</p>

<p>We talk through:<br>
    • From Colab to clusters: why scaling isn’t just about training massive models, but serving agents, handling load, and speeding up iteration<br>
    • Zero-to-two GPUs: how to get started without Kubernetes, Slurm, or a PhD in networking<br>
    • Scaling tradeoffs: when to care about interconnects, which infra bottlenecks actually matter, and how to avoid chasing performance ghosts<br>
    • The GPU middle class: strategies for training and serving on a shoestring, with just a few cards or modest credits<br>
    • Local experiments, global impact: why learning distributed systems—even just a little—can set you apart as an engineer</p>

<p>If you’ve ever stared at a Hugging Face training script and wondered how to run it on something more than your laptop: this one’s for you.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://www.linkedin.com/in/zachary-mueller-135257118/" rel="nofollow">Zach on LinkedIn</a></li>
<li><a href="https://www.linkedin.com/posts/hugo-bowne-anderson-045939a5_yesterday-i-posted-about-stop-building-ai-activity-7346942036752613376-b8-t/" rel="nofollow">Hugo&#39;s blog post on Stop Buliding AI Agents</a></li>
<li><a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">Upcoming Events on Luma</a></li>
<li><a href="https://hugobowne.substack.com/p/stop-building-agents" rel="nofollow">Hugo&#39;s recent newsletter about upcoming events and more!</a></li>
</ul>

<p>🎓 Learn more:</p>

<ul>
<li><strong>Hugo&#39;s course:</strong> <a href="https://maven.com/s/course/d56067f338" rel="nofollow">Building LLM Applications for Data Scientists and Software Engineers</a> — <a href="https://maven.com/s/course/d56067f338" rel="nofollow">https://maven.com/s/course/d56067f338</a></li>
<li><strong>Zach&#39;s course (45% off for VG listeners!):</strong> <a href="https://maven.com/walk-with-code/scratch-to-scale?promoCode=hugo39" rel="nofollow">Scratch to Scale: Large-Scale Training in the Modern World</a> -- <a href="https://maven.com/walk-with-code/scratch-to-scale?promoCode=hugo39" rel="nofollow">https://maven.com/walk-with-code/scratch-to-scale?promoCode=hugo39</a></li>
</ul>

<p>📺 <strong>Watch the video version on YouTube:</strong> <a href="https://youtube.com/live/76NAtzWZ25s?feature=share" rel="nofollow">YouTube link</a></p>]]>
  </itunes:summary>
</item>
<item>
  <title>Episode 52: Why Most LLM Products Break at Retrieval (And How to Fix Them) </title>
  <link>https://vanishinggradients.fireside.fm/52</link>
  <guid isPermaLink="false">258dd611-e817-4971-a655-f07343b967e4</guid>
  <pubDate>Thu, 03 Jul 2025 02:00:00 +1000</pubDate>
  <author>Hugo Bowne-Anderson</author>
  <enclosure url="https://aphid.fireside.fm/d/1437767933/140c3904-8258-4c39-a698-a112b7077bd7/258dd611-e817-4971-a655-f07343b967e4.mp3" length="27489267" type="audio/mpeg"/>
  <itunes:episodeType>full</itunes:episodeType>
  <itunes:season>1</itunes:season>
  <itunes:author>Hugo Bowne-Anderson</itunes:author>
  <itunes:subtitle>Most LLM-powered features do not break at the model. They break at the context. So how do you retrieve the right information to get useful results, even under vague or messy user queries?

In this episode, we hear from Eric Ma, who leads data science research in the Data Science and AI group at Moderna. He shares what it takes to move beyond toy demos and ship LLM features that actually help people do their jobs.</itunes:subtitle>
  <itunes:duration>28:38</itunes:duration>
  <itunes:explicit>no</itunes:explicit>
  <itunes:image href="https://media24.fireside.fm/file/fireside-images-2024/podcasts/images/1/140c3904-8258-4c39-a698-a112b7077bd7/cover.jpg?v=1"/>
  <description>Most LLM-powered features do not break at the model. They break at the context. So how do you retrieve the right information to get useful results, even under vague or messy user queries?
In this episode, we hear from Eric Ma, who leads data science research in the Data Science and AI group at Moderna. He shares what it takes to move beyond toy demos and ship LLM features that actually help people do their jobs.
We cover:
• How to align retrieval with user intent and why cosine similarity is not the answer
• How a dumb YAML-based system outperformed so-called smart retrieval pipelines
• Why vague queries like “what is this all about” expose real weaknesses in most systems
• When vibe checks are enough and when formal evaluation is worth the effort
• How retrieval workflows can evolve alongside your product and user needs
If you are building LLM-powered systems and care about how they work, not just whether they work, this one is for you.
LINKS
Eric's website (https://ericmjl.github.io/)
Upcoming Events on Luma (https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk)
Hugo's recent newsletter about upcoming events and more! (https://hugobowne.substack.com/p/stop-building-agents)
🎓 Learn more:
Hugo's course: Building LLM Applications for Data Scientists and Software Engineers (https://maven.com/s/course/d56067f338) — next cohort starts July 8: https://maven.com/s/course/d56067f338
📺 Watch the video version on YouTube: YouTube link (https://youtu.be/d-FaR5Ywd5k)
</description>
  <itunes:keywords>data science, machine learning, AI, LLMs, GenAI</itunes:keywords>
  <content:encoded>
    <![CDATA[<p>Most LLM-powered features do not break at the model. They break at the context. So how do you retrieve the right information to get useful results, even under vague or messy user queries?</p>

<p>In this episode, we hear from Eric Ma, who leads data science research in the Data Science and AI group at Moderna. He shares what it takes to move beyond toy demos and ship LLM features that actually help people do their jobs.</p>

<p>We cover:<br>
• How to align retrieval with user intent and why cosine similarity is not the answer<br>
• How a dumb YAML-based system outperformed so-called smart retrieval pipelines<br>
• Why vague queries like “what is this all about” expose real weaknesses in most systems<br>
• When vibe checks are enough and when formal evaluation is worth the effort<br>
• How retrieval workflows can evolve alongside your product and user needs</p>

<p>If you are building LLM-powered systems and care about how they work, not just whether they work, this one is for you.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://ericmjl.github.io/" rel="nofollow">Eric&#39;s website</a></li>
<li><a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">Upcoming Events on Luma</a></li>
<li><a href="https://hugobowne.substack.com/p/stop-building-agents" rel="nofollow">Hugo&#39;s recent newsletter about upcoming events and more!</a></li>
</ul>

<p>🎓 Learn more:</p>

<ul>
<li><strong>Hugo&#39;s course:</strong> <a href="https://maven.com/s/course/d56067f338" rel="nofollow">Building LLM Applications for Data Scientists and Software Engineers</a> — next cohort starts July 8: <a href="https://maven.com/s/course/d56067f338" rel="nofollow">https://maven.com/s/course/d56067f338</a></li>
</ul>

<p>📺 <strong>Watch the video version on YouTube:</strong> <a href="https://youtu.be/d-FaR5Ywd5k" rel="nofollow">YouTube link</a></p>]]>
  </content:encoded>
  <itunes:summary>
    <![CDATA[<p>Most LLM-powered features do not break at the model. They break at the context. So how do you retrieve the right information to get useful results, even under vague or messy user queries?</p>

<p>In this episode, we hear from Eric Ma, who leads data science research in the Data Science and AI group at Moderna. He shares what it takes to move beyond toy demos and ship LLM features that actually help people do their jobs.</p>

<p>We cover:<br>
• How to align retrieval with user intent and why cosine similarity is not the answer<br>
• How a dumb YAML-based system outperformed so-called smart retrieval pipelines<br>
• Why vague queries like “what is this all about” expose real weaknesses in most systems<br>
• When vibe checks are enough and when formal evaluation is worth the effort<br>
• How retrieval workflows can evolve alongside your product and user needs</p>

<p>If you are building LLM-powered systems and care about how they work, not just whether they work, this one is for you.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://ericmjl.github.io/" rel="nofollow">Eric&#39;s website</a></li>
<li><a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">Upcoming Events on Luma</a></li>
<li><a href="https://hugobowne.substack.com/p/stop-building-agents" rel="nofollow">Hugo&#39;s recent newsletter about upcoming events and more!</a></li>
</ul>

<p>🎓 Learn more:</p>

<ul>
<li><strong>Hugo&#39;s course:</strong> <a href="https://maven.com/s/course/d56067f338" rel="nofollow">Building LLM Applications for Data Scientists and Software Engineers</a> — next cohort starts July 8: <a href="https://maven.com/s/course/d56067f338" rel="nofollow">https://maven.com/s/course/d56067f338</a></li>
</ul>

<p>📺 <strong>Watch the video version on YouTube:</strong> <a href="https://youtu.be/d-FaR5Ywd5k" rel="nofollow">YouTube link</a></p>]]>
  </itunes:summary>
</item>
<item>
  <title>Episode 47: The Great Pacific Garbage Patch of Code Slop with Joe Reis</title>
  <link>https://vanishinggradients.fireside.fm/47</link>
  <guid isPermaLink="false">decc9c1a-f18a-41e9-947a-e58fa0957f1e</guid>
  <pubDate>Mon, 07 Apr 2025 10:00:00 +1000</pubDate>
  <author>Hugo Bowne-Anderson</author>
  <enclosure url="https://aphid.fireside.fm/d/1437767933/140c3904-8258-4c39-a698-a112b7077bd7/decc9c1a-f18a-41e9-947a-e58fa0957f1e.mp3" length="76045085" type="audio/mpeg"/>
  <itunes:episodeType>full</itunes:episodeType>
  <itunes:season>1</itunes:season>
  <itunes:author>Hugo Bowne-Anderson</itunes:author>
  <itunes:subtitle>What if the cost of writing code dropped to zero — but the cost of understanding it skyrocketed?

In this episode, Hugo sits down with Joe Reis to unpack how AI tooling is reshaping the software development lifecycle — from experimentation and prototyping to deployment, maintainability, and everything in between.</itunes:subtitle>
  <itunes:duration>1:19:12</itunes:duration>
  <itunes:explicit>yes</itunes:explicit>
  <itunes:image href="https://media24.fireside.fm/file/fireside-images-2024/podcasts/images/1/140c3904-8258-4c39-a698-a112b7077bd7/cover.jpg?v=1"/>
  <description>What if the cost of writing code dropped to zero — but the cost of understanding it skyrocketed?
In this episode, Hugo sits down with Joe Reis to unpack how AI tooling is reshaping the software development lifecycle — from experimentation and prototyping to deployment, maintainability, and everything in between.
Joe is the co-author of Fundamentals of Data Engineering and a longtime voice on the systems side of modern software. He’s also one of the sharpest critics of “vibe coding” — the emerging pattern of writing software by feel, with heavy reliance on LLMs and little regard for structure or quality.
We dive into:
    • Why “vibe coding” is more than a meme — and what it says about how we build today
    • How AI tools expand the surface area of software creation — for better and worse
    • What happens to technical debt, testing, and security when generation outpaces understanding
    • The changing definition of “production” in a world of ephemeral, internal, or just-good-enough tools
    • How AI is flattening the learning curve — and threatening the talent pipeline
    • Joe’s view on what real craftsmanship means in an age of disposable code
This conversation isn’t about doom, and it’s not about hype. It’s about mapping the real, messy terrain of what it means to build software today — and how to do it with care.
LINKS
* Joe's Practical Data Modeling Newsletter on Substack (https://practicaldatamodeling.substack.com/)
* Joe's Practical Data Modeling Server on Discord (https://discord.gg/HhSZVvWDBb)
* Vanishing Gradients YouTube Channel (https://www.youtube.com/channel/UC_NafIo-Ku2loOLrzm45ABA)  
* Upcoming Events on Luma (https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk)
🎓 Want to go deeper?
Check out my course: Building LLM Applications for Data Scientists and Software Engineers.
Learn how to design, test, and deploy production-grade LLM systems — with observability, feedback loops, and structure built in.
This isn’t about vibes or fragile agents. It’s about making LLMs reliable, testable, and actually useful.
Includes over $800 in compute credits and guest lectures from experts at DeepMind, Moderna, and more.
Cohort starts July 8 — Use this link for a 10% discount (https://maven.com/hugo-stefan/building-llm-apps-ds-and-swe-from-first-principles?promoCode=LLM10) 
</description>
  <itunes:keywords>AI, LLMs, data science, machine learning, data science, GenAI, vibe coding</itunes:keywords>
  <content:encoded>
    <![CDATA[<p>What if the cost of writing code dropped to zero — but the cost of understanding it skyrocketed?</p>

<p>In this episode, Hugo sits down with Joe Reis to unpack how AI tooling is reshaping the software development lifecycle — from experimentation and prototyping to deployment, maintainability, and everything in between.</p>

<p>Joe is the co-author of Fundamentals of Data Engineering and a longtime voice on the systems side of modern software. He’s also one of the sharpest critics of “vibe coding” — the emerging pattern of writing software by feel, with heavy reliance on LLMs and little regard for structure or quality.</p>

<p>We dive into:<br>
    • Why “vibe coding” is more than a meme — and what it says about how we build today<br>
    • How AI tools expand the surface area of software creation — for better and worse<br>
    • What happens to technical debt, testing, and security when generation outpaces understanding<br>
    • The changing definition of “production” in a world of ephemeral, internal, or just-good-enough tools<br>
    • How AI is flattening the learning curve — and threatening the talent pipeline<br>
    • Joe’s view on what real craftsmanship means in an age of disposable code</p>

<p>This conversation isn’t about doom, and it’s not about hype. It’s about mapping the real, messy terrain of what it means to build software today — and how to do it with care.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://practicaldatamodeling.substack.com/" rel="nofollow">Joe&#39;s Practical Data Modeling Newsletter on Substack</a></li>
<li><a href="https://discord.gg/HhSZVvWDBb" rel="nofollow">Joe&#39;s Practical Data Modeling Server on Discord</a></li>
<li><a href="https://www.youtube.com/channel/UC_NafIo-Ku2loOLrzm45ABA" rel="nofollow">Vanishing Gradients YouTube Channel</a><br></li>
<li><a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">Upcoming Events on Luma</a></li>
</ul>

<p>🎓 Want to go deeper?<br>
Check out my course: <em>Building LLM Applications for Data Scientists and Software Engineers.</em><br>
Learn how to design, test, and deploy production-grade LLM systems — with observability, feedback loops, and structure built in.<br>
This isn’t about vibes or fragile agents. It’s about making LLMs reliable, testable, and actually useful.</p>

<p>Includes over $800 in compute credits and guest lectures from experts at DeepMind, Moderna, and more.<br>
Cohort starts July 8 — <a href="https://maven.com/hugo-stefan/building-llm-apps-ds-and-swe-from-first-principles?promoCode=LLM10" rel="nofollow">Use this link for a 10% discount</a></p>]]>
  </content:encoded>
  <itunes:summary>
    <![CDATA[<p>What if the cost of writing code dropped to zero — but the cost of understanding it skyrocketed?</p>

<p>In this episode, Hugo sits down with Joe Reis to unpack how AI tooling is reshaping the software development lifecycle — from experimentation and prototyping to deployment, maintainability, and everything in between.</p>

<p>Joe is the co-author of Fundamentals of Data Engineering and a longtime voice on the systems side of modern software. He’s also one of the sharpest critics of “vibe coding” — the emerging pattern of writing software by feel, with heavy reliance on LLMs and little regard for structure or quality.</p>

<p>We dive into:<br>
    • Why “vibe coding” is more than a meme — and what it says about how we build today<br>
    • How AI tools expand the surface area of software creation — for better and worse<br>
    • What happens to technical debt, testing, and security when generation outpaces understanding<br>
    • The changing definition of “production” in a world of ephemeral, internal, or just-good-enough tools<br>
    • How AI is flattening the learning curve — and threatening the talent pipeline<br>
    • Joe’s view on what real craftsmanship means in an age of disposable code</p>

<p>This conversation isn’t about doom, and it’s not about hype. It’s about mapping the real, messy terrain of what it means to build software today — and how to do it with care.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://practicaldatamodeling.substack.com/" rel="nofollow">Joe&#39;s Practical Data Modeling Newsletter on Substack</a></li>
<li><a href="https://discord.gg/HhSZVvWDBb" rel="nofollow">Joe&#39;s Practical Data Modeling Server on Discord</a></li>
<li><a href="https://www.youtube.com/channel/UC_NafIo-Ku2loOLrzm45ABA" rel="nofollow">Vanishing Gradients YouTube Channel</a><br></li>
<li><a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">Upcoming Events on Luma</a></li>
</ul>

<p>🎓 Want to go deeper?<br>
Check out my course: <em>Building LLM Applications for Data Scientists and Software Engineers.</em><br>
Learn how to design, test, and deploy production-grade LLM systems — with observability, feedback loops, and structure built in.<br>
This isn’t about vibes or fragile agents. It’s about making LLMs reliable, testable, and actually useful.</p>

<p>Includes over $800 in compute credits and guest lectures from experts at DeepMind, Moderna, and more.<br>
Cohort starts July 8 — <a href="https://maven.com/hugo-stefan/building-llm-apps-ds-and-swe-from-first-principles?promoCode=LLM10" rel="nofollow">Use this link for a 10% discount</a></p>]]>
  </itunes:summary>
</item>
<item>
  <title>Episode 46: Software Composition Is the New Vibe Coding</title>
  <link>https://vanishinggradients.fireside.fm/46</link>
  <guid isPermaLink="false">dcb8396f-ece2-4636-951c-8ad44d698d15</guid>
  <pubDate>Thu, 03 Apr 2025 13:00:00 +1100</pubDate>
  <author>Hugo Bowne-Anderson</author>
  <enclosure url="https://aphid.fireside.fm/d/1437767933/140c3904-8258-4c39-a698-a112b7077bd7/dcb8396f-ece2-4636-951c-8ad44d698d15.mp3" length="99299288" type="audio/mpeg"/>
  <itunes:episodeType>full</itunes:episodeType>
  <itunes:season>1</itunes:season>
  <itunes:author>Hugo Bowne-Anderson</itunes:author>
  <itunes:subtitle>What if building software felt more like composing than coding?

In this episode, Hugo and Greg explore how LLMs are reshaping the way we think about software development—from deterministic programming to a more flexible, prompt-driven, and collaborative style of building. It’s not just hype or grift—it’s a real shift in how we express intent, reason about systems, and collaborate across roles.

Hugo speaks with Greg Ceccarelli—co-founder of SpecStory, former CPO at Pluralsight, and Director of Data Science at GitHub—about the rise of software composition and how it changes the way individuals and teams create with LLMs.</itunes:subtitle>
  <itunes:duration>1:08:57</itunes:duration>
  <itunes:explicit>no</itunes:explicit>
  <itunes:image href="https://media24.fireside.fm/file/fireside-images-2024/podcasts/images/1/140c3904-8258-4c39-a698-a112b7077bd7/cover.jpg?v=1"/>
  <description>What if building software felt more like composing than coding?
In this episode, Hugo and Greg explore how LLMs are reshaping the way we think about software development—from deterministic programming to a more flexible, prompt-driven, and collaborative style of building. It’s not just hype or grift—it’s a real shift in how we express intent, reason about systems, and collaborate across roles.
Hugo speaks with Greg Ceccarelli—co-founder of SpecStory, former CPO at Pluralsight, and Director of Data Science at GitHub—about the rise of software composition and how it changes the way individuals and teams create with LLMs.
We dive into:
- Why software composition is emerging as a serious alternative to traditional coding
- The real difference between vibe coding and production-minded prototyping
- How LLMs are expanding who gets to build software—and how
- What changes when you focus on intent, not just code
- What Greg is building with SpecStory to support collaborative, traceable AI-native workflows
- The challenges (and joys) of debugging and exploring with agentic tools like Cursor and Claude
We’ve removed the visual demos from the audio—but you can catch our live-coded Chrome extension and JFK document explorer on YouTube. Links below.
JFK Docs Vibe Coding Demo (YouTube) (https://youtu.be/JpXCkuV58QE)  
Chrome Extension Vibe Coding Demo (YouTube) (https://youtu.be/ESVKp37jDwc)  
Meditations on Tech (Greg’s Substack) (https://www.meditationsontech.com/)  
Simon Willison on Vibe Coding (https://simonwillison.net/2025/Mar/19/vibe-coding/)  
Johnno Whitaker: On Vibe Coding (https://johnowhitaker.dev/essays/vibe_coding.html)  
Tim O’Reilly – The End of Programming (https://www.oreilly.com/radar/the-end-of-programming-as-we-know-it/)  
Vanishing Gradients YouTube Channel (https://www.youtube.com/channel/UC_NafIo-Ku2loOLrzm45ABA)  
Upcoming Events on Luma (https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk)  
Greg Ceccarelli on LinkedIn (https://www.linkedin.com/in/gregceccarelli/)  
Greg’s Hacker News Post on GOOD (https://news.ycombinator.com/item?id=43557698)  
SpecStory: GOOD – Git Companion for AI Workflows (https://github.com/specstoryai/getspecstory/blob/main/GOOD.md)
🎓 Want to go deeper?
Check out my course: Building LLM Applications for Data Scientists and Software Engineers.
Learn how to design, test, and deploy production-grade LLM systems — with observability, feedback loops, and structure built in.
This isn’t about vibes or fragile agents. It’s about making LLMs reliable, testable, and actually useful.
Includes over $2,500 in compute credits and guest lectures from experts at DeepMind, Moderna, and more.
Cohort starts April 7 — Use this link for a 10% discount (https://maven.com/hugo-stefan/building-llm-apps-ds-and-swe-from-first-principles?promoCode=LLM10)
🔍 Want to help shape the future of SpecStory?
Greg and the team are looking for design partners for their new SpecStory Teams product—built for collaborative, AI-native software development.
If you're working with LLMs in a team setting and want to influence the next wave of developer tools, you can apply here:  
👉 specstory.com/teams (https://specstory.com/teams) 
</description>
  <itunes:keywords>AI, LLMs, data science, machine learning, data science, GenAI, vibe coding</itunes:keywords>
  <content:encoded>
    <![CDATA[<p>What if building software felt more like composing than coding?</p>

<p>In this episode, Hugo and Greg explore how LLMs are reshaping the way we think about software development—from deterministic programming to a more flexible, prompt-driven, and collaborative style of building. It’s not just hype or grift—it’s a real shift in how we express intent, reason about systems, and collaborate across roles.</p>

<p>Hugo speaks with Greg Ceccarelli—co-founder of SpecStory, former CPO at Pluralsight, and Director of Data Science at GitHub—about the rise of software composition and how it changes the way individuals and teams create with LLMs.</p>

<p>We dive into:</p>

<ul>
<li>Why software composition is emerging as a serious alternative to traditional coding</li>
<li>The real difference between vibe coding and production-minded prototyping</li>
<li>How LLMs are expanding who gets to build software—and how</li>
<li>What changes when you focus on intent, not just code</li>
<li>What Greg is building with SpecStory to support collaborative, traceable AI-native workflows</li>
<li>The challenges (and joys) of debugging and exploring with agentic tools like Cursor and Claude</li>
</ul>

<p>We’ve removed the visual demos from the audio—but you can catch our live-coded Chrome extension and JFK document explorer on YouTube. Links below.</p>

<ul>
<li><a href="https://youtu.be/JpXCkuV58QE" rel="nofollow">JFK Docs Vibe Coding Demo (YouTube)</a><br></li>
<li><a href="https://youtu.be/ESVKp37jDwc" rel="nofollow">Chrome Extension Vibe Coding Demo (YouTube)</a><br></li>
<li><a href="https://www.meditationsontech.com/" rel="nofollow">Meditations on Tech (Greg’s Substack)</a><br></li>
<li><a href="https://simonwillison.net/2025/Mar/19/vibe-coding/" rel="nofollow">Simon Willison on Vibe Coding</a><br></li>
<li><a href="https://johnowhitaker.dev/essays/vibe_coding.html" rel="nofollow">Johnno Whitaker: On Vibe Coding</a><br></li>
<li><a href="https://www.oreilly.com/radar/the-end-of-programming-as-we-know-it/" rel="nofollow">Tim O’Reilly – The End of Programming</a><br></li>
<li><a href="https://www.youtube.com/channel/UC_NafIo-Ku2loOLrzm45ABA" rel="nofollow">Vanishing Gradients YouTube Channel</a><br></li>
<li><a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">Upcoming Events on Luma</a><br></li>
<li><a href="https://www.linkedin.com/in/gregceccarelli/" rel="nofollow">Greg Ceccarelli on LinkedIn</a><br></li>
<li><a href="https://news.ycombinator.com/item?id=43557698" rel="nofollow">Greg’s Hacker News Post on GOOD</a><br></li>
<li><a href="https://github.com/specstoryai/getspecstory/blob/main/GOOD.md" rel="nofollow">SpecStory: GOOD – Git Companion for AI Workflows</a></li>
</ul>

<p>🎓 Want to go deeper?<br>
Check out my course: <em>Building LLM Applications for Data Scientists and Software Engineers.</em><br>
Learn how to design, test, and deploy production-grade LLM systems — with observability, feedback loops, and structure built in.<br>
This isn’t about vibes or fragile agents. It’s about making LLMs reliable, testable, and actually useful.</p>

<p>Includes over $2,500 in compute credits and guest lectures from experts at DeepMind, Moderna, and more.<br>
Cohort starts April 7 — <a href="https://maven.com/hugo-stefan/building-llm-apps-ds-and-swe-from-first-principles?promoCode=LLM10" rel="nofollow">Use this link for a 10% discount</a></p>

<h3>🔍 Want to help shape the future of SpecStory?</h3>

<p>Greg and the team are looking for <strong>design partners</strong> for their new SpecStory Teams product—built for collaborative, AI-native software development.</p>

<p>If you&#39;re working with LLMs in a team setting and want to influence the next wave of developer tools, you can apply here:<br><br>
👉 <a href="https://specstory.com/teams" rel="nofollow">specstory.com/teams</a></p>]]>
  </content:encoded>
  <itunes:summary>
    <![CDATA[<p>What if building software felt more like composing than coding?</p>

<p>In this episode, Hugo and Greg explore how LLMs are reshaping the way we think about software development—from deterministic programming to a more flexible, prompt-driven, and collaborative style of building. It’s not just hype or grift—it’s a real shift in how we express intent, reason about systems, and collaborate across roles.</p>

<p>Hugo speaks with Greg Ceccarelli—co-founder of SpecStory, former CPO at Pluralsight, and Director of Data Science at GitHub—about the rise of software composition and how it changes the way individuals and teams create with LLMs.</p>

<p>We dive into:</p>

<ul>
<li>Why software composition is emerging as a serious alternative to traditional coding</li>
<li>The real difference between vibe coding and production-minded prototyping</li>
<li>How LLMs are expanding who gets to build software—and how</li>
<li>What changes when you focus on intent, not just code</li>
<li>What Greg is building with SpecStory to support collaborative, traceable AI-native workflows</li>
<li>The challenges (and joys) of debugging and exploring with agentic tools like Cursor and Claude</li>
</ul>

<p>We’ve removed the visual demos from the audio—but you can catch our live-coded Chrome extension and JFK document explorer on YouTube. Links below.</p>

<ul>
<li><a href="https://youtu.be/JpXCkuV58QE" rel="nofollow">JFK Docs Vibe Coding Demo (YouTube)</a><br></li>
<li><a href="https://youtu.be/ESVKp37jDwc" rel="nofollow">Chrome Extension Vibe Coding Demo (YouTube)</a><br></li>
<li><a href="https://www.meditationsontech.com/" rel="nofollow">Meditations on Tech (Greg’s Substack)</a><br></li>
<li><a href="https://simonwillison.net/2025/Mar/19/vibe-coding/" rel="nofollow">Simon Willison on Vibe Coding</a><br></li>
<li><a href="https://johnowhitaker.dev/essays/vibe_coding.html" rel="nofollow">Johnno Whitaker: On Vibe Coding</a><br></li>
<li><a href="https://www.oreilly.com/radar/the-end-of-programming-as-we-know-it/" rel="nofollow">Tim O’Reilly – The End of Programming</a><br></li>
<li><a href="https://www.youtube.com/channel/UC_NafIo-Ku2loOLrzm45ABA" rel="nofollow">Vanishing Gradients YouTube Channel</a><br></li>
<li><a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">Upcoming Events on Luma</a><br></li>
<li><a href="https://www.linkedin.com/in/gregceccarelli/" rel="nofollow">Greg Ceccarelli on LinkedIn</a><br></li>
<li><a href="https://news.ycombinator.com/item?id=43557698" rel="nofollow">Greg’s Hacker News Post on GOOD</a><br></li>
<li><a href="https://github.com/specstoryai/getspecstory/blob/main/GOOD.md" rel="nofollow">SpecStory: GOOD – Git Companion for AI Workflows</a></li>
</ul>

<p>🎓 Want to go deeper?<br>
Check out my course: <em>Building LLM Applications for Data Scientists and Software Engineers.</em><br>
Learn how to design, test, and deploy production-grade LLM systems — with observability, feedback loops, and structure built in.<br>
This isn’t about vibes or fragile agents. It’s about making LLMs reliable, testable, and actually useful.</p>

<p>Includes over $2,500 in compute credits and guest lectures from experts at DeepMind, Moderna, and more.<br>
Cohort starts April 7 — <a href="https://maven.com/hugo-stefan/building-llm-apps-ds-and-swe-from-first-principles?promoCode=LLM10" rel="nofollow">Use this link for a 10% discount</a></p>

<h3>🔍 Want to help shape the future of SpecStory?</h3>

<p>Greg and the team are looking for <strong>design partners</strong> for their new SpecStory Teams product—built for collaborative, AI-native software development.</p>

<p>If you&#39;re working with LLMs in a team setting and want to influence the next wave of developer tools, you can apply here:<br><br>
👉 <a href="https://specstory.com/teams" rel="nofollow">specstory.com/teams</a></p>]]>
  </itunes:summary>
</item>
<item>
  <title>Episode 42: Learning, Teaching, and Building in the Age of AI</title>
  <link>https://vanishinggradients.fireside.fm/42</link>
  <guid isPermaLink="false">6af2e172-b72b-418b-baa6-369299f37b8b</guid>
  <pubDate>Sat, 04 Jan 2025 14:00:00 +1100</pubDate>
  <author>Hugo Bowne-Anderson</author>
  <enclosure url="https://aphid.fireside.fm/d/1437767933/140c3904-8258-4c39-a698-a112b7077bd7/6af2e172-b72b-418b-baa6-369299f37b8b.mp3" length="76860106" type="audio/mpeg"/>
  <itunes:episodeType>full</itunes:episodeType>
  <itunes:season>1</itunes:season>
  <itunes:author>Hugo Bowne-Anderson</itunes:author>
  <itunes:subtitle>The tables turn as Hugo sits down with Alex Andorra, host of Learning Bayesian Statistics. Hugo shares his journey from mathematics to AI, reflecting on how Bayesian inference shapes his approach to data science, teaching, and building AI-powered applications.</itunes:subtitle>
  <itunes:duration>1:20:03</itunes:duration>
  <itunes:explicit>no</itunes:explicit>
  <itunes:image href="https://media24.fireside.fm/file/fireside-images-2024/podcasts/images/1/140c3904-8258-4c39-a698-a112b7077bd7/cover.jpg?v=1"/>
  <description>In this episode of Vanishing Gradients, the tables turn as Hugo sits down with Alex Andorra, host of Learning Bayesian Statistics. Hugo shares his journey from mathematics to AI, reflecting on how Bayesian inference shapes his approach to data science, teaching, and building AI-powered applications.
They dive into the realities of deploying LLM applications, overcoming “proof-of-concept purgatory,” and why first principles and iteration are critical for success in AI. Whether you’re an educator, software engineer, or data scientist, this episode offers valuable insights into the intersection of AI, product development, and real-world deployment.
LINKS
The podcast on YouTube (https://www.youtube.com/watch?v=BRIYytbqtP0)
The original podcast episode (https://learnbayesstats.com/episode/122-learning-and-teaching-in-the-age-of-ai-hugo-bowne-anderson)
Alex Andorra on LinkedIn (https://www.linkedin.com/in/alex-andorra/)
Hugo on LinkedIn (https://www.linkedin.com/in/hugo-bowne-anderson-045939a5/)
Hugo on twitter (https://x.com/hugobowne)
Vanishing Gradients on twitter (https://x.com/vanishingdata)
Hugo's "Building LLM Applications for Data Scientists and Software Engineers" course (https://maven.com/s/course/d56067f338) 
</description>
  <itunes:keywords>data science, machine learning, AI, LLMs</itunes:keywords>
  <content:encoded>
    <![CDATA[<p>In this episode of Vanishing Gradients, the tables turn as Hugo sits down with Alex Andorra, host of Learning Bayesian Statistics. Hugo shares his journey from mathematics to AI, reflecting on how Bayesian inference shapes his approach to data science, teaching, and building AI-powered applications.</p>

<p>They dive into the realities of deploying LLM applications, overcoming “proof-of-concept purgatory,” and why first principles and iteration are critical for success in AI. Whether you’re an educator, software engineer, or data scientist, this episode offers valuable insights into the intersection of AI, product development, and real-world deployment.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://www.youtube.com/watch?v=BRIYytbqtP0" rel="nofollow">The podcast on YouTube</a></li>
<li><a href="https://learnbayesstats.com/episode/122-learning-and-teaching-in-the-age-of-ai-hugo-bowne-anderson" rel="nofollow">The original podcast episode</a></li>
<li><a href="https://www.linkedin.com/in/alex-andorra/" rel="nofollow">Alex Andorra on LinkedIn</a></li>
<li><a href="https://www.linkedin.com/in/hugo-bowne-anderson-045939a5/" rel="nofollow">Hugo on LinkedIn</a></li>
<li><a href="https://x.com/hugobowne" rel="nofollow">Hugo on twitter</a></li>
<li><a href="https://x.com/vanishingdata" rel="nofollow">Vanishing Gradients on twitter</a></li>
<li><a href="https://maven.com/s/course/d56067f338" rel="nofollow">Hugo&#39;s &quot;Building LLM Applications for Data Scientists and Software Engineers&quot; course</a></li>
</ul>]]>
  </content:encoded>
  <itunes:summary>
    <![CDATA[<p>In this episode of Vanishing Gradients, the tables turn as Hugo sits down with Alex Andorra, host of Learning Bayesian Statistics. Hugo shares his journey from mathematics to AI, reflecting on how Bayesian inference shapes his approach to data science, teaching, and building AI-powered applications.</p>

<p>They dive into the realities of deploying LLM applications, overcoming “proof-of-concept purgatory,” and why first principles and iteration are critical for success in AI. Whether you’re an educator, software engineer, or data scientist, this episode offers valuable insights into the intersection of AI, product development, and real-world deployment.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://www.youtube.com/watch?v=BRIYytbqtP0" rel="nofollow">The podcast on YouTube</a></li>
<li><a href="https://learnbayesstats.com/episode/122-learning-and-teaching-in-the-age-of-ai-hugo-bowne-anderson" rel="nofollow">The original podcast episode</a></li>
<li><a href="https://www.linkedin.com/in/alex-andorra/" rel="nofollow">Alex Andorra on LinkedIn</a></li>
<li><a href="https://www.linkedin.com/in/hugo-bowne-anderson-045939a5/" rel="nofollow">Hugo on LinkedIn</a></li>
<li><a href="https://x.com/hugobowne" rel="nofollow">Hugo on twitter</a></li>
<li><a href="https://x.com/vanishingdata" rel="nofollow">Vanishing Gradients on twitter</a></li>
<li><a href="https://maven.com/s/course/d56067f338" rel="nofollow">Hugo&#39;s &quot;Building LLM Applications for Data Scientists and Software Engineers&quot; course</a></li>
</ul>]]>
  </itunes:summary>
</item>
<item>
  <title>Episode 41: Beyond Prompt Engineering: Can AI Learn to Set Its Own Goals?</title>
  <link>https://vanishinggradients.fireside.fm/41</link>
  <guid isPermaLink="false">695d8cc9-b111-4f1d-9871-82962ae023f4</guid>
  <pubDate>Tue, 31 Dec 2024 10:00:00 +1100</pubDate>
  <author>Hugo Bowne-Anderson</author>
  <enclosure url="https://aphid.fireside.fm/d/1437767933/140c3904-8258-4c39-a698-a112b7077bd7/695d8cc9-b111-4f1d-9871-82962ae023f4.mp3" length="42114740" type="audio/mpeg"/>
  <itunes:episodeType>full</itunes:episodeType>
  <itunes:season>1</itunes:season>
  <itunes:author>Hugo Bowne-Anderson</itunes:author>
  <itunes:subtitle>Hugo Bowne-Anderson hosts a panel discussion from the MLOps World and Generative AI Summit in Austin, exploring the long-term growth of AI by distinguishing real problem-solving from trend-based solutions. If you're navigating the evolving landscape of generative AI, productionizing models, or questioning the hype, this episode dives into the tough questions shaping the field.</itunes:subtitle>
  <itunes:duration>43:51</itunes:duration>
  <itunes:explicit>yes</itunes:explicit>
  <itunes:image href="https://media24.fireside.fm/file/fireside-images-2024/podcasts/images/1/140c3904-8258-4c39-a698-a112b7077bd7/cover.jpg?v=1"/>
  <description>Hugo Bowne-Anderson hosts a panel discussion from the MLOps World and Generative AI Summit in Austin, exploring the long-term growth of AI by distinguishing real problem-solving from trend-based solutions. If you're navigating the evolving landscape of generative AI, productionizing models, or questioning the hype, this episode dives into the tough questions shaping the field.
The panel features:  
- Ben Taylor (Jepson) (https://www.linkedin.com/in/jepsontaylor/) – CEO and Founder at VEOX Inc., with experience in AI exploration, genetic programming, and deep learning.  
- Joe Reis (https://www.linkedin.com/in/josephreis/) – Co-founder of Ternary Data and author of Fundamentals of Data Engineering.  
- Juan Sequeda (https://www.linkedin.com/in/juansequeda/) – Principal Scientist and Head of AI Lab at Data.World, known for his expertise in knowledge graphs and the semantic web.  
The discussion unpacks essential topics such as:  
- The shift from prompt engineering to goal engineering—letting AI iterate toward well-defined objectives.  
- Whether generative AI is having an electricity moment or more of a blockchain trajectory.  
- The combinatorial power of AI to explore new solutions, drawing parallels to AlphaZero redefining strategy games.  
- The POC-to-production gap and why AI projects stall.  
- Failure modes, hallucinations, and governance risks—and how to mitigate them.  
- The disconnect between executive optimism and employee workload.  
Hugo also mentions his upcoming workshop on escaping Proof-of-Concept Purgatory, which has evolved into a Maven course "Building LLM Applications for Data Scientists and Software Engineers" launching in January (https://maven.com/hugo-stefan/building-llm-apps-ds-and-swe-from-first-principles?utm_campaign=8123d0&amp;amp;utm_medium=partner&amp;amp;utm_source=instructor). Vanishing Gradient listeners can get 25% off the course (use the code VG25), with $1,000 in Modal compute credits included.
A huge thanks to Dave Scharbach and the Toronto Machine Learning Society for organizing the conference and to the audience for their thoughtful questions.
As we head into the new year, this conversation offers a reality check amidst the growing AI agent hype.  
LINKS
Hugo on twitter (https://x.com/hugobowne)
Hugo on LinkedIn (https://www.linkedin.com/in/hugo-bowne-anderson-045939a5/)
Vanishing Gradients on twitter (https://x.com/vanishingdata)
"Building LLM Applications for Data Scientists and Software Engineers" course (https://maven.com/hugo-stefan/building-llm-apps-ds-and-swe-from-first-principles?utm_campaign=8123d0&amp;amp;utm_medium=partner&amp;amp;utm_source=instructor).
</description>
  <itunes:keywords>data science, machine learning, AI, LLMs</itunes:keywords>
  <content:encoded>
    <![CDATA[<p>Hugo Bowne-Anderson hosts a panel discussion from the MLOps World and Generative AI Summit in Austin, exploring the long-term growth of AI by distinguishing real problem-solving from trend-based solutions. If you&#39;re navigating the evolving landscape of generative AI, productionizing models, or questioning the hype, this episode dives into the tough questions shaping the field.</p>

<p>The panel features:  </p>

<ul>
<li><a href="https://www.linkedin.com/in/jepsontaylor/" rel="nofollow"><strong>Ben Taylor (Jepson)</strong></a> – CEO and Founder at VEOX Inc., with experience in AI exploration, genetic programming, and deep learning.<br></li>
<li><a href="https://www.linkedin.com/in/josephreis/" rel="nofollow"><strong>Joe Reis</strong></a> – Co-founder of Ternary Data and author of <em>Fundamentals of Data Engineering</em>.<br></li>
<li><a href="https://www.linkedin.com/in/juansequeda/" rel="nofollow"><strong>Juan Sequeda</strong></a> – Principal Scientist and Head of AI Lab at Data.World, known for his expertise in knowledge graphs and the semantic web.<br></li>
</ul>

<p>The discussion unpacks essential topics such as:  </p>

<ul>
<li>The shift from <strong>prompt engineering</strong> to <strong>goal engineering</strong>—letting AI iterate toward well-defined objectives.<br></li>
<li>Whether generative AI is having an <strong>electricity moment</strong> or more of a <strong>blockchain trajectory</strong>.<br></li>
<li>The <strong>combinatorial power of AI</strong> to explore new solutions, drawing parallels to AlphaZero redefining strategy games.<br></li>
<li>The <strong>POC-to-production gap</strong> and why AI projects stall.<br></li>
<li><strong>Failure modes, hallucinations, and governance risks</strong>—and how to mitigate them.<br></li>
<li>The disconnect between executive optimism and employee workload.<br></li>
</ul>

<p>Hugo also mentions his upcoming workshop on <strong>escaping Proof-of-Concept Purgatory</strong>, <a href="https://maven.com/hugo-stefan/building-llm-apps-ds-and-swe-from-first-principles?utm_campaign=8123d0&utm_medium=partner&utm_source=instructor" rel="nofollow">which has evolved into a <strong>Maven course &quot;Building LLM Applications for Data Scientists and Software Engineers&quot; launching in January</strong></a>. Vanishing Gradient listeners can get 25% off the course (use the code VG25), with $1,000 in Modal compute credits included.</p>

<p>A huge thanks to <strong>Dave Scharbach and the Toronto Machine Learning Society</strong> for organizing the conference and to the audience for their thoughtful questions.</p>

<p>As we head into the new year, this conversation offers a reality check amidst the growing AI agent hype.  </p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://x.com/hugobowne" rel="nofollow">Hugo on twitter</a></li>
<li><a href="https://www.linkedin.com/in/hugo-bowne-anderson-045939a5/" rel="nofollow">Hugo on LinkedIn</a></li>
<li><a href="https://x.com/vanishingdata" rel="nofollow">Vanishing Gradients on twitter</a></li>
<li><a href="https://maven.com/hugo-stefan/building-llm-apps-ds-and-swe-from-first-principles?utm_campaign=8123d0&utm_medium=partner&utm_source=instructor" rel="nofollow">&quot;Building LLM Applications for Data Scientists and Software Engineers&quot; course</a>.</li>
</ul>]]>
  </content:encoded>
  <itunes:summary>
    <![CDATA[<p>Hugo Bowne-Anderson hosts a panel discussion from the MLOps World and Generative AI Summit in Austin, exploring the long-term growth of AI by distinguishing real problem-solving from trend-based solutions. If you&#39;re navigating the evolving landscape of generative AI, productionizing models, or questioning the hype, this episode dives into the tough questions shaping the field.</p>

<p>The panel features:  </p>

<ul>
<li><a href="https://www.linkedin.com/in/jepsontaylor/" rel="nofollow"><strong>Ben Taylor (Jepson)</strong></a> – CEO and Founder at VEOX Inc., with experience in AI exploration, genetic programming, and deep learning.<br></li>
<li><a href="https://www.linkedin.com/in/josephreis/" rel="nofollow"><strong>Joe Reis</strong></a> – Co-founder of Ternary Data and author of <em>Fundamentals of Data Engineering</em>.<br></li>
<li><a href="https://www.linkedin.com/in/juansequeda/" rel="nofollow"><strong>Juan Sequeda</strong></a> – Principal Scientist and Head of AI Lab at Data.World, known for his expertise in knowledge graphs and the semantic web.<br></li>
</ul>

<p>The discussion unpacks essential topics such as:  </p>

<ul>
<li>The shift from <strong>prompt engineering</strong> to <strong>goal engineering</strong>—letting AI iterate toward well-defined objectives.<br></li>
<li>Whether generative AI is having an <strong>electricity moment</strong> or more of a <strong>blockchain trajectory</strong>.<br></li>
<li>The <strong>combinatorial power of AI</strong> to explore new solutions, drawing parallels to AlphaZero redefining strategy games.<br></li>
<li>The <strong>POC-to-production gap</strong> and why AI projects stall.<br></li>
<li><strong>Failure modes, hallucinations, and governance risks</strong>—and how to mitigate them.<br></li>
<li>The disconnect between executive optimism and employee workload.<br></li>
</ul>

<p>Hugo also mentions his upcoming workshop on <strong>escaping Proof-of-Concept Purgatory</strong>, <a href="https://maven.com/hugo-stefan/building-llm-apps-ds-and-swe-from-first-principles?utm_campaign=8123d0&utm_medium=partner&utm_source=instructor" rel="nofollow">which has evolved into a <strong>Maven course &quot;Building LLM Applications for Data Scientists and Software Engineers&quot; launching in January</strong></a>. Vanishing Gradient listeners can get 25% off the course (use the code VG25), with $1,000 in Modal compute credits included.</p>

<p>A huge thanks to <strong>Dave Scharbach and the Toronto Machine Learning Society</strong> for organizing the conference and to the audience for their thoughtful questions.</p>

<p>As we head into the new year, this conversation offers a reality check amidst the growing AI agent hype.  </p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://x.com/hugobowne" rel="nofollow">Hugo on twitter</a></li>
<li><a href="https://www.linkedin.com/in/hugo-bowne-anderson-045939a5/" rel="nofollow">Hugo on LinkedIn</a></li>
<li><a href="https://x.com/vanishingdata" rel="nofollow">Vanishing Gradients on twitter</a></li>
<li><a href="https://maven.com/hugo-stefan/building-llm-apps-ds-and-swe-from-first-principles?utm_campaign=8123d0&utm_medium=partner&utm_source=instructor" rel="nofollow">&quot;Building LLM Applications for Data Scientists and Software Engineers&quot; course</a>.</li>
</ul>]]>
  </itunes:summary>
</item>
<item>
  <title>Episode 38: The Art of Freelance AI Consulting and Products: Data, Dollars, and Deliverables</title>
  <link>https://vanishinggradients.fireside.fm/38</link>
  <guid isPermaLink="false">c1a5c8d1-777a-41b7-a123-6b06861dbc35</guid>
  <pubDate>Tue, 05 Nov 2024 10:00:00 +1100</pubDate>
  <author>Hugo Bowne-Anderson</author>
  <enclosure url="https://aphid.fireside.fm/d/1437767933/140c3904-8258-4c39-a698-a112b7077bd7/c1a5c8d1-777a-41b7-a123-6b06861dbc35.mp3" length="80443270" type="audio/mpeg"/>
  <itunes:episodeType>full</itunes:episodeType>
  <itunes:season>1</itunes:season>
  <itunes:author>Hugo Bowne-Anderson</itunes:author>
  <itunes:subtitle>Hugo speaks with Jason Liu, an independent AI consultant with experience at Meta and Stitch Fix. At Stitch Fix, Jason developed impactful AI systems, like a $50 million product similarity search and the widely adopted Flight recommendation framework. Now, he helps startups and enterprises design and deploy production-level AI applications, with a focus on retrieval-augmented generation (RAG) and scalable solutions.</itunes:subtitle>
  <itunes:duration>1:23:47</itunes:duration>
  <itunes:explicit>no</itunes:explicit>
  <itunes:image href="https://media24.fireside.fm/file/fireside-images-2024/podcasts/images/1/140c3904-8258-4c39-a698-a112b7077bd7/cover.jpg?v=1"/>
  <description>Hugo speaks with Jason Liu, an independent AI consultant with experience at Meta and Stitch Fix. At Stitch Fix, Jason developed impactful AI systems, like a $50 million product similarity search and the widely adopted Flight recommendation framework. Now, he helps startups and enterprises design and deploy production-level AI applications, with a focus on retrieval-augmented generation (RAG) and scalable solutions.
This episode is a bit of an experiment. Instead of our usual technical deep dives, we’re focusing on the world of AI consulting and freelancing. We explore Jason’s consulting playbook, covering how he structures contracts to maximize value, strategies for moving from hourly billing to securing larger deals, and the mindset shift needed to align incentives with clients. We’ll also discuss the challenges of moving from deterministic software to probabilistic AI systems and even do a live role-playing session where Jason coaches me on client engagement and pricing pitfalls.
LINKS
The livestream on YouTube (https://youtube.com/live/9CFs06UDbGI?feature=share)
Jason's Upcoming course: AI Consultant Accelerator: From Expert to High-Demand Business (https://maven.com/indie-consulting/ai-consultant-accelerator?utm_campaign=9532cc&amp;amp;utm_medium=partner&amp;amp;utm_source=instructor)
Hugo's upcoming course: Building LLM Applications for Data Scientists and Software Engineers (https://maven.com/s/course/d56067f338)
Jason's website (https://jxnl.co/)
Jason's indie consulting newsletter (https://indieconsulting.podia.com/)
Your AI Product Needs Evals by Hamel Husain (https://hamel.dev/blog/posts/evals/)
What We’ve Learned From A Year of Building with LLMs (https://applied-llms.org/)
Dear Future AI Consultant by Jason (https://jxnl.co/writing/#dear-future-ai-consultant)
Alex Hormozi's books (https://www.acquisition.com/books)
The Burnout Society by Byung-Chul Han (https://www.sup.org/books/theory-and-philosophy/burnout-society)
Jason on Twitter (https://x.com/jxnlco)
Vanishing Gradients on Twitter (https://twitter.com/vanishingdata)
Hugo on Twitter (https://twitter.com/hugobowne)
Vanishing Gradients' lu.ma calendar (https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk)
Vanishing Gradients on YouTube (https://www.youtube.com/@vanishinggradients) 
</description>
  <itunes:keywords>AI, LLMs, machine learning, data science, GenAI, consulting</itunes:keywords>
  <content:encoded>
    <![CDATA[<p>Hugo speaks with Jason Liu, an independent AI consultant with experience at Meta and Stitch Fix. At Stitch Fix, Jason developed impactful AI systems, like a $50 million product similarity search and the widely adopted Flight recommendation framework. Now, he helps startups and enterprises design and deploy production-level AI applications, with a focus on retrieval-augmented generation (RAG) and scalable solutions.</p>

<p>This episode is a bit of an experiment. Instead of our usual technical deep dives, we’re focusing on the world of AI consulting and freelancing. We explore Jason’s consulting playbook, covering how he structures contracts to maximize value, strategies for moving from hourly billing to securing larger deals, and the mindset shift needed to align incentives with clients. We’ll also discuss the challenges of moving from deterministic software to probabilistic AI systems and even do a live role-playing session where Jason coaches me on client engagement and pricing pitfalls.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://youtube.com/live/9CFs06UDbGI?feature=share" rel="nofollow">The livestream on YouTube</a></li>
<li><a href="https://maven.com/indie-consulting/ai-consultant-accelerator?utm_campaign=9532cc&utm_medium=partner&utm_source=instructor" rel="nofollow">Jason&#39;s Upcoming course: AI Consultant Accelerator: From Expert to High-Demand Business</a></li>
<li><a href="https://maven.com/s/course/d56067f338" rel="nofollow">Hugo&#39;s upcoming course: Building LLM Applications for Data Scientists and Software Engineers</a></li>
<li><a href="https://jxnl.co/" rel="nofollow">Jason&#39;s website</a></li>
<li><a href="https://indieconsulting.podia.com/" rel="nofollow">Jason&#39;s indie consulting newsletter</a></li>
<li><a href="https://hamel.dev/blog/posts/evals/" rel="nofollow">Your AI Product Needs Evals by Hamel Husain</a></li>
<li><a href="https://applied-llms.org/" rel="nofollow">What We’ve Learned From A Year of Building with LLMs</a></li>
<li><a href="https://jxnl.co/writing/#dear-future-ai-consultant" rel="nofollow">Dear Future AI Consultant by Jason</a></li>
<li><a href="https://www.acquisition.com/books" rel="nofollow">Alex Hormozi&#39;s books</a></li>
<li><a href="https://www.sup.org/books/theory-and-philosophy/burnout-society" rel="nofollow">The Burnout Society by Byung-Chul Han</a></li>
<li><a href="https://x.com/jxnlco" rel="nofollow">Jason on Twitter</a></li>
<li><a href="https://twitter.com/vanishingdata" rel="nofollow">Vanishing Gradients on Twitter</a></li>
<li><a href="https://twitter.com/hugobowne" rel="nofollow">Hugo on Twitter</a></li>
<li><a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">Vanishing Gradients&#39; lu.ma calendar</a></li>
<li><a href="https://www.youtube.com/@vanishinggradients" rel="nofollow">Vanishing Gradients on YouTube</a></li>
</ul>]]>
  </content:encoded>
  <itunes:summary>
    <![CDATA[<p>Hugo speaks with Jason Liu, an independent AI consultant with experience at Meta and Stitch Fix. At Stitch Fix, Jason developed impactful AI systems, like a $50 million product similarity search and the widely adopted Flight recommendation framework. Now, he helps startups and enterprises design and deploy production-level AI applications, with a focus on retrieval-augmented generation (RAG) and scalable solutions.</p>

<p>This episode is a bit of an experiment. Instead of our usual technical deep dives, we’re focusing on the world of AI consulting and freelancing. We explore Jason’s consulting playbook, covering how he structures contracts to maximize value, strategies for moving from hourly billing to securing larger deals, and the mindset shift needed to align incentives with clients. We’ll also discuss the challenges of moving from deterministic software to probabilistic AI systems and even do a live role-playing session where Jason coaches me on client engagement and pricing pitfalls.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://youtube.com/live/9CFs06UDbGI?feature=share" rel="nofollow">The livestream on YouTube</a></li>
<li><a href="https://maven.com/indie-consulting/ai-consultant-accelerator?utm_campaign=9532cc&utm_medium=partner&utm_source=instructor" rel="nofollow">Jason&#39;s Upcoming course: AI Consultant Accelerator: From Expert to High-Demand Business</a></li>
<li><a href="https://maven.com/s/course/d56067f338" rel="nofollow">Hugo&#39;s upcoming course: Building LLM Applications for Data Scientists and Software Engineers</a></li>
<li><a href="https://jxnl.co/" rel="nofollow">Jason&#39;s website</a></li>
<li><a href="https://indieconsulting.podia.com/" rel="nofollow">Jason&#39;s indie consulting newsletter</a></li>
<li><a href="https://hamel.dev/blog/posts/evals/" rel="nofollow">Your AI Product Needs Evals by Hamel Husain</a></li>
<li><a href="https://applied-llms.org/" rel="nofollow">What We’ve Learned From A Year of Building with LLMs</a></li>
<li><a href="https://jxnl.co/writing/#dear-future-ai-consultant" rel="nofollow">Dear Future AI Consultant by Jason</a></li>
<li><a href="https://www.acquisition.com/books" rel="nofollow">Alex Hormozi&#39;s books</a></li>
<li><a href="https://www.sup.org/books/theory-and-philosophy/burnout-society" rel="nofollow">The Burnout Society by Byung-Chul Han</a></li>
<li><a href="https://x.com/jxnlco" rel="nofollow">Jason on Twitter</a></li>
<li><a href="https://twitter.com/vanishingdata" rel="nofollow">Vanishing Gradients on Twitter</a></li>
<li><a href="https://twitter.com/hugobowne" rel="nofollow">Hugo on Twitter</a></li>
<li><a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">Vanishing Gradients&#39; lu.ma calendar</a></li>
<li><a href="https://www.youtube.com/@vanishinggradients" rel="nofollow">Vanishing Gradients on YouTube</a></li>
</ul>]]>
  </itunes:summary>
</item>
<item>
  <title>Episode 37: Prompt Engineering, Security in Generative AI, and the Future of AI Research Part 2</title>
  <link>https://vanishinggradients.fireside.fm/37</link>
  <guid isPermaLink="false">eadec2c4-f8f9-45b0-ae7e-5867f7201801</guid>
  <pubDate>Tue, 08 Oct 2024 17:00:00 +1100</pubDate>
  <author>Hugo Bowne-Anderson</author>
  <enclosure url="https://aphid.fireside.fm/d/1437767933/140c3904-8258-4c39-a698-a112b7077bd7/eadec2c4-f8f9-45b0-ae7e-5867f7201801.mp3" length="48585166" type="audio/mpeg"/>
  <itunes:episodeType>full</itunes:episodeType>
  <itunes:season>1</itunes:season>
  <itunes:author>Hugo Bowne-Anderson</itunes:author>
  <itunes:subtitle>Hugo speaks with three leading figures from the world of AI research: Sander Schulhoff, a recent University of Maryland graduate and lead contributor to the Learn Prompting initiative; Philip Resnik, professor at the University of Maryland, known for his pioneering work in computational linguistics; and Dennis Peskoff, a researcher from Princeton specializing in prompt engineering and its applications in the social sciences.</itunes:subtitle>
  <itunes:duration>50:36</itunes:duration>
  <itunes:explicit>no</itunes:explicit>
  <itunes:image href="https://media24.fireside.fm/file/fireside-images-2024/podcasts/images/1/140c3904-8258-4c39-a698-a112b7077bd7/cover.jpg?v=1"/>
  <description>Hugo speaks with three leading figures from the world of AI research: Sander Schulhoff, a recent University of Maryland graduate and lead contributor to the Learn Prompting initiative; Philip Resnik, professor at the University of Maryland, known for his pioneering work in computational linguistics; and Dennis Peskoff, a researcher from Princeton specializing in prompt engineering and its applications in the social sciences.
This is Part 2 of a special two-part episode, prompted—no pun intended—by these guys being part of a team, led by Sander, that wrote a 76-page survey analyzing prompting techniques, agents, and generative AI. The survey included contributors from OpenAI, Microsoft, the University of Maryland, Princeton, and more.
In this episode, we cover:
The Prompt Report: A comprehensive survey on prompting techniques, agents, and generative AI, including advanced evaluation methods for assessing these techniques.
Security Risks and Prompt Hacking: A detailed exploration of the security concerns surrounding prompt engineering, including Sander’s thoughts on its potential applications in cybersecurity and military contexts.
AI’s Impact Across Fields: A discussion on how generative AI is reshaping various domains, including the social sciences and security.
Multimodal AI: Updates on how large language models (LLMs) are expanding to interact with images, code, and music.
Case Study - Detecting Suicide Risk: A careful examination of how prompting techniques are being used in important areas like detecting suicide risk, showcasing the critical potential of AI in addressing sensitive, real-world challenges.
The episode concludes with a reflection on the evolving landscape of LLMs and multimodal AI, and what might be on the horizon.
If you haven’t yet, make sure to check out Part 1, where we discuss the history of NLP, prompt engineering techniques, and Sander’s development of the Learn Prompting initiative.
LINKS
The livestream on YouTube (https://youtube.com/live/FreXovgG-9A?feature=share)
The Prompt Report: A Systematic Survey of Prompting Techniques (https://arxiv.org/abs/2406.06608)
Learn Prompting: Your Guide to Communicating with AI (https://learnprompting.org/)
Vanishing Gradients on Twitter (https://twitter.com/vanishingdata)
Hugo on Twitter (https://twitter.com/hugobowne)
Vanishing Gradients' lu.ma calendar (https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk)
Vanishing Gradients on YouTube (https://www.youtube.com/@vanishinggradients)
</description>
  <itunes:keywords>AI, LLMs, machine learning, data science, GenAI, NLP</itunes:keywords>
  <content:encoded>
    <![CDATA[<p>Hugo speaks with three leading figures from the world of AI research: Sander Schulhoff, a recent University of Maryland graduate and lead contributor to the Learn Prompting initiative; Philip Resnik, professor at the University of Maryland, known for his pioneering work in computational linguistics; and Dennis Peskoff, a researcher from Princeton specializing in prompt engineering and its applications in the social sciences.</p>

<p>This is Part 2 of a special two-part episode, prompted—no pun intended—by these guys being part of a team, led by Sander, that wrote a 76-page survey analyzing prompting techniques, agents, and generative AI. The survey included contributors from OpenAI, Microsoft, the University of Maryland, Princeton, and more.</p>

<p>In this episode, we cover:</p>

<ul>
<li><p><strong>The Prompt Report:</strong> A comprehensive survey on prompting techniques, agents, and generative AI, including advanced evaluation methods for assessing these techniques.</p></li>
<li><p><strong>Security Risks and Prompt Hacking:</strong> A detailed exploration of the security concerns surrounding prompt engineering, including Sander’s thoughts on its potential applications in cybersecurity and military contexts.</p></li>
<li><p><strong>AI’s Impact Across Fields:</strong> A discussion on how generative AI is reshaping various domains, including the social sciences and security.</p></li>
<li><p><strong>Multimodal AI:</strong> Updates on how large language models (LLMs) are expanding to interact with images, code, and music.</p></li>
<li><p><strong>Case Study - Detecting Suicide Risk:</strong> A careful examination of how prompting techniques are being used in important areas like detecting suicide risk, showcasing the critical potential of AI in addressing sensitive, real-world challenges.</p></li>
</ul>

<p>The episode concludes with a reflection on the evolving landscape of <strong>LLMs</strong> and multimodal AI, and what might be on the horizon.</p>

<p>If you haven’t yet, make sure to check out <strong>Part 1</strong>, where we discuss the history of NLP, prompt engineering techniques, and Sander’s development of the Learn Prompting initiative.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://youtube.com/live/FreXovgG-9A?feature=share" rel="nofollow">The livestream on YouTube</a></li>
<li><a href="https://arxiv.org/abs/2406.06608" rel="nofollow">The Prompt Report: A Systematic Survey of Prompting Techniques</a></li>
<li><a href="https://learnprompting.org/" rel="nofollow">Learn Prompting: Your Guide to Communicating with AI</a></li>
<li><a href="https://twitter.com/vanishingdata" rel="nofollow">Vanishing Gradients on Twitter</a></li>
<li><a href="https://twitter.com/hugobowne" rel="nofollow">Hugo on Twitter</a></li>
<li><a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">Vanishing Gradients&#39; lu.ma calendar</a></li>
<li><a href="https://www.youtube.com/@vanishinggradients" rel="nofollow">Vanishing Gradients on YouTube</a></li>
</ul>]]>
  </content:encoded>
  <itunes:summary>
    <![CDATA[<p>Hugo speaks with three leading figures from the world of AI research: Sander Schulhoff, a recent University of Maryland graduate and lead contributor to the Learn Prompting initiative; Philip Resnik, professor at the University of Maryland, known for his pioneering work in computational linguistics; and Dennis Peskoff, a researcher from Princeton specializing in prompt engineering and its applications in the social sciences.</p>

<p>This is Part 2 of a special two-part episode, prompted—no pun intended—by these guys being part of a team, led by Sander, that wrote a 76-page survey analyzing prompting techniques, agents, and generative AI. The survey included contributors from OpenAI, Microsoft, the University of Maryland, Princeton, and more.</p>

<p>In this episode, we cover:</p>

<ul>
<li><p><strong>The Prompt Report:</strong> A comprehensive survey on prompting techniques, agents, and generative AI, including advanced evaluation methods for assessing these techniques.</p></li>
<li><p><strong>Security Risks and Prompt Hacking:</strong> A detailed exploration of the security concerns surrounding prompt engineering, including Sander’s thoughts on its potential applications in cybersecurity and military contexts.</p></li>
<li><p><strong>AI’s Impact Across Fields:</strong> A discussion on how generative AI is reshaping various domains, including the social sciences and security.</p></li>
<li><p><strong>Multimodal AI:</strong> Updates on how large language models (LLMs) are expanding to interact with images, code, and music.</p></li>
<li><p><strong>Case Study - Detecting Suicide Risk:</strong> A careful examination of how prompting techniques are being used in important areas like detecting suicide risk, showcasing the critical potential of AI in addressing sensitive, real-world challenges.</p></li>
</ul>

<p>The episode concludes with a reflection on the evolving landscape of <strong>LLMs</strong> and multimodal AI, and what might be on the horizon.</p>

<p>If you haven’t yet, make sure to check out <strong>Part 1</strong>, where we discuss the history of NLP, prompt engineering techniques, and Sander’s development of the Learn Prompting initiative.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://youtube.com/live/FreXovgG-9A?feature=share" rel="nofollow">The livestream on YouTube</a></li>
<li><a href="https://arxiv.org/abs/2406.06608" rel="nofollow">The Prompt Report: A Systematic Survey of Prompting Techniques</a></li>
<li><a href="https://learnprompting.org/" rel="nofollow">Learn Prompting: Your Guide to Communicating with AI</a></li>
<li><a href="https://twitter.com/vanishingdata" rel="nofollow">Vanishing Gradients on Twitter</a></li>
<li><a href="https://twitter.com/hugobowne" rel="nofollow">Hugo on Twitter</a></li>
<li><a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">Vanishing Gradients&#39; lu.ma calendar</a></li>
<li><a href="https://www.youtube.com/@vanishinggradients" rel="nofollow">Vanishing Gradients on YouTube</a></li>
</ul>]]>
  </itunes:summary>
</item>
<item>
  <title>Episode 36: Prompt Engineering, Security in Generative AI, and the Future of AI Research Part 1</title>
  <link>https://vanishinggradients.fireside.fm/36</link>
  <guid isPermaLink="false">acd8aaec-1788-459d-a4e9-10feae67a19a</guid>
  <pubDate>Mon, 30 Sep 2024 18:00:00 +1000</pubDate>
  <author>Hugo Bowne-Anderson</author>
  <enclosure url="https://aphid.fireside.fm/d/1437767933/140c3904-8258-4c39-a698-a112b7077bd7/acd8aaec-1788-459d-a4e9-10feae67a19a.mp3" length="61232193" type="audio/mpeg"/>
  <itunes:episodeType>full</itunes:episodeType>
  <itunes:season>1</itunes:season>
  <itunes:author>Hugo Bowne-Anderson</itunes:author>
  <itunes:subtitle>Hugo speaks with three leading figures from the world of AI research: Sander Schulhoff, a recent University of Maryland graduate and lead contributor to the Learn Prompting initiative; Philip Resnik, professor at the University of Maryland, known for his pioneering work in computational linguistics; and Dennis Peskoff, a researcher from Princeton specializing in prompt engineering and its applications in the social sciences.</itunes:subtitle>
  <itunes:duration>1:03:46</itunes:duration>
  <itunes:explicit>no</itunes:explicit>
  <itunes:image href="https://media24.fireside.fm/file/fireside-images-2024/podcasts/images/1/140c3904-8258-4c39-a698-a112b7077bd7/cover.jpg?v=1"/>
  <description>Hugo speaks with three leading figures from the world of AI research: Sander Schulhoff, a recent University of Maryland graduate and lead contributor to the Learn Prompting initiative; Philip Resnik, professor at the University of Maryland, known for his pioneering work in computational linguistics; and Dennis Peskoff, a researcher from Princeton specializing in prompt engineering and its applications in the social sciences.
This is Part 1 of a special two-part episode, prompted—no pun intended—by these guys being part of a team, led by Sander, that wrote a 76-page survey analyzing prompting techniques, agents, and generative AI. The survey included contributors from OpenAI, Microsoft, the University of Maryland, Princeton, and more.
In this first part, 
* we’ll explore the critical role of prompt engineering, 
* &amp;amp; diving into adversarial techniques like prompt hacking and 
* the challenges of evaluating these techniques. 
* we’ll examine the impact of few-shot learning and 
* the groundbreaking taxonomy of prompting techniques from the Prompt Report.
Along the way, 
* we’ll uncover the rich history of natural language processing (NLP) and AI, showing how modern prompting techniques evolved from early rule-based systems and statistical methods. 
* we’ll also hear how Sander’s experimentation with GPT-3 for diplomatic tasks led him to develop Learn Prompting, and 
* how Dennis highlights the accessibility of AI through prompting, which allows non-technical users to interact with AI without needing to code.
Finally, we’ll explore the future of multimodal AI, where LLMs interact with images, code, and even music creation. Make sure to tune in to Part 2, where we dive deeper into security risks, prompt hacking, and more.
LINKS
The livestream on YouTube (https://youtube.com/live/FreXovgG-9A?feature=share)
The Prompt Report: A Systematic Survey of Prompting Techniques (https://arxiv.org/abs/2406.06608)
Learn Prompting: Your Guide to Communicating with AI (https://learnprompting.org/)
Vanishing Gradients on Twitter (https://twitter.com/vanishingdata)
Hugo on Twitter (https://twitter.com/hugobowne)
Vanishing Gradients' lu.ma calendar (https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk)
Vanishing Gradients on YouTube (https://www.youtube.com/@vanishinggradients)
</description>
  <itunes:keywords>AI, LLMs, damachine learning, data science, GenAI, prompt engineering</itunes:keywords>
  <content:encoded>
    <![CDATA[<p>Hugo speaks with three leading figures from the world of AI research: Sander Schulhoff, a recent University of Maryland graduate and lead contributor to the Learn Prompting initiative; Philip Resnik, professor at the University of Maryland, known for his pioneering work in computational linguistics; and Dennis Peskoff, a researcher from Princeton specializing in prompt engineering and its applications in the social sciences.</p>

<p>This is Part 1 of a special two-part episode, prompted—no pun intended—by these guys being part of a team, led by Sander, that wrote a 76-page survey analyzing prompting techniques, agents, and generative AI. The survey included contributors from OpenAI, Microsoft, the University of Maryland, Princeton, and more.</p>

<p>In this first part, </p>

<ul>
<li>we’ll explore the critical role of prompt engineering, </li>
<li>&amp; diving into adversarial techniques like prompt hacking and </li>
<li>the challenges of evaluating these techniques. </li>
<li>we’ll examine the impact of few-shot learning and </li>
<li>the groundbreaking taxonomy of prompting techniques from the Prompt Report.</li>
</ul>

<p>Along the way, </p>

<ul>
<li>we’ll uncover the rich history of natural language processing (NLP) and AI, showing how modern prompting techniques evolved from early rule-based systems and statistical methods. </li>
<li>we’ll also hear how Sander’s experimentation with GPT-3 for diplomatic tasks led him to develop Learn Prompting, and </li>
<li>how Dennis highlights the accessibility of AI through prompting, which allows non-technical users to interact with AI without needing to code.</li>
</ul>

<p>Finally, we’ll explore the future of multimodal AI, where LLMs interact with images, code, and even music creation. Make sure to tune in to Part 2, where we dive deeper into security risks, prompt hacking, and more.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://youtube.com/live/FreXovgG-9A?feature=share" rel="nofollow">The livestream on YouTube</a></li>
<li><a href="https://arxiv.org/abs/2406.06608" rel="nofollow">The Prompt Report: A Systematic Survey of Prompting Techniques</a></li>
<li><a href="https://learnprompting.org/" rel="nofollow">Learn Prompting: Your Guide to Communicating with AI</a></li>
<li><a href="https://twitter.com/vanishingdata" rel="nofollow">Vanishing Gradients on Twitter</a></li>
<li><a href="https://twitter.com/hugobowne" rel="nofollow">Hugo on Twitter</a></li>
<li><a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">Vanishing Gradients&#39; lu.ma calendar</a></li>
<li><a href="https://www.youtube.com/@vanishinggradients" rel="nofollow">Vanishing Gradients on YouTube</a></li>
</ul>]]>
  </content:encoded>
  <itunes:summary>
    <![CDATA[<p>Hugo speaks with three leading figures from the world of AI research: Sander Schulhoff, a recent University of Maryland graduate and lead contributor to the Learn Prompting initiative; Philip Resnik, professor at the University of Maryland, known for his pioneering work in computational linguistics; and Dennis Peskoff, a researcher from Princeton specializing in prompt engineering and its applications in the social sciences.</p>

<p>This is Part 1 of a special two-part episode, prompted—no pun intended—by these guys being part of a team, led by Sander, that wrote a 76-page survey analyzing prompting techniques, agents, and generative AI. The survey included contributors from OpenAI, Microsoft, the University of Maryland, Princeton, and more.</p>

<p>In this first part, </p>

<ul>
<li>we’ll explore the critical role of prompt engineering, </li>
<li>&amp; diving into adversarial techniques like prompt hacking and </li>
<li>the challenges of evaluating these techniques. </li>
<li>we’ll examine the impact of few-shot learning and </li>
<li>the groundbreaking taxonomy of prompting techniques from the Prompt Report.</li>
</ul>

<p>Along the way, </p>

<ul>
<li>we’ll uncover the rich history of natural language processing (NLP) and AI, showing how modern prompting techniques evolved from early rule-based systems and statistical methods. </li>
<li>we’ll also hear how Sander’s experimentation with GPT-3 for diplomatic tasks led him to develop Learn Prompting, and </li>
<li>how Dennis highlights the accessibility of AI through prompting, which allows non-technical users to interact with AI without needing to code.</li>
</ul>

<p>Finally, we’ll explore the future of multimodal AI, where LLMs interact with images, code, and even music creation. Make sure to tune in to Part 2, where we dive deeper into security risks, prompt hacking, and more.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://youtube.com/live/FreXovgG-9A?feature=share" rel="nofollow">The livestream on YouTube</a></li>
<li><a href="https://arxiv.org/abs/2406.06608" rel="nofollow">The Prompt Report: A Systematic Survey of Prompting Techniques</a></li>
<li><a href="https://learnprompting.org/" rel="nofollow">Learn Prompting: Your Guide to Communicating with AI</a></li>
<li><a href="https://twitter.com/vanishingdata" rel="nofollow">Vanishing Gradients on Twitter</a></li>
<li><a href="https://twitter.com/hugobowne" rel="nofollow">Hugo on Twitter</a></li>
<li><a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">Vanishing Gradients&#39; lu.ma calendar</a></li>
<li><a href="https://www.youtube.com/@vanishinggradients" rel="nofollow">Vanishing Gradients on YouTube</a></li>
</ul>]]>
  </itunes:summary>
</item>
<item>
  <title>Episode 35: Open Science at NASA -- Measuring Impact and the Future of AI</title>
  <link>https://vanishinggradients.fireside.fm/35</link>
  <guid isPermaLink="false">feeeecc8-a170-48c7-ae4c-8dd64484c64c</guid>
  <pubDate>Thu, 19 Sep 2024 17:00:00 +1000</pubDate>
  <author>Hugo Bowne-Anderson</author>
  <enclosure url="https://aphid.fireside.fm/d/1437767933/140c3904-8258-4c39-a698-a112b7077bd7/feeeecc8-a170-48c7-ae4c-8dd64484c64c.mp3" length="55905303" type="audio/mpeg"/>
  <itunes:episodeType>full</itunes:episodeType>
  <itunes:season>1</itunes:season>
  <itunes:author>Hugo Bowne-Anderson</itunes:author>
  <itunes:subtitle>Hugo speaks with Dr. Chelle Gentemann, Open Science Program Scientist for NASA’s Office of the Chief Science Data Officer, about NASA’s ambitious efforts to integrate AI across the research lifecycle. In this episode, we’ll dive deeper into how AI is transforming NASA’s approach to science, making data more accessible and advancing open science practices.</itunes:subtitle>
  <itunes:duration>58:13</itunes:duration>
  <itunes:explicit>no</itunes:explicit>
  <itunes:image href="https://media24.fireside.fm/file/fireside-images-2024/podcasts/images/1/140c3904-8258-4c39-a698-a112b7077bd7/cover.jpg?v=1"/>
  <description>Hugo speaks with Dr. Chelle Gentemann, Open Science Program Scientist for NASA’s Office of the Chief Science Data Officer, about NASA’s ambitious efforts to integrate AI across the research lifecycle. In this episode, we’ll dive deeper into how AI is transforming NASA’s approach to science, making data more accessible and advancing open science practices. We explore
Measuring the Impact of Open Science: How NASA is developing new metrics to evaluate the effectiveness of open science, moving beyond traditional publication-based assessments.
The Process of Scientific Discovery: Insights into the collaborative nature of research and how breakthroughs are achieved at NASA.
** AI Applications in NASA’s Science:** From rats in space to exploring the origins of the universe, we cover how AI is being applied across NASA’s divisions to improve data accessibility and analysis.
Addressing Challenges in Open Science: The complexities of implementing open science within government agencies and research environments.
Reforming Incentive Systems: How NASA is reconsidering traditional metrics like publications and citations, and starting to recognize contributions such as software development and data sharing.
The Future of Open Science: How open science is shaping the future of research, fostering interdisciplinary collaboration, and increasing accessibility.
This conversation offers valuable insights for researchers, data scientists, and those interested in the practical applications of AI and open science. Join us as we discuss how NASA is working to make science more collaborative, reproducible, and impactful.
LINKS
The livestream on YouTube (https://youtube.com/live/VJDg3ZbkNOE?feature=share)
NASA's Open Science 101 course &amp;lt;-- do it to learn and also to get NASA Swag! (https://openscience101.org/)
Science Cast (https://sciencecast.org/)
NASA and IBM Openly Release Geospatial AI Foundation Model for NASA Earth Observation Data (https://www.earthdata.nasa.gov/news/impact-ibm-hls-foundation-model)
Jake VanderPlas' daily conundrum tweet from 2013 (https://x.com/jakevdp/status/408678764705378304)
Replit, "an AI-powered software development &amp;amp; deployment platform for building, sharing, and shipping software fast." (https://replit.com/) 
</description>
  <itunes:keywords>AI, LLMs, damachine learning, data science, GenAI</itunes:keywords>
  <content:encoded>
    <![CDATA[<p>Hugo speaks with Dr. Chelle Gentemann, Open Science Program Scientist for NASA’s Office of the Chief Science Data Officer, about NASA’s ambitious efforts to integrate AI across the research lifecycle. In this episode, we’ll dive deeper into how AI is transforming NASA’s approach to science, making data more accessible and advancing open science practices. We explore</p>

<ul>
<li><strong>Measuring the Impact of Open Science:</strong> How NASA is developing new metrics to evaluate the effectiveness of open science, moving beyond traditional publication-based assessments.</li>
<li><strong>The Process of Scientific Discovery:</strong> Insights into the collaborative nature of research and how breakthroughs are achieved at NASA.</li>
<li>** AI Applications in NASA’s Science:** From rats in space to exploring the origins of the universe, we cover how AI is being applied across NASA’s divisions to improve data accessibility and analysis.</li>
<li><strong>Addressing Challenges in Open Science:</strong> The complexities of implementing open science within government agencies and research environments.</li>
<li><strong>Reforming Incentive Systems:</strong> How NASA is reconsidering traditional metrics like publications and citations, and starting to recognize contributions such as software development and data sharing.</li>
<li><strong>The Future of Open Science:</strong> How open science is shaping the future of research, fostering interdisciplinary collaboration, and increasing accessibility.</li>
</ul>

<p>This conversation offers valuable insights for researchers, data scientists, and those interested in the practical applications of AI and open science. Join us as we discuss how NASA is working to make science more collaborative, reproducible, and impactful.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://youtube.com/live/VJDg3ZbkNOE?feature=share" rel="nofollow">The livestream on YouTube</a></li>
<li><a href="https://openscience101.org/" rel="nofollow">NASA&#39;s Open Science 101 course &lt;-- do it to learn and also to get NASA Swag!</a></li>
<li><a href="https://sciencecast.org/" rel="nofollow">Science Cast</a></li>
<li><a href="https://www.earthdata.nasa.gov/news/impact-ibm-hls-foundation-model" rel="nofollow">NASA and IBM Openly Release Geospatial AI Foundation Model for NASA Earth Observation Data</a></li>
<li><a href="https://x.com/jakevdp/status/408678764705378304" rel="nofollow">Jake VanderPlas&#39; daily conundrum tweet from 2013</a></li>
<li><a href="https://replit.com/" rel="nofollow">Replit, &quot;an AI-powered software development &amp; deployment platform for building, sharing, and shipping software fast.&quot;</a></li>
</ul>]]>
  </content:encoded>
  <itunes:summary>
    <![CDATA[<p>Hugo speaks with Dr. Chelle Gentemann, Open Science Program Scientist for NASA’s Office of the Chief Science Data Officer, about NASA’s ambitious efforts to integrate AI across the research lifecycle. In this episode, we’ll dive deeper into how AI is transforming NASA’s approach to science, making data more accessible and advancing open science practices. We explore</p>

<ul>
<li><strong>Measuring the Impact of Open Science:</strong> How NASA is developing new metrics to evaluate the effectiveness of open science, moving beyond traditional publication-based assessments.</li>
<li><strong>The Process of Scientific Discovery:</strong> Insights into the collaborative nature of research and how breakthroughs are achieved at NASA.</li>
<li>** AI Applications in NASA’s Science:** From rats in space to exploring the origins of the universe, we cover how AI is being applied across NASA’s divisions to improve data accessibility and analysis.</li>
<li><strong>Addressing Challenges in Open Science:</strong> The complexities of implementing open science within government agencies and research environments.</li>
<li><strong>Reforming Incentive Systems:</strong> How NASA is reconsidering traditional metrics like publications and citations, and starting to recognize contributions such as software development and data sharing.</li>
<li><strong>The Future of Open Science:</strong> How open science is shaping the future of research, fostering interdisciplinary collaboration, and increasing accessibility.</li>
</ul>

<p>This conversation offers valuable insights for researchers, data scientists, and those interested in the practical applications of AI and open science. Join us as we discuss how NASA is working to make science more collaborative, reproducible, and impactful.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://youtube.com/live/VJDg3ZbkNOE?feature=share" rel="nofollow">The livestream on YouTube</a></li>
<li><a href="https://openscience101.org/" rel="nofollow">NASA&#39;s Open Science 101 course &lt;-- do it to learn and also to get NASA Swag!</a></li>
<li><a href="https://sciencecast.org/" rel="nofollow">Science Cast</a></li>
<li><a href="https://www.earthdata.nasa.gov/news/impact-ibm-hls-foundation-model" rel="nofollow">NASA and IBM Openly Release Geospatial AI Foundation Model for NASA Earth Observation Data</a></li>
<li><a href="https://x.com/jakevdp/status/408678764705378304" rel="nofollow">Jake VanderPlas&#39; daily conundrum tweet from 2013</a></li>
<li><a href="https://replit.com/" rel="nofollow">Replit, &quot;an AI-powered software development &amp; deployment platform for building, sharing, and shipping software fast.&quot;</a></li>
</ul>]]>
  </itunes:summary>
</item>
<item>
  <title>Episode 34: The AI Revolution Will Not Be Monopolized</title>
  <link>https://vanishinggradients.fireside.fm/34</link>
  <guid isPermaLink="false">8c18d59e-9b79-4682-8e3c-ba682daf1c1c</guid>
  <pubDate>Thu, 22 Aug 2024 17:00:00 +1000</pubDate>
  <author>Hugo Bowne-Anderson</author>
  <enclosure url="https://aphid.fireside.fm/d/1437767933/140c3904-8258-4c39-a698-a112b7077bd7/8c18d59e-9b79-4682-8e3c-ba682daf1c1c.mp3" length="98751972" type="audio/mpeg"/>
  <itunes:episodeType>full</itunes:episodeType>
  <itunes:season>1</itunes:season>
  <itunes:author>Hugo Bowne-Anderson</itunes:author>
  <itunes:subtitle>Hugo speaks with Ines Montani and Matthew Honnibal, the creators of spaCy and founders of Explosion AI. Collectively, they've had a huge impact on the fields of industrial natural language processing (NLP), ML, and AI through their widely-used open-source library spaCy and their innovative annotation tool Prodigy.</itunes:subtitle>
  <itunes:duration>1:42:51</itunes:duration>
  <itunes:explicit>yes</itunes:explicit>
  <itunes:image href="https://media24.fireside.fm/file/fireside-images-2024/podcasts/images/1/140c3904-8258-4c39-a698-a112b7077bd7/cover.jpg?v=1"/>
  <description>Hugo speaks with Ines Montani and Matthew Honnibal, the creators of spaCy and founders of Explosion AI. Collectively, they've had a huge impact on the fields of industrial natural language processing (NLP), ML, and AI through their widely-used open-source library spaCy and their innovative annotation tool Prodigy. These tools have become essential for many data scientists and NLP practitioners in industry and academia alike.
In this wide-ranging discussion, we dive into:
• The evolution of applied NLP and its role in industry
• The balance between large language models and smaller, specialized models
• Human-in-the-loop distillation for creating faster, more data-private AI systems
• The challenges and opportunities in NLP, including modularity, transparency, and privacy
• The future of AI and software development
• The potential impact of AI regulation on innovation and competition
We also touch on their recent transition back to a smaller, more independent-minded company structure and the lessons learned from their journey in the AI startup world.
Ines and Matt offer invaluable insights for data scientists, machine learning practitioners, and anyone interested in the practical applications of AI. They share their thoughts on how to approach NLP projects, the importance of data quality, and the role of open-source in advancing the field.
Whether you're a seasoned NLP practitioner or just getting started with AI, this episode offers a wealth of knowledge from two of the field's most respected figures. Join us for a discussion that explores the current landscape of AI development, with insights that bridge the gap between cutting-edge research and real-world applications.
LINKS
The livestream on YouTube (https://youtube.com/live/-6o5-3cP0ik?feature=share)
How S&amp;amp;P Global is making markets more transparent with NLP, spaCy and Prodigy (https://explosion.ai/blog/sp-global-commodities)
A practical guide to human-in-the-loop distillation (https://explosion.ai/blog/human-in-the-loop-distillation)
Laws of Tech: Commoditize Your Complement (https://gwern.net/complement)
spaCy: Industrial-Strength Natural Language Processing (https://spacy.io/)
LLMs with spaCy (https://spacy.io/usage/large-language-models)
Explosion, building developer tools for AI, Machine Learning and Natural Language Processing (https://explosion.ai/)
Back to our roots: Company update and future plans, by Matt and Ines (https://explosion.ai/blog/back-to-our-roots-company-update)
Matt's detailed blog post: back to our roots (https://honnibal.dev/blog/back-to-our-roots)
Ines on twitter (https://x.com/_inesmontani)
Matt on twitter (https://x.com/honnibal)
Vanishing Gradients on Twitter (https://twitter.com/vanishingdata)
Hugo on Twitter (https://twitter.com/hugobowne)
Check out and subcribe to our lu.ma calendar (https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk) for upcoming livestreams!
</description>
  <itunes:keywords>AI, LLMs, machine learning, data science, GenAI, NLP</itunes:keywords>
  <content:encoded>
    <![CDATA[<p>Hugo speaks with Ines Montani and Matthew Honnibal, the creators of spaCy and founders of Explosion AI. Collectively, they&#39;ve had a huge impact on the fields of industrial natural language processing (NLP), ML, and AI through their widely-used open-source library spaCy and their innovative annotation tool Prodigy. These tools have become essential for many data scientists and NLP practitioners in industry and academia alike.</p>

<p>In this wide-ranging discussion, we dive into:</p>

<p>• The evolution of applied NLP and its role in industry<br>
• The balance between large language models and smaller, specialized models<br>
• Human-in-the-loop distillation for creating faster, more data-private AI systems<br>
• The challenges and opportunities in NLP, including modularity, transparency, and privacy<br>
• The future of AI and software development<br>
• The potential impact of AI regulation on innovation and competition</p>

<p>We also touch on their recent transition back to a smaller, more independent-minded company structure and the lessons learned from their journey in the AI startup world.</p>

<p>Ines and Matt offer invaluable insights for data scientists, machine learning practitioners, and anyone interested in the practical applications of AI. They share their thoughts on how to approach NLP projects, the importance of data quality, and the role of open-source in advancing the field.</p>

<p>Whether you&#39;re a seasoned NLP practitioner or just getting started with AI, this episode offers a wealth of knowledge from two of the field&#39;s most respected figures. Join us for a discussion that explores the current landscape of AI development, with insights that bridge the gap between cutting-edge research and real-world applications.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://youtube.com/live/-6o5-3cP0ik?feature=share" rel="nofollow">The livestream on YouTube</a></li>
<li><a href="https://explosion.ai/blog/sp-global-commodities" rel="nofollow">How S&amp;P Global is making markets more transparent with NLP, spaCy and Prodigy</a></li>
<li><a href="https://explosion.ai/blog/human-in-the-loop-distillation" rel="nofollow">A practical guide to human-in-the-loop distillation</a></li>
<li><a href="https://gwern.net/complement" rel="nofollow">Laws of Tech: Commoditize Your Complement</a></li>
<li><a href="https://spacy.io/" rel="nofollow">spaCy: Industrial-Strength Natural Language Processing</a></li>
<li><a href="https://spacy.io/usage/large-language-models" rel="nofollow">LLMs with spaCy</a></li>
<li><a href="https://explosion.ai/" rel="nofollow">Explosion, building developer tools for AI, Machine Learning and Natural Language Processing</a></li>
<li><a href="https://explosion.ai/blog/back-to-our-roots-company-update" rel="nofollow">Back to our roots: Company update and future plans, by Matt and Ines</a></li>
<li><a href="https://honnibal.dev/blog/back-to-our-roots" rel="nofollow">Matt&#39;s detailed blog post: back to our roots</a></li>
<li><a href="https://x.com/_inesmontani" rel="nofollow">Ines on twitter</a></li>
<li><a href="https://x.com/honnibal" rel="nofollow">Matt on twitter</a></li>
<li><a href="https://twitter.com/vanishingdata" rel="nofollow">Vanishing Gradients on Twitter</a></li>
<li><a href="https://twitter.com/hugobowne" rel="nofollow">Hugo on Twitter</a></li>
</ul>

<p>Check out and subcribe to our <a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">lu.ma calendar</a> for upcoming livestreams!</p>]]>
  </content:encoded>
  <itunes:summary>
    <![CDATA[<p>Hugo speaks with Ines Montani and Matthew Honnibal, the creators of spaCy and founders of Explosion AI. Collectively, they&#39;ve had a huge impact on the fields of industrial natural language processing (NLP), ML, and AI through their widely-used open-source library spaCy and their innovative annotation tool Prodigy. These tools have become essential for many data scientists and NLP practitioners in industry and academia alike.</p>

<p>In this wide-ranging discussion, we dive into:</p>

<p>• The evolution of applied NLP and its role in industry<br>
• The balance between large language models and smaller, specialized models<br>
• Human-in-the-loop distillation for creating faster, more data-private AI systems<br>
• The challenges and opportunities in NLP, including modularity, transparency, and privacy<br>
• The future of AI and software development<br>
• The potential impact of AI regulation on innovation and competition</p>

<p>We also touch on their recent transition back to a smaller, more independent-minded company structure and the lessons learned from their journey in the AI startup world.</p>

<p>Ines and Matt offer invaluable insights for data scientists, machine learning practitioners, and anyone interested in the practical applications of AI. They share their thoughts on how to approach NLP projects, the importance of data quality, and the role of open-source in advancing the field.</p>

<p>Whether you&#39;re a seasoned NLP practitioner or just getting started with AI, this episode offers a wealth of knowledge from two of the field&#39;s most respected figures. Join us for a discussion that explores the current landscape of AI development, with insights that bridge the gap between cutting-edge research and real-world applications.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://youtube.com/live/-6o5-3cP0ik?feature=share" rel="nofollow">The livestream on YouTube</a></li>
<li><a href="https://explosion.ai/blog/sp-global-commodities" rel="nofollow">How S&amp;P Global is making markets more transparent with NLP, spaCy and Prodigy</a></li>
<li><a href="https://explosion.ai/blog/human-in-the-loop-distillation" rel="nofollow">A practical guide to human-in-the-loop distillation</a></li>
<li><a href="https://gwern.net/complement" rel="nofollow">Laws of Tech: Commoditize Your Complement</a></li>
<li><a href="https://spacy.io/" rel="nofollow">spaCy: Industrial-Strength Natural Language Processing</a></li>
<li><a href="https://spacy.io/usage/large-language-models" rel="nofollow">LLMs with spaCy</a></li>
<li><a href="https://explosion.ai/" rel="nofollow">Explosion, building developer tools for AI, Machine Learning and Natural Language Processing</a></li>
<li><a href="https://explosion.ai/blog/back-to-our-roots-company-update" rel="nofollow">Back to our roots: Company update and future plans, by Matt and Ines</a></li>
<li><a href="https://honnibal.dev/blog/back-to-our-roots" rel="nofollow">Matt&#39;s detailed blog post: back to our roots</a></li>
<li><a href="https://x.com/_inesmontani" rel="nofollow">Ines on twitter</a></li>
<li><a href="https://x.com/honnibal" rel="nofollow">Matt on twitter</a></li>
<li><a href="https://twitter.com/vanishingdata" rel="nofollow">Vanishing Gradients on Twitter</a></li>
<li><a href="https://twitter.com/hugobowne" rel="nofollow">Hugo on Twitter</a></li>
</ul>

<p>Check out and subcribe to our <a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">lu.ma calendar</a> for upcoming livestreams!</p>]]>
  </itunes:summary>
</item>
<item>
  <title>Episode 32: Building Reliable and Robust ML/AI Pipelines</title>
  <link>https://vanishinggradients.fireside.fm/32</link>
  <guid isPermaLink="false">3aa4ba58-30aa-4a85-a139-e9057629171c</guid>
  <pubDate>Sat, 27 Jul 2024 13:00:00 +1000</pubDate>
  <author>Hugo Bowne-Anderson</author>
  <enclosure url="https://aphid.fireside.fm/d/1437767933/140c3904-8258-4c39-a698-a112b7077bd7/3aa4ba58-30aa-4a85-a139-e9057629171c.mp3" length="72173111" type="audio/mpeg"/>
  <itunes:episodeType>full</itunes:episodeType>
  <itunes:season>1</itunes:season>
  <itunes:author>Hugo Bowne-Anderson</itunes:author>
  <itunes:subtitle>Hugo speaks with Shreya Shankar, a researcher at UC Berkeley focusing on data management systems with a human-centered approach. Shreya's work is at the cutting edge of human-computer interaction (HCI) and AI, particularly in the realm of large language models (LLMs). Her impressive background includes being the first ML engineer at Viaduct, doing research engineering at Google Brain, and software engineering at Facebook.</itunes:subtitle>
  <itunes:duration>1:15:10</itunes:duration>
  <itunes:explicit>yes</itunes:explicit>
  <itunes:image href="https://media24.fireside.fm/file/fireside-images-2024/podcasts/images/1/140c3904-8258-4c39-a698-a112b7077bd7/cover.jpg?v=1"/>
  <description>Hugo speaks with Shreya Shankar, a researcher at UC Berkeley focusing on data management systems with a human-centered approach. Shreya's work is at the cutting edge of human-computer interaction (HCI) and AI, particularly in the realm of large language models (LLMs). Her impressive background includes being the first ML engineer at Viaduct, doing research engineering at Google Brain, and software engineering at Facebook.
In this episode, we dive deep into the world of LLMs and the critical challenges of building reliable AI pipelines. We'll explore:
The fascinating journey from classic machine learning to the current LLM revolution
Why Shreya believes most ML problems are actually data management issues
The concept of "data flywheels" for LLM applications and how to implement them
The intriguing world of evaluating AI systems - who validates the validators?
Shreya's work on SPADE and EvalGen, innovative tools for synthesizing data quality assertions and aligning LLM evaluations with human preferences
The importance of human-in-the-loop processes in AI development
The future of low-code and no-code tools in the AI landscape
We'll also touch on the potential pitfalls of over-relying on LLMs, the concept of "Habsburg AI," and how to avoid disappearing up our own proverbial arseholes in the world of recursive AI processes.
Whether you're a seasoned AI practitioner, a curious data scientist, or someone interested in the human side of AI development, this conversation offers valuable insights into building more robust, reliable, and human-centered AI systems.
LINKS
The livestream on YouTube (https://youtube.com/live/hKV6xSJZkB0?feature=share)
Shreya's website (https://www.sh-reya.com/)
Shreya on Twitter (https://x.com/sh_reya)
Data Flywheels for LLM Applications (https://www.sh-reya.com/blog/ai-engineering-flywheel/)
SPADE: Synthesizing Data Quality Assertions for Large Language Model Pipelines (https://arxiv.org/abs/2401.03038)
What We’ve Learned From A Year of Building with LLMs (https://applied-llms.org/)
Who Validates the Validators? Aligning LLM-Assisted Evaluation of LLM Outputs with Human Preferences (https://arxiv.org/abs/2404.12272)
Operationalizing Machine Learning: An Interview Study (https://arxiv.org/abs/2209.09125)
Vanishing Gradients on Twitter (https://twitter.com/vanishingdata)
Hugo on Twitter (https://twitter.com/hugobowne)
In the podcast, Hugo also mentioned that this was the 5th time he and Shreya chatted publicly. which is wild!
If you want to dive deep into Shreya's work and related topics through their chats, you can check them all out here:
Outerbounds' Fireside Chat: Operationalizing ML -- Patterns and Pain Points from MLOps Practitioners (https://www.youtube.com/watch?v=7zB6ESFto_U)
The Past, Present, and Future of Generative AI (https://youtu.be/q0A9CdGWXqc?si=XmaUnQmZiXL2eagS)
LLMs, OpenAI Dev Day, and the Existential Crisis for Machine Learning Engineering (https://www.youtube.com/live/MTJHvgJtynU?si=Ncjqn5YuFBemvOJ0)
Lessons from a Year of Building with LLMs (https://youtube.com/live/c0gcsprsFig?feature=share)
Check out and subcribe to our lu.ma calendar (https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk) for upcoming livestreams! 
</description>
  <itunes:keywords>AI, LLMs, machine learning, data science, GenAI</itunes:keywords>
  <content:encoded>
    <![CDATA[<p>Hugo speaks with Shreya Shankar, a researcher at UC Berkeley focusing on data management systems with a human-centered approach. Shreya&#39;s work is at the cutting edge of human-computer interaction (HCI) and AI, particularly in the realm of large language models (LLMs). Her impressive background includes being the first ML engineer at Viaduct, doing research engineering at Google Brain, and software engineering at Facebook.</p>

<p>In this episode, we dive deep into the world of LLMs and the critical challenges of building reliable AI pipelines. We&#39;ll explore:</p>

<ul>
<li>The fascinating journey from classic machine learning to the current LLM revolution</li>
<li>Why Shreya believes most ML problems are actually data management issues</li>
<li>The concept of &quot;data flywheels&quot; for LLM applications and how to implement them</li>
<li>The intriguing world of evaluating AI systems - who validates the validators?</li>
<li>Shreya&#39;s work on SPADE and EvalGen, innovative tools for synthesizing data quality assertions and aligning LLM evaluations with human preferences</li>
<li>The importance of human-in-the-loop processes in AI development</li>
<li>The future of low-code and no-code tools in the AI landscape</li>
</ul>

<p>We&#39;ll also touch on the potential pitfalls of over-relying on LLMs, the concept of &quot;Habsburg AI,&quot; and how to avoid disappearing up our own proverbial arseholes in the world of recursive AI processes.</p>

<p>Whether you&#39;re a seasoned AI practitioner, a curious data scientist, or someone interested in the human side of AI development, this conversation offers valuable insights into building more robust, reliable, and human-centered AI systems.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://youtube.com/live/hKV6xSJZkB0?feature=share" rel="nofollow">The livestream on YouTube</a></li>
<li><a href="https://www.sh-reya.com/" rel="nofollow">Shreya&#39;s website</a></li>
<li><a href="https://x.com/sh_reya" rel="nofollow">Shreya on Twitter</a></li>
<li><a href="https://www.sh-reya.com/blog/ai-engineering-flywheel/" rel="nofollow">Data Flywheels for LLM Applications</a></li>
<li><a href="https://arxiv.org/abs/2401.03038" rel="nofollow">SPADE: Synthesizing Data Quality Assertions for Large Language Model Pipelines</a></li>
<li><a href="https://applied-llms.org/" rel="nofollow">What We’ve Learned From A Year of Building with LLMs</a></li>
<li><a href="https://arxiv.org/abs/2404.12272" rel="nofollow">Who Validates the Validators? Aligning LLM-Assisted Evaluation of LLM Outputs with Human Preferences</a></li>
<li><a href="https://arxiv.org/abs/2209.09125" rel="nofollow">Operationalizing Machine Learning: An Interview Study</a></li>
<li><a href="https://twitter.com/vanishingdata" rel="nofollow">Vanishing Gradients on Twitter</a></li>
<li><a href="https://twitter.com/hugobowne" rel="nofollow">Hugo on Twitter</a></li>
</ul>

<p>In the podcast, Hugo also mentioned that this was the 5th time he and Shreya chatted publicly. which is wild!</p>

<p>If you want to dive deep into Shreya&#39;s work and related topics through their chats, you can check them all out here:</p>

<ol>
<li><a href="https://www.youtube.com/watch?v=7zB6ESFto_U" rel="nofollow">Outerbounds&#39; Fireside Chat: Operationalizing ML -- Patterns and Pain Points from MLOps Practitioners</a></li>
<li><a href="https://youtu.be/q0A9CdGWXqc?si=XmaUnQmZiXL2eagS" rel="nofollow">The Past, Present, and Future of Generative AI</a></li>
<li><a href="https://www.youtube.com/live/MTJHvgJtynU?si=Ncjqn5YuFBemvOJ0" rel="nofollow">LLMs, OpenAI Dev Day, and the Existential Crisis for Machine Learning Engineering</a></li>
<li><a href="https://youtube.com/live/c0gcsprsFig?feature=share" rel="nofollow">Lessons from a Year of Building with LLMs</a></li>
</ol>

<p>Check out and subcribe to our <a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">lu.ma calendar</a> for upcoming livestreams!</p>]]>
  </content:encoded>
  <itunes:summary>
    <![CDATA[<p>Hugo speaks with Shreya Shankar, a researcher at UC Berkeley focusing on data management systems with a human-centered approach. Shreya&#39;s work is at the cutting edge of human-computer interaction (HCI) and AI, particularly in the realm of large language models (LLMs). Her impressive background includes being the first ML engineer at Viaduct, doing research engineering at Google Brain, and software engineering at Facebook.</p>

<p>In this episode, we dive deep into the world of LLMs and the critical challenges of building reliable AI pipelines. We&#39;ll explore:</p>

<ul>
<li>The fascinating journey from classic machine learning to the current LLM revolution</li>
<li>Why Shreya believes most ML problems are actually data management issues</li>
<li>The concept of &quot;data flywheels&quot; for LLM applications and how to implement them</li>
<li>The intriguing world of evaluating AI systems - who validates the validators?</li>
<li>Shreya&#39;s work on SPADE and EvalGen, innovative tools for synthesizing data quality assertions and aligning LLM evaluations with human preferences</li>
<li>The importance of human-in-the-loop processes in AI development</li>
<li>The future of low-code and no-code tools in the AI landscape</li>
</ul>

<p>We&#39;ll also touch on the potential pitfalls of over-relying on LLMs, the concept of &quot;Habsburg AI,&quot; and how to avoid disappearing up our own proverbial arseholes in the world of recursive AI processes.</p>

<p>Whether you&#39;re a seasoned AI practitioner, a curious data scientist, or someone interested in the human side of AI development, this conversation offers valuable insights into building more robust, reliable, and human-centered AI systems.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://youtube.com/live/hKV6xSJZkB0?feature=share" rel="nofollow">The livestream on YouTube</a></li>
<li><a href="https://www.sh-reya.com/" rel="nofollow">Shreya&#39;s website</a></li>
<li><a href="https://x.com/sh_reya" rel="nofollow">Shreya on Twitter</a></li>
<li><a href="https://www.sh-reya.com/blog/ai-engineering-flywheel/" rel="nofollow">Data Flywheels for LLM Applications</a></li>
<li><a href="https://arxiv.org/abs/2401.03038" rel="nofollow">SPADE: Synthesizing Data Quality Assertions for Large Language Model Pipelines</a></li>
<li><a href="https://applied-llms.org/" rel="nofollow">What We’ve Learned From A Year of Building with LLMs</a></li>
<li><a href="https://arxiv.org/abs/2404.12272" rel="nofollow">Who Validates the Validators? Aligning LLM-Assisted Evaluation of LLM Outputs with Human Preferences</a></li>
<li><a href="https://arxiv.org/abs/2209.09125" rel="nofollow">Operationalizing Machine Learning: An Interview Study</a></li>
<li><a href="https://twitter.com/vanishingdata" rel="nofollow">Vanishing Gradients on Twitter</a></li>
<li><a href="https://twitter.com/hugobowne" rel="nofollow">Hugo on Twitter</a></li>
</ul>

<p>In the podcast, Hugo also mentioned that this was the 5th time he and Shreya chatted publicly. which is wild!</p>

<p>If you want to dive deep into Shreya&#39;s work and related topics through their chats, you can check them all out here:</p>

<ol>
<li><a href="https://www.youtube.com/watch?v=7zB6ESFto_U" rel="nofollow">Outerbounds&#39; Fireside Chat: Operationalizing ML -- Patterns and Pain Points from MLOps Practitioners</a></li>
<li><a href="https://youtu.be/q0A9CdGWXqc?si=XmaUnQmZiXL2eagS" rel="nofollow">The Past, Present, and Future of Generative AI</a></li>
<li><a href="https://www.youtube.com/live/MTJHvgJtynU?si=Ncjqn5YuFBemvOJ0" rel="nofollow">LLMs, OpenAI Dev Day, and the Existential Crisis for Machine Learning Engineering</a></li>
<li><a href="https://youtube.com/live/c0gcsprsFig?feature=share" rel="nofollow">Lessons from a Year of Building with LLMs</a></li>
</ol>

<p>Check out and subcribe to our <a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">lu.ma calendar</a> for upcoming livestreams!</p>]]>
  </itunes:summary>
</item>
<item>
  <title>Episode 31: Rethinking Data Science, Machine Learning, and AI</title>
  <link>https://vanishinggradients.fireside.fm/31</link>
  <guid isPermaLink="false">455d1587-7ba6-4850-920e-360d8cbe33d3</guid>
  <pubDate>Tue, 09 Jul 2024 19:00:00 +1000</pubDate>
  <author>Hugo Bowne-Anderson</author>
  <enclosure url="https://aphid.fireside.fm/d/1437767933/140c3904-8258-4c39-a698-a112b7077bd7/455d1587-7ba6-4850-920e-360d8cbe33d3.mp3" length="92236825" type="audio/mpeg"/>
  <itunes:episodeType>full</itunes:episodeType>
  <itunes:season>1</itunes:season>
  <itunes:author>Hugo Bowne-Anderson</itunes:author>
  <itunes:subtitle>Hugo speaks with Vincent Warmerdam, a senior data professional and machine learning engineer at :probabl, the exclusive brand operator of scikit-learn. Vincent is known for challenging common assumptions and exploring innovative approaches in data science and machine learning.</itunes:subtitle>
  <itunes:duration>1:36:04</itunes:duration>
  <itunes:explicit>no</itunes:explicit>
  <itunes:image href="https://media24.fireside.fm/file/fireside-images-2024/podcasts/images/1/140c3904-8258-4c39-a698-a112b7077bd7/cover.jpg?v=1"/>
  <description>Hugo speaks with Vincent Warmerdam, a senior data professional and machine learning engineer at :probabl, the exclusive brand operator of scikit-learn. Vincent is known for challenging common assumptions and exploring innovative approaches in data science and machine learning.
In this episode, they dive deep into rethinking established methods in data science, machine learning, and AI. We explore Vincent's principled approach to the field, including:
The critical importance of exposing yourself to real-world problems before applying ML solutions
Framing problems correctly and understanding the data generating process
The power of visualization and human intuition in data analysis
Questioning whether algorithms truly meet the actual problem at hand
The value of simple, interpretable models and when to consider more complex approaches
The importance of UI and user experience in data science tools
Strategies for preventing algorithmic failures by rethinking evaluation metrics and data quality
The potential and limitations of LLMs in the current data science landscape
The benefits of open-source collaboration and knowledge sharing in the community
Throughout the conversation, Vincent illustrates these principles with vivid, real-world examples from his extensive experience in the field. They also discuss Vincent's thoughts on the future of data science and his call to action for more knowledge sharing in the community through blogging and open dialogue.
LINKS
The livestream on YouTube (https://youtube.com/live/-CD66CI1pEo?feature=share)
Vincent's blog (https://koaning.io/)
CalmCode (https://calmcode.io/)
scikit-lego (https://koaning.github.io/scikit-lego/)
Vincent's book Data Science Fiction (WIP) (https://calmcode.io/book)
The Deon Checklist, an ethics checklist for data scientists (https://deon.drivendata.org/)
Of oaths and checklists, by DJ Patil, Hilary Mason and Mike Loukides (https://www.oreilly.com/radar/of-oaths-and-checklists/)
Vincent's Getting Started with NLP and spaCy Course course on Talk Python (https://training.talkpython.fm/courses/getting-started-with-spacy)
Vincent on twitter (https://x.com/fishnets88)
:probabl. on twitter (https://x.com/probabl_ai)
Vincent's PyData Amsterdam Keynote "Natural Intelligence is All You Need [tm]" (https://www.youtube.com/watch?v=C9p7suS-NGk)
Vincent's PyData Amsterdam 2019 talk: The profession of solving (the wrong problem)  (https://www.youtube.com/watch?v=kYMfE9u-lMo)
Vanishing Gradients on Twitter (https://twitter.com/vanishingdata)
Hugo on Twitter (https://twitter.com/hugobowne)
Check out and subcribe to our lu.ma calendar (https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk) for upcoming livestreams! 
</description>
  <itunes:keywords>AI, LLMs, machine learning, data science, GenAI</itunes:keywords>
  <content:encoded>
    <![CDATA[<p>Hugo speaks with Vincent Warmerdam, a senior data professional and machine learning engineer at :probabl, the exclusive brand operator of scikit-learn. Vincent is known for challenging common assumptions and exploring innovative approaches in data science and machine learning.</p>

<p>In this episode, they dive deep into rethinking established methods in data science, machine learning, and AI. We explore Vincent&#39;s principled approach to the field, including:</p>

<ul>
<li>The critical importance of exposing yourself to real-world problems before applying ML solutions</li>
<li>Framing problems correctly and understanding the data generating process</li>
<li>The power of visualization and human intuition in data analysis</li>
<li>Questioning whether algorithms truly meet the actual problem at hand</li>
<li>The value of simple, interpretable models and when to consider more complex approaches</li>
<li>The importance of UI and user experience in data science tools</li>
<li>Strategies for preventing algorithmic failures by rethinking evaluation metrics and data quality</li>
<li>The potential and limitations of LLMs in the current data science landscape</li>
<li>The benefits of open-source collaboration and knowledge sharing in the community</li>
</ul>

<p>Throughout the conversation, Vincent illustrates these principles with vivid, real-world examples from his extensive experience in the field. They also discuss Vincent&#39;s thoughts on the future of data science and his call to action for more knowledge sharing in the community through blogging and open dialogue.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://youtube.com/live/-CD66CI1pEo?feature=share" rel="nofollow">The livestream on YouTube</a></li>
<li><a href="https://koaning.io/" rel="nofollow">Vincent&#39;s blog</a></li>
<li><a href="https://calmcode.io/" rel="nofollow">CalmCode</a></li>
<li><a href="https://koaning.github.io/scikit-lego/" rel="nofollow">scikit-lego</a></li>
<li><a href="https://calmcode.io/book" rel="nofollow">Vincent&#39;s book Data Science Fiction (WIP)</a></li>
<li><a href="https://deon.drivendata.org/" rel="nofollow">The Deon Checklist, an ethics checklist for data scientists</a></li>
<li><a href="https://www.oreilly.com/radar/of-oaths-and-checklists/" rel="nofollow">Of oaths and checklists, by DJ Patil, Hilary Mason and Mike Loukides</a></li>
<li><a href="https://training.talkpython.fm/courses/getting-started-with-spacy" rel="nofollow">Vincent&#39;s Getting Started with NLP and spaCy Course course on Talk Python</a></li>
<li><a href="https://x.com/fishnets88" rel="nofollow">Vincent on twitter</a></li>
<li><a href="https://x.com/probabl_ai" rel="nofollow">:probabl. on twitter</a></li>
<li><a href="https://www.youtube.com/watch?v=C9p7suS-NGk" rel="nofollow">Vincent&#39;s PyData Amsterdam Keynote &quot;Natural Intelligence is All You Need [tm]&quot;</a></li>
<li><a href="https://www.youtube.com/watch?v=kYMfE9u-lMo" rel="nofollow">Vincent&#39;s PyData Amsterdam 2019 talk: The profession of solving (the wrong problem) </a></li>
<li><a href="https://twitter.com/vanishingdata" rel="nofollow">Vanishing Gradients on Twitter</a></li>
<li><a href="https://twitter.com/hugobowne" rel="nofollow">Hugo on Twitter</a></li>
</ul>

<p>Check out and subcribe to our <a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">lu.ma calendar</a> for upcoming livestreams!</p>]]>
  </content:encoded>
  <itunes:summary>
    <![CDATA[<p>Hugo speaks with Vincent Warmerdam, a senior data professional and machine learning engineer at :probabl, the exclusive brand operator of scikit-learn. Vincent is known for challenging common assumptions and exploring innovative approaches in data science and machine learning.</p>

<p>In this episode, they dive deep into rethinking established methods in data science, machine learning, and AI. We explore Vincent&#39;s principled approach to the field, including:</p>

<ul>
<li>The critical importance of exposing yourself to real-world problems before applying ML solutions</li>
<li>Framing problems correctly and understanding the data generating process</li>
<li>The power of visualization and human intuition in data analysis</li>
<li>Questioning whether algorithms truly meet the actual problem at hand</li>
<li>The value of simple, interpretable models and when to consider more complex approaches</li>
<li>The importance of UI and user experience in data science tools</li>
<li>Strategies for preventing algorithmic failures by rethinking evaluation metrics and data quality</li>
<li>The potential and limitations of LLMs in the current data science landscape</li>
<li>The benefits of open-source collaboration and knowledge sharing in the community</li>
</ul>

<p>Throughout the conversation, Vincent illustrates these principles with vivid, real-world examples from his extensive experience in the field. They also discuss Vincent&#39;s thoughts on the future of data science and his call to action for more knowledge sharing in the community through blogging and open dialogue.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://youtube.com/live/-CD66CI1pEo?feature=share" rel="nofollow">The livestream on YouTube</a></li>
<li><a href="https://koaning.io/" rel="nofollow">Vincent&#39;s blog</a></li>
<li><a href="https://calmcode.io/" rel="nofollow">CalmCode</a></li>
<li><a href="https://koaning.github.io/scikit-lego/" rel="nofollow">scikit-lego</a></li>
<li><a href="https://calmcode.io/book" rel="nofollow">Vincent&#39;s book Data Science Fiction (WIP)</a></li>
<li><a href="https://deon.drivendata.org/" rel="nofollow">The Deon Checklist, an ethics checklist for data scientists</a></li>
<li><a href="https://www.oreilly.com/radar/of-oaths-and-checklists/" rel="nofollow">Of oaths and checklists, by DJ Patil, Hilary Mason and Mike Loukides</a></li>
<li><a href="https://training.talkpython.fm/courses/getting-started-with-spacy" rel="nofollow">Vincent&#39;s Getting Started with NLP and spaCy Course course on Talk Python</a></li>
<li><a href="https://x.com/fishnets88" rel="nofollow">Vincent on twitter</a></li>
<li><a href="https://x.com/probabl_ai" rel="nofollow">:probabl. on twitter</a></li>
<li><a href="https://www.youtube.com/watch?v=C9p7suS-NGk" rel="nofollow">Vincent&#39;s PyData Amsterdam Keynote &quot;Natural Intelligence is All You Need [tm]&quot;</a></li>
<li><a href="https://www.youtube.com/watch?v=kYMfE9u-lMo" rel="nofollow">Vincent&#39;s PyData Amsterdam 2019 talk: The profession of solving (the wrong problem) </a></li>
<li><a href="https://twitter.com/vanishingdata" rel="nofollow">Vanishing Gradients on Twitter</a></li>
<li><a href="https://twitter.com/hugobowne" rel="nofollow">Hugo on Twitter</a></li>
</ul>

<p>Check out and subcribe to our <a href="https://lu.ma/calendar/cal-8ImWFDQ3IEIxNWk" rel="nofollow">lu.ma calendar</a> for upcoming livestreams!</p>]]>
  </itunes:summary>
</item>
<item>
  <title>Episode 28: Beyond Supervised Learning: The Rise of In-Context Learning with LLMs</title>
  <link>https://vanishinggradients.fireside.fm/28</link>
  <guid isPermaLink="false">b268a89e-4fc9-4f9f-a2a5-c7636b3fbd70</guid>
  <pubDate>Mon, 10 Jun 2024 08:00:00 +1000</pubDate>
  <author>Hugo Bowne-Anderson</author>
  <enclosure url="https://aphid.fireside.fm/d/1437767933/140c3904-8258-4c39-a698-a112b7077bd7/b268a89e-4fc9-4f9f-a2a5-c7636b3fbd70.mp3" length="63014789" type="audio/mpeg"/>
  <itunes:episodeType>full</itunes:episodeType>
  <itunes:season>1</itunes:season>
  <itunes:author>Hugo Bowne-Anderson</itunes:author>
  <itunes:subtitle>Hugo speaks with Alan Nichol, co-founder and CTO of Rasa, where they build software to enable developers to create enterprise-grade conversational AI and chatbot systems across industries like telcos, healthcare, fintech, and government.</itunes:subtitle>
  <itunes:duration>1:05:38</itunes:duration>
  <itunes:explicit>no</itunes:explicit>
  <itunes:image href="https://media24.fireside.fm/file/fireside-images-2024/podcasts/images/1/140c3904-8258-4c39-a698-a112b7077bd7/cover.jpg?v=1"/>
  <description>Hugo speaks with Alan Nichol, co-founder and CTO of Rasa, where they build software to enable developers to create enterprise-grade conversational AI and chatbot systems across industries like telcos, healthcare, fintech, and government.
What's super cool is that Alan and the Rasa team have been doing this type of thing for over a decade, giving them a wealth of wisdom on how to effectively incorporate LLMs into chatbots - and how not to. For example, if you want a chatbot that takes specific and important actions like transferring money, do you want to fully entrust the conversation to one big LLM like ChatGPT, or secure what the LLMs can do inside key business logic?
In this episode, they also dive into the history of conversational AI and explore how the advent of LLMs is reshaping the field. Alan shares his perspective on how supervised learning has failed us in some ways and discusses what he sees as the most overrated and underrated aspects of LLMs.
Alan offers advice for those looking to work with LLMs and conversational AI, emphasizing the importance of not sleeping on proven techniques and looking beyond the latest hype. In a live demo, he showcases Rasa's Calm (Conversational AI with Language Models), which allows developers to define business logic declaratively and separate it from the LLM, enabling reliable execution of conversational flows.
LINKS
The livestream on YouTube (https://www.youtube.com/live/kMFBYC2pB30?si=yV5sGq1iuC47LBSi)
Alan's Rasa CALM Demo: Building Conversational AI with LLMs  (https://youtu.be/4UnxaJ-GcT0?si=6uLY3GD5DkOmWiBW)
Alan on twitter.com (https://x.com/alanmnichol)
Rasa (https://rasa.com/)
CALM, an LLM-native approach to building reliable conversational AI (https://rasa.com/docs/rasa-pro/calm/)
Task-Oriented Dialogue with In-Context Learning (https://arxiv.org/abs/2402.12234)
'We don’t know how to build conversational software yet' by Alan Nicol (https://medium.com/rasa-blog/we-don-t-know-how-to-build-conversational-software-yet-a18301db0e4b)
Vanishing Gradients on Twitter (https://twitter.com/vanishingdata)
Hugo on Twitter (https://twitter.com/hugobowne)
Upcoming Livestreams
Lessons from a Year of Building with LLMs (https://lu.ma/e8huz3s6?utm_source=vgan)
VALIDATING THE VALIDATORS with Shreya Shanker (https://lu.ma/zz3qic45?utm_source=vgan)
</description>
  <itunes:keywords>AI, LLMs, machine learning, data science</itunes:keywords>
  <content:encoded>
    <![CDATA[<p>Hugo speaks with Alan Nichol, co-founder and CTO of Rasa, where they build software to enable developers to create enterprise-grade conversational AI and chatbot systems across industries like telcos, healthcare, fintech, and government.</p>

<p>What&#39;s super cool is that Alan and the Rasa team have been doing this type of thing for over a decade, giving them a wealth of wisdom on how to effectively incorporate LLMs into chatbots - and how not to. For example, if you want a chatbot that takes specific and important actions like transferring money, do you want to fully entrust the conversation to one big LLM like ChatGPT, or secure what the LLMs can do inside key business logic?</p>

<p>In this episode, they also dive into the history of conversational AI and explore how the advent of LLMs is reshaping the field. Alan shares his perspective on how supervised learning has failed us in some ways and discusses what he sees as the most overrated and underrated aspects of LLMs.</p>

<p>Alan offers advice for those looking to work with LLMs and conversational AI, emphasizing the importance of not sleeping on proven techniques and looking beyond the latest hype. In a live demo, he showcases Rasa&#39;s Calm (Conversational AI with Language Models), which allows developers to define business logic declaratively and separate it from the LLM, enabling reliable execution of conversational flows.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://www.youtube.com/live/kMFBYC2pB30?si=yV5sGq1iuC47LBSi" rel="nofollow">The livestream on YouTube</a></li>
<li><a href="https://youtu.be/4UnxaJ-GcT0?si=6uLY3GD5DkOmWiBW" rel="nofollow">Alan&#39;s Rasa CALM Demo: Building Conversational AI with LLMs </a></li>
<li><a href="https://x.com/alanmnichol" rel="nofollow">Alan on twitter.com</a></li>
<li><a href="https://rasa.com/" rel="nofollow">Rasa</a></li>
<li><a href="https://rasa.com/docs/rasa-pro/calm/" rel="nofollow">CALM, an LLM-native approach to building reliable conversational AI</a></li>
<li><a href="https://arxiv.org/abs/2402.12234" rel="nofollow">Task-Oriented Dialogue with In-Context Learning</a></li>
<li><a href="https://medium.com/rasa-blog/we-don-t-know-how-to-build-conversational-software-yet-a18301db0e4b" rel="nofollow">&#39;We don’t know how to build conversational software yet&#39; by Alan Nicol</a></li>
<li><a href="https://twitter.com/vanishingdata" rel="nofollow">Vanishing Gradients on Twitter</a></li>
<li><a href="https://twitter.com/hugobowne" rel="nofollow">Hugo on Twitter</a></li>
</ul>

<p><strong>Upcoming Livestreams</strong></p>

<ul>
<li><a href="https://lu.ma/e8huz3s6?utm_source=vgan" rel="nofollow">Lessons from a Year of Building with LLMs</a></li>
<li><a href="https://lu.ma/zz3qic45?utm_source=vgan" rel="nofollow">VALIDATING THE VALIDATORS with Shreya Shanker</a></li>
</ul>]]>
  </content:encoded>
  <itunes:summary>
    <![CDATA[<p>Hugo speaks with Alan Nichol, co-founder and CTO of Rasa, where they build software to enable developers to create enterprise-grade conversational AI and chatbot systems across industries like telcos, healthcare, fintech, and government.</p>

<p>What&#39;s super cool is that Alan and the Rasa team have been doing this type of thing for over a decade, giving them a wealth of wisdom on how to effectively incorporate LLMs into chatbots - and how not to. For example, if you want a chatbot that takes specific and important actions like transferring money, do you want to fully entrust the conversation to one big LLM like ChatGPT, or secure what the LLMs can do inside key business logic?</p>

<p>In this episode, they also dive into the history of conversational AI and explore how the advent of LLMs is reshaping the field. Alan shares his perspective on how supervised learning has failed us in some ways and discusses what he sees as the most overrated and underrated aspects of LLMs.</p>

<p>Alan offers advice for those looking to work with LLMs and conversational AI, emphasizing the importance of not sleeping on proven techniques and looking beyond the latest hype. In a live demo, he showcases Rasa&#39;s Calm (Conversational AI with Language Models), which allows developers to define business logic declaratively and separate it from the LLM, enabling reliable execution of conversational flows.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://www.youtube.com/live/kMFBYC2pB30?si=yV5sGq1iuC47LBSi" rel="nofollow">The livestream on YouTube</a></li>
<li><a href="https://youtu.be/4UnxaJ-GcT0?si=6uLY3GD5DkOmWiBW" rel="nofollow">Alan&#39;s Rasa CALM Demo: Building Conversational AI with LLMs </a></li>
<li><a href="https://x.com/alanmnichol" rel="nofollow">Alan on twitter.com</a></li>
<li><a href="https://rasa.com/" rel="nofollow">Rasa</a></li>
<li><a href="https://rasa.com/docs/rasa-pro/calm/" rel="nofollow">CALM, an LLM-native approach to building reliable conversational AI</a></li>
<li><a href="https://arxiv.org/abs/2402.12234" rel="nofollow">Task-Oriented Dialogue with In-Context Learning</a></li>
<li><a href="https://medium.com/rasa-blog/we-don-t-know-how-to-build-conversational-software-yet-a18301db0e4b" rel="nofollow">&#39;We don’t know how to build conversational software yet&#39; by Alan Nicol</a></li>
<li><a href="https://twitter.com/vanishingdata" rel="nofollow">Vanishing Gradients on Twitter</a></li>
<li><a href="https://twitter.com/hugobowne" rel="nofollow">Hugo on Twitter</a></li>
</ul>

<p><strong>Upcoming Livestreams</strong></p>

<ul>
<li><a href="https://lu.ma/e8huz3s6?utm_source=vgan" rel="nofollow">Lessons from a Year of Building with LLMs</a></li>
<li><a href="https://lu.ma/zz3qic45?utm_source=vgan" rel="nofollow">VALIDATING THE VALIDATORS with Shreya Shanker</a></li>
</ul>]]>
  </itunes:summary>
</item>
<item>
  <title>Episode 27: How to Build Terrible AI Systems</title>
  <link>https://vanishinggradients.fireside.fm/27</link>
  <guid isPermaLink="false">d42a2479-a220-4f72-bf48-946c4a393efa</guid>
  <pubDate>Fri, 31 May 2024 10:00:00 +1000</pubDate>
  <author>Hugo Bowne-Anderson</author>
  <enclosure url="https://aphid.fireside.fm/d/1437767933/140c3904-8258-4c39-a698-a112b7077bd7/d42a2479-a220-4f72-bf48-946c4a393efa.mp3" length="88718026" type="audio/mpeg"/>
  <itunes:episodeType>full</itunes:episodeType>
  <itunes:season>1</itunes:season>
  <itunes:author>Hugo Bowne-Anderson</itunes:author>
  <itunes:subtitle>Hugo speaks with Jason Liu, an independent consultant who uses his expertise in recommendation systems to help fast-growing startups build out their RAG applications. He was previously at Meta and Stitch Fix is also the creator of Instructor, Flight, and an ML and data science educator.</itunes:subtitle>
  <itunes:duration>1:32:24</itunes:duration>
  <itunes:explicit>no</itunes:explicit>
  <itunes:image href="https://media24.fireside.fm/file/fireside-images-2024/podcasts/images/1/140c3904-8258-4c39-a698-a112b7077bd7/cover.jpg?v=1"/>
  <description>Hugo speaks with Jason Liu, an independent consultant who uses his expertise in recommendation systems to help fast-growing startups build out their RAG applications. He was previously at Meta and Stitch Fix is also the creator of Instructor, Flight, and an ML and data science educator.
They talk about how Jason approaches consulting companies across many industries, including construction and sales, in building production LLM apps, his playbook for getting ML and AI up and running to build and maintain such apps, and the future of tooling to do so.
They take an inverted thinking approach, envisaging all the failure modes that would result in building terrible AI systems, and then figure out how to avoid such pitfalls.
LINKS
The livestream on YouTube (https://youtube.com/live/USTG6sQlB6s?feature=share)
Jason's website (https://jxnl.co/)
PyDdantic is all you need, Jason's Keynote at AI Engineer Summit, 2023 (https://youtu.be/yj-wSRJwrrc?si=JIGhN0mx0i50dUR9)
How to build a terrible RAG system by Jason (https://jxnl.co/writing/2024/01/07/inverted-thinking-rag/)
To express interest in Jason's Systematically improving RAG Applications course (https://q7gjsgfstrp.typeform.com/ragcourse?typeform-source=vg)
Vanishing Gradients on Twitter (https://twitter.com/vanishingdata)
Hugo on Twitter (https://twitter.com/hugobowne)
Upcoming Livestreams
Good Riddance to Supervised Learning with Alan Nichol (CTO and co-founder, Rasa) (https://lu.ma/gphzzyyn?utm_source=vgj)
Lessons from a Year of Building with LLMs (https://lu.ma/e8huz3s6?utm_source=vgj)
</description>
  <itunes:keywords>AI, LLMs, machine learning, data science</itunes:keywords>
  <content:encoded>
    <![CDATA[<p>Hugo speaks with Jason Liu, an independent consultant who uses his expertise in recommendation systems to help fast-growing startups build out their RAG applications. He was previously at Meta and Stitch Fix is also the creator of Instructor, Flight, and an ML and data science educator.</p>

<p>They talk about how Jason approaches consulting companies across many industries, including construction and sales, in building production LLM apps, his playbook for getting ML and AI up and running to build and maintain such apps, and the future of tooling to do so.</p>

<p>They take an inverted thinking approach, envisaging all the failure modes that would result in building terrible AI systems, and then figure out how to avoid such pitfalls.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://youtube.com/live/USTG6sQlB6s?feature=share" rel="nofollow">The livestream on YouTube</a></li>
<li><a href="https://jxnl.co/" rel="nofollow">Jason&#39;s website</a></li>
<li><a href="https://youtu.be/yj-wSRJwrrc?si=JIGhN0mx0i50dUR9" rel="nofollow">PyDdantic is all you need, Jason&#39;s Keynote at AI Engineer Summit, 2023</a></li>
<li><a href="https://jxnl.co/writing/2024/01/07/inverted-thinking-rag/" rel="nofollow">How to build a terrible RAG system by Jason</a></li>
<li><a href="https://q7gjsgfstrp.typeform.com/ragcourse?typeform-source=vg" rel="nofollow">To express interest in Jason&#39;s <em>Systematically improving RAG Applications</em> course</a></li>
<li><a href="https://twitter.com/vanishingdata" rel="nofollow">Vanishing Gradients on Twitter</a></li>
<li><a href="https://twitter.com/hugobowne" rel="nofollow">Hugo on Twitter</a></li>
</ul>

<p><strong>Upcoming Livestreams</strong></p>

<ul>
<li><a href="https://lu.ma/gphzzyyn?utm_source=vgj" rel="nofollow">Good Riddance to Supervised Learning with Alan Nichol (CTO and co-founder, Rasa)</a></li>
<li><a href="https://lu.ma/e8huz3s6?utm_source=vgj" rel="nofollow">Lessons from a Year of Building with LLMs</a></li>
</ul>]]>
  </content:encoded>
  <itunes:summary>
    <![CDATA[<p>Hugo speaks with Jason Liu, an independent consultant who uses his expertise in recommendation systems to help fast-growing startups build out their RAG applications. He was previously at Meta and Stitch Fix is also the creator of Instructor, Flight, and an ML and data science educator.</p>

<p>They talk about how Jason approaches consulting companies across many industries, including construction and sales, in building production LLM apps, his playbook for getting ML and AI up and running to build and maintain such apps, and the future of tooling to do so.</p>

<p>They take an inverted thinking approach, envisaging all the failure modes that would result in building terrible AI systems, and then figure out how to avoid such pitfalls.</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://youtube.com/live/USTG6sQlB6s?feature=share" rel="nofollow">The livestream on YouTube</a></li>
<li><a href="https://jxnl.co/" rel="nofollow">Jason&#39;s website</a></li>
<li><a href="https://youtu.be/yj-wSRJwrrc?si=JIGhN0mx0i50dUR9" rel="nofollow">PyDdantic is all you need, Jason&#39;s Keynote at AI Engineer Summit, 2023</a></li>
<li><a href="https://jxnl.co/writing/2024/01/07/inverted-thinking-rag/" rel="nofollow">How to build a terrible RAG system by Jason</a></li>
<li><a href="https://q7gjsgfstrp.typeform.com/ragcourse?typeform-source=vg" rel="nofollow">To express interest in Jason&#39;s <em>Systematically improving RAG Applications</em> course</a></li>
<li><a href="https://twitter.com/vanishingdata" rel="nofollow">Vanishing Gradients on Twitter</a></li>
<li><a href="https://twitter.com/hugobowne" rel="nofollow">Hugo on Twitter</a></li>
</ul>

<p><strong>Upcoming Livestreams</strong></p>

<ul>
<li><a href="https://lu.ma/gphzzyyn?utm_source=vgj" rel="nofollow">Good Riddance to Supervised Learning with Alan Nichol (CTO and co-founder, Rasa)</a></li>
<li><a href="https://lu.ma/e8huz3s6?utm_source=vgj" rel="nofollow">Lessons from a Year of Building with LLMs</a></li>
</ul>]]>
  </itunes:summary>
</item>
<item>
  <title>Episode 26: Developing and Training LLMs From Scratch</title>
  <link>https://vanishinggradients.fireside.fm/26</link>
  <guid isPermaLink="false">d56cd02b-11cb-4be9-a2a7-31f783ef9c1a</guid>
  <pubDate>Wed, 15 May 2024 13:00:00 +1000</pubDate>
  <author>Hugo Bowne-Anderson</author>
  <enclosure url="https://aphid.fireside.fm/d/1437767933/140c3904-8258-4c39-a698-a112b7077bd7/d56cd02b-11cb-4be9-a2a7-31f783ef9c1a.mp3" length="53564523" type="audio/mpeg"/>
  <itunes:episodeType>full</itunes:episodeType>
  <itunes:season>1</itunes:season>
  <itunes:author>Hugo Bowne-Anderson</itunes:author>
  <itunes:subtitle>Hugo speaks with Sebastian Raschka, a machine learning &amp; AI researcher, programmer, and author.They’ll tell you everything you need to know about LLMs, but were too afraid to ask: from covering the entire LLM lifecycle, what type of skills you need to work with them, what type of resources and hardware, prompt engineering vs fine-tuning vs RAG, how to build an LLM from scratch, and much more.</itunes:subtitle>
  <itunes:duration>1:51:35</itunes:duration>
  <itunes:explicit>no</itunes:explicit>
  <itunes:image href="https://media24.fireside.fm/file/fireside-images-2024/podcasts/images/1/140c3904-8258-4c39-a698-a112b7077bd7/cover.jpg?v=1"/>
  <description>Hugo speaks with Sebastian Raschka, a machine learning &amp;amp; AI researcher, programmer, and author. As Staff Research Engineer at Lightning AI, he focuses on the intersection of AI research, software development, and large language models (LLMs).
How do you build LLMs? How can you use them, both in prototype and production settings? What are the building blocks you need to know about?
​In this episode, we’ll tell you everything you need to know about LLMs, but were too afraid to ask: from covering the entire LLM lifecycle, what type of skills you need to work with them, what type of resources and hardware, prompt engineering vs fine-tuning vs RAG, how to build an LLM from scratch, and much more.
The idea here is not that you’ll need to use an LLM you’ve built from scratch, but that we’ll learn a lot about LLMs and how to use them in the process.
Near the end we also did some live coding to fine-tune GPT-2 in order to create a spam classifier! 
LINKS
The livestream on YouTube (https://youtube.com/live/qL4JY6Y5pmA)
Sebastian's website (https://sebastianraschka.com/)
Machine Learning Q and AI: 30 Essential Questions and Answers on Machine Learning and AI by Sebastian (https://nostarch.com/machine-learning-q-and-ai)
Build a Large Language Model (From Scratch) by Sebastian (https://www.manning.com/books/build-a-large-language-model-from-scratch)
PyTorch Lightning (https://lightning.ai/docs/pytorch/stable/)
Lightning Fabric (https://lightning.ai/docs/fabric/stable/)
LitGPT (https://github.com/Lightning-AI/litgpt)
Sebastian's notebook for finetuning GPT-2 for spam classification! (https://github.com/rasbt/LLMs-from-scratch/blob/main/ch06/01_main-chapter-code/ch06.ipynb)
The end of fine-tuning: Jeremy Howard on the Latent Space Podcast (https://www.latent.space/p/fastai)
Our next livestream: How to Build Terrible AI Systems with Jason Liu (https://lu.ma/terrible-ai-systems?utm_source=vg)
Vanishing Gradients on Twitter (https://twitter.com/vanishingdata)
Hugo on Twitter (https://twitter.com/hugobowne) 
</description>
  <itunes:keywords>AI, LLMs, machine learning, data science, OpenAI</itunes:keywords>
  <content:encoded>
    <![CDATA[<p>Hugo speaks with Sebastian Raschka, a machine learning &amp; AI researcher, programmer, and author. As Staff Research Engineer at Lightning AI, he focuses on the intersection of AI research, software development, and large language models (LLMs).</p>

<p>How do you build LLMs? How can you use them, both in prototype and production settings? What are the building blocks you need to know about?</p>

<p>​In this episode, we’ll tell you everything you need to know about LLMs, but were too afraid to ask: from covering the entire LLM lifecycle, what type of skills you need to work with them, what type of resources and hardware, prompt engineering vs fine-tuning vs RAG, how to build an LLM from scratch, and much more.</p>

<p>The idea here is not that you’ll need to use an LLM you’ve built from scratch, but that we’ll learn a lot about LLMs and how to use them in the process.</p>

<p>Near the end we also did some live coding to fine-tune GPT-2 in order to create a spam classifier! </p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://youtube.com/live/qL4JY6Y5pmA" rel="nofollow">The livestream on YouTube</a></li>
<li><a href="https://sebastianraschka.com/" rel="nofollow">Sebastian&#39;s website</a></li>
<li><a href="https://nostarch.com/machine-learning-q-and-ai" rel="nofollow">Machine Learning Q and AI: 30 Essential Questions and Answers on Machine Learning and AI by Sebastian</a></li>
<li><a href="https://www.manning.com/books/build-a-large-language-model-from-scratch" rel="nofollow">Build a Large Language Model (From Scratch) by Sebastian</a></li>
<li><a href="https://lightning.ai/docs/pytorch/stable/" rel="nofollow">PyTorch Lightning</a></li>
<li><a href="https://lightning.ai/docs/fabric/stable/" rel="nofollow">Lightning Fabric</a></li>
<li><a href="https://github.com/Lightning-AI/litgpt" rel="nofollow">LitGPT</a></li>
<li><a href="https://github.com/rasbt/LLMs-from-scratch/blob/main/ch06/01_main-chapter-code/ch06.ipynb" rel="nofollow">Sebastian&#39;s notebook for finetuning GPT-2 for spam classification!</a></li>
<li><a href="https://www.latent.space/p/fastai" rel="nofollow">The end of fine-tuning: Jeremy Howard on the Latent Space Podcast</a></li>
<li><a href="https://lu.ma/terrible-ai-systems?utm_source=vg" rel="nofollow">Our next livestream: How to Build Terrible AI Systems with Jason Liu</a></li>
<li><a href="https://twitter.com/vanishingdata" rel="nofollow">Vanishing Gradients on Twitter</a></li>
<li><a href="https://twitter.com/hugobowne" rel="nofollow">Hugo on Twitter</a></li>
</ul>]]>
  </content:encoded>
  <itunes:summary>
    <![CDATA[<p>Hugo speaks with Sebastian Raschka, a machine learning &amp; AI researcher, programmer, and author. As Staff Research Engineer at Lightning AI, he focuses on the intersection of AI research, software development, and large language models (LLMs).</p>

<p>How do you build LLMs? How can you use them, both in prototype and production settings? What are the building blocks you need to know about?</p>

<p>​In this episode, we’ll tell you everything you need to know about LLMs, but were too afraid to ask: from covering the entire LLM lifecycle, what type of skills you need to work with them, what type of resources and hardware, prompt engineering vs fine-tuning vs RAG, how to build an LLM from scratch, and much more.</p>

<p>The idea here is not that you’ll need to use an LLM you’ve built from scratch, but that we’ll learn a lot about LLMs and how to use them in the process.</p>

<p>Near the end we also did some live coding to fine-tune GPT-2 in order to create a spam classifier! </p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://youtube.com/live/qL4JY6Y5pmA" rel="nofollow">The livestream on YouTube</a></li>
<li><a href="https://sebastianraschka.com/" rel="nofollow">Sebastian&#39;s website</a></li>
<li><a href="https://nostarch.com/machine-learning-q-and-ai" rel="nofollow">Machine Learning Q and AI: 30 Essential Questions and Answers on Machine Learning and AI by Sebastian</a></li>
<li><a href="https://www.manning.com/books/build-a-large-language-model-from-scratch" rel="nofollow">Build a Large Language Model (From Scratch) by Sebastian</a></li>
<li><a href="https://lightning.ai/docs/pytorch/stable/" rel="nofollow">PyTorch Lightning</a></li>
<li><a href="https://lightning.ai/docs/fabric/stable/" rel="nofollow">Lightning Fabric</a></li>
<li><a href="https://github.com/Lightning-AI/litgpt" rel="nofollow">LitGPT</a></li>
<li><a href="https://github.com/rasbt/LLMs-from-scratch/blob/main/ch06/01_main-chapter-code/ch06.ipynb" rel="nofollow">Sebastian&#39;s notebook for finetuning GPT-2 for spam classification!</a></li>
<li><a href="https://www.latent.space/p/fastai" rel="nofollow">The end of fine-tuning: Jeremy Howard on the Latent Space Podcast</a></li>
<li><a href="https://lu.ma/terrible-ai-systems?utm_source=vg" rel="nofollow">Our next livestream: How to Build Terrible AI Systems with Jason Liu</a></li>
<li><a href="https://twitter.com/vanishingdata" rel="nofollow">Vanishing Gradients on Twitter</a></li>
<li><a href="https://twitter.com/hugobowne" rel="nofollow">Hugo on Twitter</a></li>
</ul>]]>
  </itunes:summary>
</item>
<item>
  <title>Episode 25: Fully Reproducible ML &amp; AI Workflows</title>
  <link>https://vanishinggradients.fireside.fm/25</link>
  <guid isPermaLink="false">2e66472b-34f3-4068-b6f9-4942dc757325</guid>
  <pubDate>Mon, 18 Mar 2024 23:00:00 +1100</pubDate>
  <author>Hugo Bowne-Anderson</author>
  <enclosure url="https://aphid.fireside.fm/d/1437767933/140c3904-8258-4c39-a698-a112b7077bd7/2e66472b-34f3-4068-b6f9-4942dc757325.mp3" length="77423933" type="audio/mpeg"/>
  <itunes:episodeType>full</itunes:episodeType>
  <itunes:season>1</itunes:season>
  <itunes:author>Hugo Bowne-Anderson</itunes:author>
  <itunes:subtitle>Hugo speaks with Omoju Miller, a machine learning guru and founder and CEO of Fimio, where she is building 21st century dev tooling.</itunes:subtitle>
  <itunes:duration>1:20:38</itunes:duration>
  <itunes:explicit>no</itunes:explicit>
  <itunes:image href="https://media24.fireside.fm/file/fireside-images-2024/podcasts/images/1/140c3904-8258-4c39-a698-a112b7077bd7/cover.jpg?v=1"/>
  <description>Hugo speaks with Omoju Miller, a machine learning guru and founder and CEO of Fimio, where she is building 21st century dev tooling. In the past, she was Technical Advisor to the CEO at GitHub, spent time co-leading non-profit investment in Computer Science Education for Google, and served as a volunteer advisor to the Obama administration’s White House Presidential Innovation Fellows.
We need open tools, open data, provenance, and the ability to build fully reproducible, transparent machine learning workflows. With the advent of closed-source, vendor-based APIs and compute becoming a form of gate-keeping, developer tools are at the risk of becoming commoditized and developers becoming consumers.
We’ll  talk about how ideas for escaping these burgeoning walled gardens. We’ll dive into
What fully reproducible ML workflows would look like, including git for the workflow build process,
The need for loosely coupled and composable tools that embrace a UNIX-like philosophy,
What a much more scientific toolchain would look like,
What a future open sources commons for Generative AI could look like,
What an open compute ecosystem could look like,
How to create LLMs and tooling so everyone can use them to build production-ready apps,
And much more!
LINKS
The livestream on YouTube (https://www.youtube.com/live/n81PWNsHSMk?si=pgX2hH5xADATdJMu)
Omoju on Twitter (https://twitter.com/omojumiller)
Hugo on Twitter (https://twitter.com/hugobowne)
Vanishing Gradients on Twitter (https://twitter.com/vanishingdata)
Lu.ma Calendar that includes details of Hugo's European Tour for Outerbounds (https://lu.ma/Outerbounds)
Blog post that includes details of Hugo's European Tour for Outerbounds (https://outerbounds.com/blog/ob-on-the-road-2024-h1/)
</description>
  <itunes:keywords>AI, LLMs, machine learning, data science, OpenAI</itunes:keywords>
  <content:encoded>
    <![CDATA[<p>Hugo speaks with Omoju Miller, a machine learning guru and founder and CEO of Fimio, where she is building 21st century dev tooling. In the past, she was Technical Advisor to the CEO at GitHub, spent time co-leading non-profit investment in Computer Science Education for Google, and served as a volunteer advisor to the Obama administration’s White House Presidential Innovation Fellows.</p>

<p>We need open tools, open data, provenance, and the ability to build fully reproducible, transparent machine learning workflows. With the advent of closed-source, vendor-based APIs and compute becoming a form of gate-keeping, developer tools are at the risk of becoming commoditized and developers becoming consumers.</p>

<p>We’ll  talk about how ideas for escaping these burgeoning walled gardens. We’ll dive into</p>

<ul>
<li>What fully reproducible ML workflows would look like, including git for the workflow build process,</li>
<li>The need for loosely coupled and composable tools that embrace a UNIX-like philosophy,</li>
<li>What a much more scientific toolchain would look like,</li>
<li>What a future open sources commons for Generative AI could look like,</li>
<li>What an open compute ecosystem could look like,</li>
<li>How to create LLMs and tooling so everyone can use them to build production-ready apps,</li>
</ul>

<p>And much more!</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://www.youtube.com/live/n81PWNsHSMk?si=pgX2hH5xADATdJMu" rel="nofollow">The livestream on YouTube</a></li>
<li><a href="https://twitter.com/omojumiller" rel="nofollow">Omoju on Twitter</a></li>
<li><a href="https://twitter.com/hugobowne" rel="nofollow">Hugo on Twitter</a></li>
<li><a href="https://twitter.com/vanishingdata" rel="nofollow">Vanishing Gradients on Twitter</a></li>
<li><a href="https://lu.ma/Outerbounds" rel="nofollow">Lu.ma Calendar that includes details of Hugo&#39;s European Tour for Outerbounds</a></li>
<li><a href="https://outerbounds.com/blog/ob-on-the-road-2024-h1/" rel="nofollow">Blog post that includes details of Hugo&#39;s European Tour for Outerbounds</a></li>
</ul>]]>
  </content:encoded>
  <itunes:summary>
    <![CDATA[<p>Hugo speaks with Omoju Miller, a machine learning guru and founder and CEO of Fimio, where she is building 21st century dev tooling. In the past, she was Technical Advisor to the CEO at GitHub, spent time co-leading non-profit investment in Computer Science Education for Google, and served as a volunteer advisor to the Obama administration’s White House Presidential Innovation Fellows.</p>

<p>We need open tools, open data, provenance, and the ability to build fully reproducible, transparent machine learning workflows. With the advent of closed-source, vendor-based APIs and compute becoming a form of gate-keeping, developer tools are at the risk of becoming commoditized and developers becoming consumers.</p>

<p>We’ll  talk about how ideas for escaping these burgeoning walled gardens. We’ll dive into</p>

<ul>
<li>What fully reproducible ML workflows would look like, including git for the workflow build process,</li>
<li>The need for loosely coupled and composable tools that embrace a UNIX-like philosophy,</li>
<li>What a much more scientific toolchain would look like,</li>
<li>What a future open sources commons for Generative AI could look like,</li>
<li>What an open compute ecosystem could look like,</li>
<li>How to create LLMs and tooling so everyone can use them to build production-ready apps,</li>
</ul>

<p>And much more!</p>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://www.youtube.com/live/n81PWNsHSMk?si=pgX2hH5xADATdJMu" rel="nofollow">The livestream on YouTube</a></li>
<li><a href="https://twitter.com/omojumiller" rel="nofollow">Omoju on Twitter</a></li>
<li><a href="https://twitter.com/hugobowne" rel="nofollow">Hugo on Twitter</a></li>
<li><a href="https://twitter.com/vanishingdata" rel="nofollow">Vanishing Gradients on Twitter</a></li>
<li><a href="https://lu.ma/Outerbounds" rel="nofollow">Lu.ma Calendar that includes details of Hugo&#39;s European Tour for Outerbounds</a></li>
<li><a href="https://outerbounds.com/blog/ob-on-the-road-2024-h1/" rel="nofollow">Blog post that includes details of Hugo&#39;s European Tour for Outerbounds</a></li>
</ul>]]>
  </itunes:summary>
</item>
<item>
  <title>Episode 24: LLM and GenAI Accessibility</title>
  <link>https://vanishinggradients.fireside.fm/24</link>
  <guid isPermaLink="false">c6ebf900-c625-493a-b4c5-27a7f31da24f</guid>
  <pubDate>Tue, 27 Feb 2024 17:00:00 +1100</pubDate>
  <author>Hugo Bowne-Anderson</author>
  <enclosure url="https://aphid.fireside.fm/d/1437767933/140c3904-8258-4c39-a698-a112b7077bd7/c6ebf900-c625-493a-b4c5-27a7f31da24f.mp3" length="86459792" type="audio/mpeg"/>
  <itunes:episodeType>full</itunes:episodeType>
  <itunes:season>1</itunes:season>
  <itunes:author>Hugo Bowne-Anderson</itunes:author>
  <itunes:subtitle>Hugo speaks with Johno Whitaker, a Data Scientist/AI Researcher doing R&amp;D with answer.ai, about where we’ve come from regarding tooling and accessibility for foundation models, ML, and AI, where we are, and where we’re going.</itunes:subtitle>
  <itunes:duration>1:30:03</itunes:duration>
  <itunes:explicit>no</itunes:explicit>
  <itunes:image href="https://media24.fireside.fm/file/fireside-images-2024/podcasts/images/1/140c3904-8258-4c39-a698-a112b7077bd7/cover.jpg?v=1"/>
  <description>Hugo speaks with Johno Whitaker, a Data Scientist/AI Researcher doing R&amp;amp;D with answer.ai. His current focus is on generative AI, flitting between different modalities. He also likes teaching and making courses, having worked with both Hugging Face and fast.ai in these capacities.
Johno recently reminded Hugo how hard everything was 10 years ago: “Want to install TensorFlow? Good luck. Need data? Perhaps try ImageNet. But now you can use big models from Hugging Face with hi-res satellite data and do all of this in a Colab notebook. Or think ecology and vision models… or medicine and multimodal models!”
We  talk about where we’ve come from regarding tooling and accessibility for foundation models, ML, and AI, where we are, and where we’re going. We’ll delve into
What the Generative AI mindset is, in terms of using atomic building blocks, and how it evolved from both the data science and ML mindsets;
How fast.ai democratized access to deep learning, what successes they had, and what was learned;
The moving parts now required to make GenAI and ML as accessible as possible;
The importance of focusing on UX and the application in the world of generative AI and foundation models;
The skillset and toolkit needed to be an LLM and AI guru;
What they’re up to at answer.ai to democratize LLMs and foundation models.
LINKS
The livestream on YouTube (https://youtube.com/live/hxZX6fBi-W8?feature=share)
Zindi, the largest professional network for data scientists in Africa (https://zindi.africa/)
A new old kind of R&amp;amp;D lab: Announcing Answer.AI (http://www.answer.ai/posts/2023-12-12-launch.html)
Why and how I’m shifting focus to LLMs by Johno Whitaker (https://johnowhitaker.dev/dsc/2023-07-01-why-and-how-im-shifting-focus-to-llms.html)
Applying AI to Immune Cell Networks by Rachel Thomas (https://www.fast.ai/posts/2024-01-23-cytokines/)
Replicate -- a cool place to explore GenAI models, among other things (https://replicate.com/explore)
Hands-On Generative AI with Transformers and Diffusion Models (https://www.oreilly.com/library/view/hands-on-generative-ai/9781098149239/)
Johno on Twitter (https://twitter.com/johnowhitaker)
Hugo on Twitter (https://twitter.com/hugobowne)
Vanishing Gradients on Twitter (https://twitter.com/vanishingdata)
SciPy 2024 CFP (https://www.scipy2024.scipy.org/#CFP)
Escaping Generative AI Walled Gardens with Omoju Miller, a Vanishing Gradients Livestream (https://lu.ma/xonnjqe4)
</description>
  <itunes:keywords>genAI, machine learning, data science</itunes:keywords>
  <content:encoded>
    <![CDATA[<p>Hugo speaks with Johno Whitaker, a Data Scientist/AI Researcher doing R&amp;D with answer.ai. His current focus is on generative AI, flitting between different modalities. He also likes teaching and making courses, having worked with both Hugging Face and fast.ai in these capacities.</p>

<p>Johno recently reminded Hugo how hard everything was 10 years ago: “Want to install TensorFlow? Good luck. Need data? Perhaps try ImageNet. But now you can use big models from Hugging Face with hi-res satellite data and do all of this in a Colab notebook. Or think ecology and vision models… or medicine and multimodal models!”</p>

<p>We  talk about where we’ve come from regarding tooling and accessibility for foundation models, ML, and AI, where we are, and where we’re going. We’ll delve into</p>

<ul>
<li>What the Generative AI mindset is, in terms of using atomic building blocks, and how it evolved from both the data science and ML mindsets;</li>
<li>How fast.ai democratized access to deep learning, what successes they had, and what was learned;</li>
<li>The moving parts now required to make GenAI and ML as accessible as possible;</li>
<li>The importance of focusing on UX and the application in the world of generative AI and foundation models;</li>
<li>The skillset and toolkit needed to be an LLM and AI guru;</li>
<li>What they’re up to at answer.ai to democratize LLMs and foundation models.</li>
</ul>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://youtube.com/live/hxZX6fBi-W8?feature=share" rel="nofollow">The livestream on YouTube</a></li>
<li><a href="https://zindi.africa/" rel="nofollow">Zindi, the largest professional network for data scientists in Africa</a></li>
<li><a href="http://www.answer.ai/posts/2023-12-12-launch.html" rel="nofollow">A new old kind of R&amp;D lab: Announcing Answer.AI</a></li>
<li><a href="https://johnowhitaker.dev/dsc/2023-07-01-why-and-how-im-shifting-focus-to-llms.html" rel="nofollow">Why and how I’m shifting focus to LLMs by Johno Whitaker</a></li>
<li><a href="https://www.fast.ai/posts/2024-01-23-cytokines/" rel="nofollow">Applying AI to Immune Cell Networks by Rachel Thomas</a></li>
<li><a href="https://replicate.com/explore" rel="nofollow">Replicate -- a cool place to explore GenAI models, among other things</a></li>
<li><a href="https://www.oreilly.com/library/view/hands-on-generative-ai/9781098149239/" rel="nofollow">Hands-On Generative AI with Transformers and Diffusion Models</a></li>
<li><a href="https://twitter.com/johnowhitaker" rel="nofollow">Johno on Twitter</a></li>
<li><a href="https://twitter.com/hugobowne" rel="nofollow">Hugo on Twitter</a></li>
<li><a href="https://twitter.com/vanishingdata" rel="nofollow">Vanishing Gradients on Twitter</a></li>
<li><a href="https://www.scipy2024.scipy.org/#CFP" rel="nofollow">SciPy 2024 CFP</a></li>
<li><a href="https://lu.ma/xonnjqe4" rel="nofollow">Escaping Generative AI Walled Gardens with Omoju Miller, a Vanishing Gradients Livestream</a></li>
</ul>]]>
  </content:encoded>
  <itunes:summary>
    <![CDATA[<p>Hugo speaks with Johno Whitaker, a Data Scientist/AI Researcher doing R&amp;D with answer.ai. His current focus is on generative AI, flitting between different modalities. He also likes teaching and making courses, having worked with both Hugging Face and fast.ai in these capacities.</p>

<p>Johno recently reminded Hugo how hard everything was 10 years ago: “Want to install TensorFlow? Good luck. Need data? Perhaps try ImageNet. But now you can use big models from Hugging Face with hi-res satellite data and do all of this in a Colab notebook. Or think ecology and vision models… or medicine and multimodal models!”</p>

<p>We  talk about where we’ve come from regarding tooling and accessibility for foundation models, ML, and AI, where we are, and where we’re going. We’ll delve into</p>

<ul>
<li>What the Generative AI mindset is, in terms of using atomic building blocks, and how it evolved from both the data science and ML mindsets;</li>
<li>How fast.ai democratized access to deep learning, what successes they had, and what was learned;</li>
<li>The moving parts now required to make GenAI and ML as accessible as possible;</li>
<li>The importance of focusing on UX and the application in the world of generative AI and foundation models;</li>
<li>The skillset and toolkit needed to be an LLM and AI guru;</li>
<li>What they’re up to at answer.ai to democratize LLMs and foundation models.</li>
</ul>

<p><strong>LINKS</strong></p>

<ul>
<li><a href="https://youtube.com/live/hxZX6fBi-W8?feature=share" rel="nofollow">The livestream on YouTube</a></li>
<li><a href="https://zindi.africa/" rel="nofollow">Zindi, the largest professional network for data scientists in Africa</a></li>
<li><a href="http://www.answer.ai/posts/2023-12-12-launch.html" rel="nofollow">A new old kind of R&amp;D lab: Announcing Answer.AI</a></li>
<li><a href="https://johnowhitaker.dev/dsc/2023-07-01-why-and-how-im-shifting-focus-to-llms.html" rel="nofollow">Why and how I’m shifting focus to LLMs by Johno Whitaker</a></li>
<li><a href="https://www.fast.ai/posts/2024-01-23-cytokines/" rel="nofollow">Applying AI to Immune Cell Networks by Rachel Thomas</a></li>
<li><a href="https://replicate.com/explore" rel="nofollow">Replicate -- a cool place to explore GenAI models, among other things</a></li>
<li><a href="https://www.oreilly.com/library/view/hands-on-generative-ai/9781098149239/" rel="nofollow">Hands-On Generative AI with Transformers and Diffusion Models</a></li>
<li><a href="https://twitter.com/johnowhitaker" rel="nofollow">Johno on Twitter</a></li>
<li><a href="https://twitter.com/hugobowne" rel="nofollow">Hugo on Twitter</a></li>
<li><a href="https://twitter.com/vanishingdata" rel="nofollow">Vanishing Gradients on Twitter</a></li>
<li><a href="https://www.scipy2024.scipy.org/#CFP" rel="nofollow">SciPy 2024 CFP</a></li>
<li><a href="https://lu.ma/xonnjqe4" rel="nofollow">Escaping Generative AI Walled Gardens with Omoju Miller, a Vanishing Gradients Livestream</a></li>
</ul>]]>
  </itunes:summary>
</item>
  </channel>
</rss>
