Search a title or topic

Over 20 million podcasts, powered by 

Player FM logo
Artwork

Content provided by Hugo Bowne-Anderson. All podcast content including episodes, graphics, and podcast descriptions are uploaded and provided directly by Hugo Bowne-Anderson or their podcast platform partner. If you believe someone is using your copyrighted work without your permission, you can follow the process outlined here https://staging.podcastplayer.com/legal.
Player FM - Podcast App
Go offline with the Player FM app!

Episode 32: Building Reliable and Robust ML/AI Pipelines

1:15:10
 
Share
 

Manage episode 430925619 series 3317544
Content provided by Hugo Bowne-Anderson. All podcast content including episodes, graphics, and podcast descriptions are uploaded and provided directly by Hugo Bowne-Anderson or their podcast platform partner. If you believe someone is using your copyrighted work without your permission, you can follow the process outlined here https://staging.podcastplayer.com/legal.

Hugo speaks with Shreya Shankar, a researcher at UC Berkeley focusing on data management systems with a human-centered approach. Shreya's work is at the cutting edge of human-computer interaction (HCI) and AI, particularly in the realm of large language models (LLMs). Her impressive background includes being the first ML engineer at Viaduct, doing research engineering at Google Brain, and software engineering at Facebook.

In this episode, we dive deep into the world of LLMs and the critical challenges of building reliable AI pipelines. We'll explore:

  • The fascinating journey from classic machine learning to the current LLM revolution
  • Why Shreya believes most ML problems are actually data management issues
  • The concept of "data flywheels" for LLM applications and how to implement them
  • The intriguing world of evaluating AI systems - who validates the validators?
  • Shreya's work on SPADE and EvalGen, innovative tools for synthesizing data quality assertions and aligning LLM evaluations with human preferences
  • The importance of human-in-the-loop processes in AI development
  • The future of low-code and no-code tools in the AI landscape

We'll also touch on the potential pitfalls of over-relying on LLMs, the concept of "Habsburg AI," and how to avoid disappearing up our own proverbial arseholes in the world of recursive AI processes.

Whether you're a seasoned AI practitioner, a curious data scientist, or someone interested in the human side of AI development, this conversation offers valuable insights into building more robust, reliable, and human-centered AI systems.

LINKS

In the podcast, Hugo also mentioned that this was the 5th time he and Shreya chatted publicly. which is wild!

If you want to dive deep into Shreya's work and related topics through their chats, you can check them all out here:

  1. Outerbounds' Fireside Chat: Operationalizing ML -- Patterns and Pain Points from MLOps Practitioners
  2. The Past, Present, and Future of Generative AI
  3. LLMs, OpenAI Dev Day, and the Existential Crisis for Machine Learning Engineering
  4. Lessons from a Year of Building with LLMs

Check out and subcribe to our lu.ma calendar for upcoming livestreams!

  continue reading

47 episodes

Artwork
iconShare
 
Manage episode 430925619 series 3317544
Content provided by Hugo Bowne-Anderson. All podcast content including episodes, graphics, and podcast descriptions are uploaded and provided directly by Hugo Bowne-Anderson or their podcast platform partner. If you believe someone is using your copyrighted work without your permission, you can follow the process outlined here https://staging.podcastplayer.com/legal.

Hugo speaks with Shreya Shankar, a researcher at UC Berkeley focusing on data management systems with a human-centered approach. Shreya's work is at the cutting edge of human-computer interaction (HCI) and AI, particularly in the realm of large language models (LLMs). Her impressive background includes being the first ML engineer at Viaduct, doing research engineering at Google Brain, and software engineering at Facebook.

In this episode, we dive deep into the world of LLMs and the critical challenges of building reliable AI pipelines. We'll explore:

  • The fascinating journey from classic machine learning to the current LLM revolution
  • Why Shreya believes most ML problems are actually data management issues
  • The concept of "data flywheels" for LLM applications and how to implement them
  • The intriguing world of evaluating AI systems - who validates the validators?
  • Shreya's work on SPADE and EvalGen, innovative tools for synthesizing data quality assertions and aligning LLM evaluations with human preferences
  • The importance of human-in-the-loop processes in AI development
  • The future of low-code and no-code tools in the AI landscape

We'll also touch on the potential pitfalls of over-relying on LLMs, the concept of "Habsburg AI," and how to avoid disappearing up our own proverbial arseholes in the world of recursive AI processes.

Whether you're a seasoned AI practitioner, a curious data scientist, or someone interested in the human side of AI development, this conversation offers valuable insights into building more robust, reliable, and human-centered AI systems.

LINKS

In the podcast, Hugo also mentioned that this was the 5th time he and Shreya chatted publicly. which is wild!

If you want to dive deep into Shreya's work and related topics through their chats, you can check them all out here:

  1. Outerbounds' Fireside Chat: Operationalizing ML -- Patterns and Pain Points from MLOps Practitioners
  2. The Past, Present, and Future of Generative AI
  3. LLMs, OpenAI Dev Day, and the Existential Crisis for Machine Learning Engineering
  4. Lessons from a Year of Building with LLMs

Check out and subcribe to our lu.ma calendar for upcoming livestreams!

  continue reading

47 episodes

All episodes

×
 
Loading …

Welcome to Player FM!

Player FM is scanning the web for high-quality podcasts for you to enjoy right now. It's the best podcast app and works on Android, iPhone, and the web. Signup to sync subscriptions across devices.

 

Listen to this show while you explore
Play