Archived series ("Inactive feed" status)
When? This feed was archived on October 30, 2025 18:17 (). Last successful fetch was on April 21, 2025 08:41 ()
Why? Inactive feed status. Our servers were unable to retrieve a valid podcast feed for a sustained period.
What now? You might be able to find a more up-to-date version using the search function. This series will no longer be checked for updates. If you believe this to be in error, please check if the publisher's feed link below is valid and contact support to request the feed be restored or if you have any other concerns about this.
Manage episode 437345387 series 3596668
Sanyam Bhutani, a leading figure in the data science community. Sanyam is a Sr. Data Scientist at H2O.ai, with previous tenures at Weights & Biases and H2O.ai, and an International Fellow at fast.ai. As a Kaggle Grandmaster, his contributions to the field are well-recognized and highly respected.
Sanyam delves into the nuances of fine-tuning and optimizing Large Language Models (LLMs). He provides a detailed exploration of the current state and future potential of LLMs, breaking down their architecture and functionality in a way that's accessible to both newcomers and seasoned data scientists. Sanyam discusses the importance of fine-tuning in enhancing the performance and applicability of LLMs, providing practical insights and strategies for effective implementation.
📲 Radek Osmulski Socials:
LinkedIn: https://www.linkedin.com/in/sanyambhutani/
Twitter: https://x.com/bhutanisanyam1?lang=en
📲 Mark Moyou, PhD Socials:
LinkedIn: https://www.linkedin.com/in/markmoyou/
Twitter: https://twitter.com/MarkMoyou
📗 Chapters
00:00 Intro
02:46 200 days of LLMs
06:16 Venture Capital
08:40 Setting Goals in Public
09:45 Fine tuning Experiment
14:02 Kaggle Grandmasters Team
15:55 Doing Challenges & Reading Research Papers
17:47 Hardest topic to learn in AI
19:05 Are you afraid to ask stupid questions?
20:43 Learning how LLMs work
22:54 Academic vs Product First Mindset
27:51 Training or Inference on LLMs
29:15 Favorite LLM Agent
32:10 How to go about learning LLMs?
36:55 Open Source LLMs on Research Papers
37:41 Capability of Modern GPUs
45:48 Journey to H20.ai
50:07 Why Sanyam stopped podcasting?
56:25 Podcasting Experience
58:39 Top Data Scientists
01:00:19 Advice for New Podcasts
01:03:32 Breaking into Data Science
01:12:23 Career Optimization Function
01:14:02 Making Progress Everyday
01:15:05 Advice for New Professionals
01:17:00 Book Recommendations
01:18:04 Rapid Round
22 episodes