Manage episode 521202981 series 3506872
Ilya & I discuss SSI’s strategy, the problems with pre-training, how to improve the generalization of AI models, and how to ensure AGI goes well.
Watch on YouTube; read the transcript.
Sponsors
* Gemini 3 is the first model I’ve used that can find connections I haven’t anticipated. I recently wrote a blog post on RL’s information efficiency, and Gemini 3 helped me think it all through. It also generated the relevant charts and ran toy ML experiments for me with zero bugs. Try Gemini 3 today at gemini.google
* Labelbox helped me create a tool to transcribe our episodes! I’ve struggled with transcription in the past because I don’t just want verbatim transcripts, I want transcripts reworded to read like essays. Labelbox helped me generate the exact data I needed for this. If you want to learn how Labelbox can help you (or if you want to try out the transcriber tool yourself), go to labelbox.com/dwarkesh
* Sardine is an AI risk management platform that brings together thousands of device, behavior, and identity signals to help you assess a user’s risk of fraud & abuse. Sardine also offers a suite of agents to automate investigations so that as fraudsters use AI to scale their attacks, you can use AI to scale your defenses. Learn more at sardine.ai/dwarkesh
To sponsor a future episode, visit dwarkesh.com/advertise.
Timestamps
(00:00:00) – Explaining model jaggedness
(00:09:39) - Emotions and value functions
(00:18:49) – What are we scaling?
(00:25:13) – Why humans generalize better than models
(00:35:45) – SSI’s plan to straight-shot superintelligence
(00:46:47) – SSI’s model will learn from deployment
(00:55:07) – How to think about powerful AGIs
(01:18:13) – “We are squarely an age of research company”
(01:30:26) – Self-play and multi-agent
(01:32:42) – Research taste
Get full access to Dwarkesh Podcast at www.dwarkesh.com/subscribe
143 episodes