Artwork
iconShare
 
Manage episode 523867878 series 3705596
Content provided by Keith Bourne. All podcast content including episodes, graphics, and podcast descriptions are uploaded and provided directly by Keith Bourne or their podcast platform partner. If you believe someone is using your copyrighted work without your permission, you can follow the process outlined here https://staging.podcastplayer.com/legal.

Unlock the inner workings of Retrieval-Augmented Generation (RAG) pipelines using LangChain in this episode of Memriq Inference Digest - Engineering Edition. We bring insights directly from Keith Bourne, author of 'Unlocking Data with Generative AI and RAG,' as we explore modular vector stores, retrievers, and LLM integrations critical for building scalable, flexible AI systems.

In this episode:

- Explore LangChain’s modular architecture for building RAG pipelines

- Compare popular vector stores: Chroma, FAISS, Weaviate, and Pinecone

- Understand retriever strategies: BM25, dense, and ensemble approaches

- Dive into LLM integrations like OpenAI’s ChatOpenAI and Together AI’s ChatTogether

- Discuss engineering trade-offs, GPU acceleration, and production considerations

- Highlight real-world use cases and challenges in scaling retrieval

Key tools and technologies mentioned:

- LangChain framework

- Vector stores: Chroma, FAISS, Weaviate, Pinecone

- Retrievers: BM25, Dense, Ensemble Retriever

- LLMs: OpenAI ChatOpenAI, Together AI ChatTogether

- FAISS GPU acceleration

Timestamps:

00:00 - Introduction & episode overview

02:15 - LangChain modularity and design philosophy

05:30 - Vector store comparisons and scale trade-offs

09:00 - Retriever types and ensemble approaches

12:30 - Under the hood: pipeline walkthrough

15:00 - Performance metrics and latency improvements

17:00 - Real-world applications and challenges

19:00 - Final thoughts and book spotlight

Resources:

- "Unlocking Data with Generative AI and RAG" by Keith Bourne - Search for 'Keith Bourne' on Amazon and grab the 2nd edition

- Visit Memriq.ai for AI infrastructure deep dives, practical guides, and research breakdowns

Thanks for tuning in to Memriq Inference Digest - Engineering Edition. Stay curious and keep building!

  continue reading

22 episodes