Go offline with the Player FM app!
Quarkus and LangChain4J - A Match Made in Heaven
Fetch error
Hmmm there seems to be a problem fetching this series right now. Last successful fetch was on May 04, 2025 15:33 (
What now? This series will be checked again in the next day. If you believe it should be working, please verify the publisher's feed link below is valid and includes actual episode links. You can contact support to request the feed be immediately fetched.
Manage episode 445997956 series 2469611
discussion on integrating langchain4j with quarkus for enterprise AI applications, similarities between LLM integration and microservice architecture, benefits of using Java and MicroProfile for AI development, explanation of AI services, chat memory, and tools in LangChain4J, importance of session management and fault tolerance in LLM applications, vector databases and embeddings for efficient information retrieval, RAG (Retrieve Augmented Generation) implementation in enterprise settings, Quarkus dev mode features for LLM experimentation, native image support with GraalVM, local inference possibilities with Java 21's Vector API and quantized models, challenges in prompt engineering and model selection, upcoming features in LangChain4J including Ollama tool support and improved result streaming, future developments in Java for AI and GPU support with Project Babylon, importance of enterprise-grade features like CI/CD, testing, and cloud deployment for LLM applications
Georgios Andrianakis on twitter: @geoand86
345 episodes
Fetch error
Hmmm there seems to be a problem fetching this series right now. Last successful fetch was on May 04, 2025 15:33 (
What now? This series will be checked again in the next day. If you believe it should be working, please verify the publisher's feed link below is valid and includes actual episode links. You can contact support to request the feed be immediately fetched.
Manage episode 445997956 series 2469611
discussion on integrating langchain4j with quarkus for enterprise AI applications, similarities between LLM integration and microservice architecture, benefits of using Java and MicroProfile for AI development, explanation of AI services, chat memory, and tools in LangChain4J, importance of session management and fault tolerance in LLM applications, vector databases and embeddings for efficient information retrieval, RAG (Retrieve Augmented Generation) implementation in enterprise settings, Quarkus dev mode features for LLM experimentation, native image support with GraalVM, local inference possibilities with Java 21's Vector API and quantized models, challenges in prompt engineering and model selection, upcoming features in LangChain4J including Ollama tool support and improved result streaming, future developments in Java for AI and GPU support with Project Babylon, importance of enterprise-grade features like CI/CD, testing, and cloud deployment for LLM applications
Georgios Andrianakis on twitter: @geoand86
345 episodes
All episodes
×Welcome to Player FM!
Player FM is scanning the web for high-quality podcasts for you to enjoy right now. It's the best podcast app and works on Android, iPhone, and the web. Signup to sync subscriptions across devices.