Kevin Dubois

Infusing Generative AI in your Java Apps with LangChain4j

Turn natural language commands into executable Java code. Learn how the `@Tool` annotation in LangChain4j connects LLM prompts directly to your business logic.

Infusing Generative AI in your Java Apps with LangChain4j
#1about 3 minutes

Integrating generative AI into Java applications with LangChain4j

LangChain4j simplifies consuming AI model APIs for Java developers, avoiding the need for deep data science expertise.

#2about 3 minutes

Creating a new Quarkus project with LangChain4j

Use the Quarkus CLI to bootstrap a new Java application and add the necessary LangChain4j dependency for OpenAI integration.

#3about 2 minutes

Using prompts and AI services in LangChain4j

Define AI interactions using the @RegisterAIService annotation, system messages for context, and user messages with dynamic placeholders.

#4about 2 minutes

Managing conversational context with memory

LangChain4j uses memory to retain context across multiple calls, with the @MemoryId annotation enabling parallel conversations.

#5about 2 minutes

Connecting AI models to business logic with tools

Use the @Tool annotation to expose Java methods to the AI model, allowing it to execute business logic like sending an email.

#6about 5 minutes

Live demo of prompts, tools, and the Dev UI

A practical demonstration shows how to generate a haiku using a prompt and then use a custom tool to send it via email, verified with Mailpit.

#7about 3 minutes

Providing custom knowledge with retrieval-augmented generation (RAG)

Enhance LLM responses with your own business data by using an embedding store or Quarkus's simplified 'Easy RAG' feature.

#8about 6 minutes

Building a chatbot with a custom knowledge base

A chatbot demo uses a terms of service document via RAG to correctly enforce a business rule for booking cancellations.

#9about 2 minutes

Using local models and implementing fault tolerance

Run LLMs on your local machine with Podman AI Lab and make your application resilient to failures using SmallRye Fault Tolerance annotations.

#10about 4 minutes

Demonstrating fault tolerance with a local LLM

A final demo shows an application calling a locally-run model and triggering a fallback mechanism when the model service is unavailable.

Related jobs
Jobs that call for the skills explored in this talk.

test

Milly
Vienna, Austria

Intermediate

test

Milly
Vienna, Austria

Intermediate

d

Saby Company
Delebio, Italy

Junior

Featured Partners

Related Articles

View all articles
DC
Daniel Cranney
Stephan Gillich - Bringing AI Everywhere
In the ever-evolving world of technology, AI continues to be the frontier for innovation and transformation. Stephan Gillich, from the AI Center of Excellence at Intel, dove into the subject in a recent session titled "Bringing AI Everywhere," sheddi...
Stephan Gillich - Bringing AI Everywhere
CH
Chris Heilmann
Exploring AI: Opportunities and Risks for Developers
In today's rapidly evolving tech landscape, the integration of Artificial Intelligence (AI) in development presents both exciting opportunities and notable risks. This dynamic was the focus of a recent panel discussion featuring industry experts Kent...
Exploring AI: Opportunities and Risks for Developers
DC
Daniel Cranney
How to Use Generative AI to Accelerate Learning to Code
It’s undeniable that generative-AI and LLMs have transformed how developers work. Hours of hunting Stack Overflow can be avoided by asking your AI-code assistant, multi-file context can be fed to the AI from inside your IDE, and applications can be b...
How to Use Generative AI to Accelerate Learning to Code

From learning to earning

Jobs that call for the skills explored in this talk.