Skip to content

Exploring interactions with LLMs : Practical insights with Spring AI

License

Notifications You must be signed in to change notification settings

AxaFrance/spring-ai-workshop

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

5 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

cover picture

Exploring interactions with LLMs : Practical insights with Spring AI

LLM (Large Language Model) concept is widely used to build chatbots, question-answering systems, and other conversational applications.

In this workshop, we will discover how to interact with a LLM as we craft a conversational agent step-by-step. Throughout this journey, we will understand the capabilities and limitations of LLMs, and explore prompt engineering techniques to elevate their standard behavior.

For the practical part, we will learn about Spring AI's chat completion features to interact with LLMs. Spring AI serves as an abstraction layer able to interact with several LLM providers. For the workshop's purposes, we will use Ollama, a solution for running LLMs locally.

🔧 Environment setup

🔮 Install Mistral 7B model

Execute this command only once to pull the model.

docker exec -it ollama sh -c "ollama pull mistral:7b"

🏗️ Build the project

Execute this command after code update.

mvn clean install

🏃‍♀️ Run the application

mvn spring-boot:run

🚀 Hands-on

  1. Zero-shot prompting (15min)
  2. Conversational memory (15min)
  3. Information extraction (20min)
  4. Retrieval Augmented Generation (RAG) (30min)
  5. Conclusion

⭐ Contributors

This workshop is proposed by AXA France's Development Guild.

Maintainers:

  • Clément GIGUEL :octocat:
  • Lamine DIAKHABY :octocat: