LLM (Large Language Model) concept is widely used to build chatbots, question-answering systems, and other conversational applications.
In this workshop, we will discover how to interact with a LLM as we craft a conversational agent step-by-step. Throughout this journey, we will understand the capabilities and limitations of LLMs, and explore prompt engineering techniques to elevate their standard behavior.
For the practical part, we will learn about Spring AI's chat completion features to interact with LLMs. Spring AI serves as an abstraction layer able to interact with several LLM providers. For the workshop's purposes, we will use Ollama, a solution for running LLMs locally.
- GitHub Codespaces (5min)
- GitPod (5min)
- Local Hosting (15min)
Execute this command only once to pull the model.
docker exec -it ollama sh -c "ollama pull mistral:7b"
Execute this command after code update.
mvn clean install
mvn spring-boot:run
- Zero-shot prompting (15min)
- Conversational memory (15min)
- Information extraction (20min)
- Retrieval Augmented Generation (RAG) (30min)
- Conclusion
This workshop is proposed by AXA France's Development Guild.
Maintainers: