Shared Ollama LLM runtime for the GenAI Portfolio Suite. GPU-accelerated Ollama container with shared Docker network.
-
Updated
Feb 12, 2026 - Python
Shared Ollama LLM runtime for the GenAI Portfolio Suite. GPU-accelerated Ollama container with shared Docker network.
Multi-model inference API and playground powered by Ollama. Serve, switch, compare, and benchmark 6 local LLMs through FastAPI and Streamlit.
Full RAG implementation with explicit prompt assembly and operator visibility for debugging and validation.
Add a description, image, and links to the portfolio-suite topic page so that developers can more easily learn about it.
To associate your repository with the portfolio-suite topic, visit your repo's landing page and select "manage topics."