-
Notifications
You must be signed in to change notification settings - Fork 22
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Draft: Ollama API with Podman Compose #93
base: main
Are you sure you want to change the base?
Conversation
Make some changes required to use podman instead of docker
docs/ai.md
Outdated
ports: | ||
- 11434:11434 | ||
volumes: | ||
- ./ollama_v:/root/.ollama:z |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Permission denied in podman without :z
docs/ai.md
Outdated
volumes: | ||
- ./ollama_v:/root/.ollama:z | ||
devices: | ||
- nvidia.com/gpu=all |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
This might work with docker-compose as well. If so we can probably unify the 2 blocks (:z
worked with docker-compose)
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Does not work for docker
Error response from daemon: could not select device driver "cdi" with capabilities: []
We could also just put a quadlet in there and tell people to paste it into the right file? I'm thinking, if we're going podman, we should go full podman/quadlet which is what they prefer. We should also leave the docker example there too, if someone needs to get something done and they need ollama they shouldn't have to learn podman that same time, so offering both feels great. What do you think? |
@castrojo I think that makes sense. I suspect that it will still need to run as root, but I'll give it a try and report back. |
split podman from the docker section, also add a quadlet version
ContainerName=ollama | ||
AutoUpdate=yes | ||
PublishPort=11434:11434 | ||
Volume=./ollama_v:/root/.ollama:z |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Not sure what the volume path should be in the quadlet
OK so the quadlet kind of works.
|
WIP for using podman-compose instead of docker compose with the Ollama API docs.
Needs someone else to verify before merging these updates. I have only tested on my Aurora install with Nvidia GPU.
Nvidia GPU passthrough seems to need
sudo
until this bug is fixed containers/podman#19338