HTTP_PORT=8888 LLM=deepseek-coder OLLAMA_BASE_URL=http://ollama-service:11434 docker compose --profile container up
# or LLM=deepseek-coder OLLAMA_BASE_URL=http://ollama-service:11434 docker compose --profile container watch
The first time only, you must wait for the complete downloading of the model.
To do for the first time only:
LLM=deepseek-coder
ollama pull ${LLM}
HTTP_PORT=8888 LLM=deepseek-coder OLLAMA_BASE_URL=http://host.docker.internal:11434 docker compose --profile webapp up
# or LLM=deepseek-coder OLLAMA_BASE_URL=http://host.docker.internal:11434 docker compose --profile webapp watch
docker compose --env-file deepseek-coder-instruct.env --profile webapp up
ollama pull gemma:2b-instruct
docker compose --env-file gemma-2b-instruct.env --profile webapp up
cd webapp
# yarn install
LLM=deepseek-coder
OLLAMA_BASE_URL=http://localhost:11434
node index.mjs