mirror of
https://github.com/zylon-ai/private-gpt.git
synced 2025-12-22 07:40:12 +01:00
fix: naming image and ollama-cpu (#2056)
Some checks failed
release-please / release-please (push) Has been cancelled
tests / setup (push) Has been cancelled
tests / ${{ matrix.quality-command }} (black) (push) Has been cancelled
tests / ${{ matrix.quality-command }} (mypy) (push) Has been cancelled
tests / ${{ matrix.quality-command }} (ruff) (push) Has been cancelled
tests / test (push) Has been cancelled
tests / all_checks_passed (push) Has been cancelled
Some checks failed
release-please / release-please (push) Has been cancelled
tests / setup (push) Has been cancelled
tests / ${{ matrix.quality-command }} (black) (push) Has been cancelled
tests / ${{ matrix.quality-command }} (mypy) (push) Has been cancelled
tests / ${{ matrix.quality-command }} (ruff) (push) Has been cancelled
tests / test (push) Has been cancelled
tests / all_checks_passed (push) Has been cancelled
This commit is contained in:
parent
22904ca8ad
commit
89477ea9d3
1 changed files with 3 additions and 4 deletions
|
|
@ -7,7 +7,7 @@ services:
|
||||||
# Private-GPT service for the Ollama CPU and GPU modes
|
# Private-GPT service for the Ollama CPU and GPU modes
|
||||||
# This service builds from an external Dockerfile and runs the Ollama mode.
|
# This service builds from an external Dockerfile and runs the Ollama mode.
|
||||||
private-gpt-ollama:
|
private-gpt-ollama:
|
||||||
image: ${PGPT_IMAGE:-zylonai/private-gpt}${PGPT_TAG:-0.6.2}-ollama # x-release-please-version
|
image: ${PGPT_IMAGE:-zylonai/private-gpt}:${PGPT_TAG:-0.6.2}-ollama # x-release-please-version
|
||||||
build:
|
build:
|
||||||
context: .
|
context: .
|
||||||
dockerfile: Dockerfile.ollama
|
dockerfile: Dockerfile.ollama
|
||||||
|
|
@ -31,7 +31,7 @@ services:
|
||||||
# Private-GPT service for the local mode
|
# Private-GPT service for the local mode
|
||||||
# This service builds from a local Dockerfile and runs the application in local mode.
|
# This service builds from a local Dockerfile and runs the application in local mode.
|
||||||
private-gpt-llamacpp-cpu:
|
private-gpt-llamacpp-cpu:
|
||||||
image: ${PGPT_IMAGE:-zylonai/private-gpt}${PGPT_TAG:-0.6.2}-llamacpp-cpu # x-release-please-version
|
image: ${PGPT_IMAGE:-zylonai/private-gpt}:${PGPT_TAG:-0.6.2}-llamacpp-cpu # x-release-please-version
|
||||||
build:
|
build:
|
||||||
context: .
|
context: .
|
||||||
dockerfile: Dockerfile.llamacpp-cpu
|
dockerfile: Dockerfile.llamacpp-cpu
|
||||||
|
|
@ -57,7 +57,6 @@ services:
|
||||||
ollama:
|
ollama:
|
||||||
image: traefik:v2.10
|
image: traefik:v2.10
|
||||||
ports:
|
ports:
|
||||||
- "11435:11434"
|
|
||||||
- "8081:8080"
|
- "8081:8080"
|
||||||
command:
|
command:
|
||||||
- "--providers.file.filename=/etc/router.yml"
|
- "--providers.file.filename=/etc/router.yml"
|
||||||
|
|
@ -84,7 +83,7 @@ services:
|
||||||
- ./models:/root/.ollama
|
- ./models:/root/.ollama
|
||||||
profiles:
|
profiles:
|
||||||
- ""
|
- ""
|
||||||
- ollama
|
- ollama-cpu
|
||||||
|
|
||||||
# Ollama service for the CUDA mode
|
# Ollama service for the CUDA mode
|
||||||
ollama-cuda:
|
ollama-cuda:
|
||||||
|
|
|
||||||
Loading…
Add table
Add a link
Reference in a new issue