mirror of
https://github.com/Haxxnet/Compose-Examples.git
synced 2025-11-23 17:04:27 +01:00
add llms
This commit is contained in:
30
examples/ollama-ui/README.md
Normal file
30
examples/ollama-ui/README.md
Normal file
@@ -0,0 +1,30 @@
|
||||
# References
|
||||
|
||||
- https://github.com/ollama/ollama
|
||||
- https://hub.docker.com/r/ollama/ollama
|
||||
- https://github.com/open-webui/open-webui
|
||||
|
||||
# Notes
|
||||
|
||||
You should spawn ollama first and download the respective LLM models:
|
||||
|
||||
````
|
||||
# spawn ollama
|
||||
docker compose up -d ollama
|
||||
|
||||
# download an llm model
|
||||
docker exec ollama ollama run llama3:8b
|
||||
````
|
||||
|
||||
Afterwards, we can spawn Open WebUI and register our first user account:
|
||||
|
||||
````
|
||||
# spawn ui
|
||||
docker compose up -d ui
|
||||
````
|
||||
|
||||
Finally, we may want to disable open user registration for Open WebUI by uncommenting the env `ENABLE_SIGNUP` variable and restarting the container.
|
||||
|
||||
> [!TIP]
|
||||
>
|
||||
> You likely want to pass a GPU into the Ollama container. Please read [this](https://hub.docker.com/r/ollama/ollama).
|
||||
39
examples/ollama-ui/docker-compose.yml
Normal file
39
examples/ollama-ui/docker-compose.yml
Normal file
@@ -0,0 +1,39 @@
|
||||
services:
|
||||
|
||||
ui:
|
||||
image: ghcr.io/open-webui/open-webui:main
|
||||
container_name: ollama-ui
|
||||
restart: always
|
||||
ports:
|
||||
- 8080
|
||||
expose:
|
||||
- 8080
|
||||
volumes:
|
||||
- ${DOCKER_VOLUME_STORAGE:-/mnt/docker-volumes}/ollama/open-webui:/app/backend/data
|
||||
environment:
|
||||
#- "ENABLE_SIGNUP=false"
|
||||
- "OLLAMA_BASE_URL=http://ollama:11434"
|
||||
#networks:
|
||||
# - proxy
|
||||
#labels:
|
||||
# - traefik.enable=true
|
||||
# - traefik.docker.network=proxy
|
||||
# - traefik.http.routers.ollama-ui.rule=Host(`ai.example.com`)
|
||||
# - traefik.http.services.ollama-ui.loadbalancer.server.port=8080
|
||||
# # Optional part for traefik middlewares
|
||||
# - traefik.http.routers.ollama-ui.middlewares=local-ipwhitelist@file,authelia@docker
|
||||
|
||||
ollama:
|
||||
image: ollama/ollama:latest
|
||||
container_name: ollama
|
||||
restart: always
|
||||
expose:
|
||||
- 11434
|
||||
volumes:
|
||||
- ${DOCKER_VOLUME_STORAGE:-/mnt/docker-volumes}/ollama/data:/root/.ollama
|
||||
#networks:
|
||||
# - proxy
|
||||
|
||||
#networks:
|
||||
# proxy:
|
||||
# external: true
|
||||
7
examples/serge/README.md
Normal file
7
examples/serge/README.md
Normal file
@@ -0,0 +1,7 @@
|
||||
# References
|
||||
|
||||
- https://github.com/serge-chat/serge
|
||||
|
||||
# Notes
|
||||
|
||||
TBD
|
||||
26
examples/serge/docker-compose.yml
Normal file
26
examples/serge/docker-compose.yml
Normal file
@@ -0,0 +1,26 @@
|
||||
services:
|
||||
|
||||
serge:
|
||||
image: ghcr.io/serge-chat/serge:main
|
||||
container_name: serge
|
||||
restart: unless-stopped
|
||||
ports:
|
||||
- 8008
|
||||
expose:
|
||||
- 8008
|
||||
volumes:
|
||||
- ${DOCKER_VOLUME_STORAGE:-/mnt/docker-volumes}/serge/weights:/usr/src/app/weights
|
||||
- ${DOCKER_VOLUME_STORAGE:-/mnt/docker-volumes}/serge/datadb:/data/db/
|
||||
#networks:
|
||||
# - proxy
|
||||
#labels:
|
||||
# - traefik.enable=true
|
||||
# - traefik.docker.network=proxy
|
||||
# - traefik.http.routers.serge.rule=Host(`serge.example.com`)
|
||||
# - traefik.http.services.serge.loadbalancer.server.port=8080
|
||||
# # Optional part for traefik middlewares
|
||||
# - traefik.http.routers.serge.middlewares=local-ipwhitelist@file,authelia@docker
|
||||
|
||||
#networks:
|
||||
# proxy:
|
||||
# external: true
|
||||
Reference in New Issue
Block a user