差異處
這裏顯示兩個版本的差異處。
兩邊的前次修訂版 前次修改 下次修改 | 前次修改 | ||
tech:ollama [2024/05/11 23:33] – [使用 docker compose 快速安裝測試] jonathan | tech:ollama [2024/12/18 18:34] (目前版本) – [Open WebUI 設定 Gemini API] jonathan | ||
---|---|---|---|
行 1: | 行 1: | ||
- | ====== Ollama + Open-WebUI | + | ====== Ollama + Open-WebUI ====== |
* 安裝環境 VM 規格 : | * 安裝環境 VM 規格 : | ||
* vCPU : 8 ([host] i7-11700 支援 avx2) | * vCPU : 8 ([host] i7-11700 支援 avx2) | ||
* RAM : 8G | * RAM : 8G | ||
* SSD : 64G | * SSD : 64G | ||
- | * OS : Ubuntu 20.04 | + | * OS : Alpine 3.19 + docker compose |
===== 使用 docker compose 快速安裝測試 ===== | ===== 使用 docker compose 快速安裝測試 ===== | ||
行 17: | 行 17: | ||
- ./ | - ./ | ||
container_name: | container_name: | ||
+ | environment: | ||
+ | - OLLAMA_NUM_PARALLEL=4 | ||
+ | - OLLAMA_MAX_LOADED_MODELS=4 | ||
pull_policy: | pull_policy: | ||
tty: true | tty: true | ||
行 22: | 行 25: | ||
networks: | networks: | ||
- ollama-docker | - ollama-docker | ||
+ | #deploy: | ||
+ | # resources: | ||
+ | # reservations: | ||
+ | # devices: | ||
+ | # - driver: nvidia | ||
+ | # count: 1 | ||
+ | # capabilities: | ||
ollama-webui: | ollama-webui: | ||
行 37: | 行 47: | ||
- host.docker.internal: | - host.docker.internal: | ||
restart: unless-stopped | restart: unless-stopped | ||
+ | networks: | ||
+ | - ollama-docker | ||
+ | |||
+ | pipelines: | ||
+ | image: ghcr.io/ | ||
+ | container_name: | ||
+ | ports: | ||
+ | - " | ||
+ | volumes: | ||
+ | - ./ | ||
+ | extra_hosts: | ||
+ | - " | ||
+ | restart: always | ||
networks: | networks: | ||
- ollama-docker | - ollama-docker | ||
行 50: | 行 73: | ||
* Ollama API 網址 - http:// | * Ollama API 網址 - http:// | ||
+ | ==== Open WebUI 設定 pipeline 使用雲端 LLM API === | ||
+ | * 採用 pipeline 機制 - https:// | ||
+ | - 設定 **連線** 新增 **http:// | ||
+ | |||
+ | === Anthropic API === | ||
+ | - **管線(pipeline)** 從 ++GitHub URL 安裝| \\ Repo - https:// | ||
+ | - ++下拉模型就可以選到 anthropic/ | ||
+ | - ++將不想使用的模型隱藏| \\ {{: | ||
+ | |||
+ | === Gemini API === | ||
+ | - **管線(pipeline)** 從 ++GitHub URL 安裝| \\ Repo - https:// | ||
+ | |||
+ | ==== Model ===== | ||
+ | * [[https:// | ||
+ | * [[https:// | ||
+ | * [[https:// | ||
+ | |||
+ | * 下載語法 Exp. ycchen/ | ||
+ | docker exec ollama ollama pull ycchen/ | ||
+ | </ | ||
+ | |||
+ | ==== 評估安裝環境運行 Model 的效能 ===== | ||
+ | * Exp. phi3 <cli> | ||
+ | pve-ollama-221: | ||
+ | NAME ID SIZE MODIFIED | ||
+ | phi3: | ||
+ | codestral: | ||
+ | yabi/ | ||
+ | phi3: | ||
+ | |||
+ | pve-ollama-221: | ||
+ | >>> | ||
+ | | ||
+ | |||
+ | total duration: | ||
+ | load duration: | ||
+ | prompt eval count: | ||
+ | prompt eval duration: 101.53ms | ||
+ | prompt eval rate: 49.25 tokens/s | ||
+ | eval count: | ||
+ | eval duration: | ||
+ | eval rate: 11.08 tokens/s | ||
+ | </ | ||
===== VM 內直接安裝與測試 Ollama ===== | ===== VM 內直接安裝與測試 Ollama ===== | ||
* 安裝程序< | * 安裝程序< |