差異處
這裏顯示兩個版本的差異處。
下次修改 | 前次修改 | ||
tech:ollama [2024/05/10 23:37] – 建立 jonathan | tech:ollama [2024/12/18 18:34] (目前版本) – [Open WebUI 設定 Gemini API] jonathan | ||
---|---|---|---|
行 1: | 行 1: | ||
- | ====== | + | ====== Ollama |
* 安裝環境 VM 規格 : | * 安裝環境 VM 規格 : | ||
- | * vCPU [host] i7-11700 | + | * vCPU : 8 ([host] i7-11700 |
* RAM : 8G | * RAM : 8G | ||
* SSD : 64G | * SSD : 64G | ||
- | * OS : Ubuntu 20.04 | + | * OS : Alpine 3.19 + docker compose |
- | ===== 安裝與測試 ===== | + | ===== 使用 docker compose 快速安裝測試 ===== |
+ | * 編輯 docker-compose.yml < | ||
+ | services: | ||
+ | ollama: | ||
+ | image: ollama/ | ||
+ | ports: | ||
+ | - 11434: | ||
+ | volumes: | ||
+ | - .:/code | ||
+ | - ./ | ||
+ | container_name: | ||
+ | environment: | ||
+ | - OLLAMA_NUM_PARALLEL=4 | ||
+ | - OLLAMA_MAX_LOADED_MODELS=4 | ||
+ | pull_policy: | ||
+ | tty: true | ||
+ | restart: always | ||
+ | networks: | ||
+ | - ollama-docker | ||
+ | #deploy: | ||
+ | # resources: | ||
+ | # reservations: | ||
+ | # devices: | ||
+ | # - driver: nvidia | ||
+ | # count: 1 | ||
+ | # capabilities: | ||
+ | |||
+ | ollama-webui: | ||
+ | image: ghcr.io/ | ||
+ | container_name: | ||
+ | volumes: | ||
+ | - ./ | ||
+ | depends_on: | ||
+ | - ollama | ||
+ | ports: | ||
+ | - 8080:8080 | ||
+ | environment: | ||
+ | - '/ | ||
+ | extra_hosts: | ||
+ | - host.docker.internal: | ||
+ | restart: unless-stopped | ||
+ | networks: | ||
+ | - ollama-docker | ||
+ | |||
+ | pipelines: | ||
+ | image: ghcr.io/ | ||
+ | container_name: | ||
+ | ports: | ||
+ | - " | ||
+ | volumes: | ||
+ | - ./ | ||
+ | extra_hosts: | ||
+ | - " | ||
+ | restart: always | ||
+ | networks: | ||
+ | - ollama-docker | ||
+ | |||
+ | networks: | ||
+ | ollama-docker: | ||
+ | external: false | ||
+ | </ | ||
+ | * 啟動 docker compose < | ||
+ | docker compose up -d | ||
+ | </ | ||
+ | * Open-WebUI 網址 - http:// | ||
+ | * Ollama API 網址 - http:// | ||
+ | |||
+ | ==== Open WebUI 設定 pipeline 使用雲端 LLM API === | ||
+ | * 採用 pipeline 機制 - https:// | ||
+ | - 設定 **連線** 新增 **http:// | ||
+ | |||
+ | === Anthropic API === | ||
+ | - **管線(pipeline)** 從 ++GitHub URL 安裝| \\ Repo - https:// | ||
+ | - ++下拉模型就可以選到 anthropic/ | ||
+ | - ++將不想使用的模型隱藏| \\ {{: | ||
+ | |||
+ | === Gemini API === | ||
+ | - **管線(pipeline)** 從 ++GitHub URL 安裝| \\ Repo - https:// | ||
+ | |||
+ | ==== Model ===== | ||
+ | * [[https:// | ||
+ | * [[https:// | ||
+ | * [[https:// | ||
+ | |||
+ | * 下載語法 Exp. ycchen/ | ||
+ | docker exec ollama ollama pull ycchen/ | ||
+ | </ | ||
+ | |||
+ | ==== 評估安裝環境運行 Model 的效能 ===== | ||
+ | * Exp. phi3 < | ||
+ | pve-ollama-221: | ||
+ | NAME ID SIZE MODIFIED | ||
+ | phi3: | ||
+ | codestral: | ||
+ | yabi/ | ||
+ | phi3: | ||
+ | |||
+ | pve-ollama-221: | ||
+ | >>> | ||
+ | | ||
+ | |||
+ | total duration: | ||
+ | load duration: | ||
+ | prompt eval count: | ||
+ | prompt eval duration: 101.53ms | ||
+ | prompt eval rate: 49.25 tokens/s | ||
+ | eval count: | ||
+ | eval duration: | ||
+ | eval rate: 11.08 tokens/s | ||
+ | </ | ||
+ | ===== VM 內直接安裝與測試 | ||
* 安裝程序< | * 安裝程序< | ||
curl -fsSL https:// | curl -fsSL https:// | ||
行 65: | 行 175: | ||
>>> | >>> | ||
</ | </ | ||
+ | |||
===== 參考網址 ===== | ===== 參考網址 ===== | ||
* https:// | * https:// | ||
+ | * https:// | ||
+ | * https:// | ||
* https:// | * https:// | ||
{{tag> | {{tag> | ||