From d4c43315fb3b266f55f1f1b8f3a378e233d66063 Mon Sep 17 00:00:00 2001 From: deepgeek Date: Fri, 18 Apr 2025 12:38:06 +0800 Subject: [PATCH] =?UTF-8?q?feat(*):=E5=A2=9E=E5=8A=A0=E4=BA=86=E4=B8=80?= =?UTF-8?q?=E4=BA=9B=E6=A8=A1=E5=9E=8B?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- anythinllm/.env | 2 +- chattts/compose.yaml | 2 +- comfyui/compose.yaml | 10 +- dbgpt/.env | 1 + dbgpt/compose.yaml | 23 +++ nlp/compose.yaml | 2 +- ollama/compose.yaml | 2 +- open-webui/compose.yaml | 7 +- owlsam/compose.yaml | 2 +- ragflow/.env | 148 ++++++++++++++++ ragflow/compose.yaml | 181 ++++++++++++++++++++ stable-diffusion-webui/compose.yaml | 12 +- toolbox-ocr/compose.yaml | 2 +- vllm-deepseek-r1-32b/compose.yaml | 8 +- vllm-deepseek-r1-70b/compose.yaml | 12 +- vllm-qwen25-72b-instruct-awq/compose.yaml | 12 +- vllm-qwen25-coder-32b-instruct/compose.yaml | 4 +- vllm-qwen25-vl-32b-instruct/compose.yaml | 4 +- vllm-qwq-32b/compose.yaml | 8 +- 19 files changed, 398 insertions(+), 44 deletions(-) create mode 100644 dbgpt/.env create mode 100644 dbgpt/compose.yaml create mode 100644 ragflow/.env create mode 100644 ragflow/compose.yaml mode change 100644 => 100755 vllm-deepseek-r1-32b/compose.yaml mode change 100644 => 100755 vllm-deepseek-r1-70b/compose.yaml mode change 100644 => 100755 vllm-qwen25-72b-instruct-awq/compose.yaml mode change 100644 => 100755 vllm-qwen25-coder-32b-instruct/compose.yaml mode change 100644 => 100755 vllm-qwen25-vl-32b-instruct/compose.yaml mode change 100644 => 100755 vllm-qwq-32b/compose.yaml diff --git a/anythinllm/.env b/anythinllm/.env index 6416fc5..70bce6e 100755 --- a/anythinllm/.env +++ b/anythinllm/.env @@ -35,7 +35,7 @@ JWT_SECRET="Mnx6P2gPXDz1FngbX3Vmn9SB4T2EVeE4JDkrqM2biA0o6nrWxDNE34QTXzxSR7ToKpTn # LLM_PROVIDER='localai' LOCAL_AI_BASE_PATH='http://host.docker.internal:10580/v1' -LOCAL_AI_MODEL_PREF='DeepSeek-R1-Distill-Llama-70B-AWQ' +LOCAL_AI_MODEL_PREF='deepseek-r1:671b' LOCAL_AI_MODEL_TOKEN_LIMIT=8192 LOCAL_AI_API_KEY="O8Is3NSYnp5fICWFbhkbwpLWgvMLkdCSuXR5ZggLmgwTKNPEWsjx1NqUxkyU7wLX" diff --git a/chattts/compose.yaml b/chattts/compose.yaml index 4f8a22b..e672b4f 100755 --- a/chattts/compose.yaml +++ b/chattts/compose.yaml @@ -13,7 +13,7 @@ services: count: all capabilities: - gpu - restart: always + restart: unless-stopped command: python webui.py runtime: nvidia x-dockge: diff --git a/comfyui/compose.yaml b/comfyui/compose.yaml index c71bdc2..ea8419b 100755 --- a/comfyui/compose.yaml +++ b/comfyui/compose.yaml @@ -1,14 +1,14 @@ version: "3.8" services: comfyui: - image: docker.citory.tech/mirror/yanwk/comfyui-boot:cu121 + image: docker.citory.tech/mirror/yanwk/comfyui-boot:cu124-cn container_name: comfyui ports: - 10587:8188 volumes: - - /home/deepgeek/data/data_local/server/comfyui/storage:/home/runner - - /home/deepgeek/data/data_local/server/sd-models/checkpoints:/home/runner/ComfyUI/models/checkpoints - - /home/deepgeek/data/data_local/server/sd-models/loras:/home/runner/ComfyUI/models/loras + - /home/deepgeek/data/data_base/comfyui/storage/ComfyUI:/root/ComfyUI + - /home/deepgeek/data/data_base/sd-models/checkpoints:/root/ComfyUI/models/checkpoints + - /home/deepgeek/data/data_base/sd-models/loras:/root/ComfyUI/models/loras deploy: resources: reservations: @@ -19,7 +19,7 @@ services: - compute - utility - gpu - restart: always + restart: unless-stopped runtime: nvidia x-dockge: urls: diff --git a/dbgpt/.env b/dbgpt/.env new file mode 100644 index 0000000..c23c858 --- /dev/null +++ b/dbgpt/.env @@ -0,0 +1 @@ +# VARIABLE=value #comment \ No newline at end of file diff --git a/dbgpt/compose.yaml b/dbgpt/compose.yaml new file mode 100644 index 0000000..a96bfe9 --- /dev/null +++ b/dbgpt/compose.yaml @@ -0,0 +1,23 @@ +services: + dbgpt: + image: docker.citory.tech/mirror/eosphorosai/dbgpt-openai:latest + container_name: dbgpt + environment: + - SILICONFLOW_API_KEY=sk-tlsrjadpcrvebqdmptabfrxznnuiimsawowtpnzbokpanfcx + volumes: + - dbgpt-data:/app/pilot/data + - dbgpt-message:/app/pilot/message + ports: + - 10506:5670/tcp + # webserver may be failed, it must wait all sqls in /docker-entrypoint-initdb.d execute finish. + restart: unless-stopped + ipc: host + tty: true + stdin_open: true +volumes: + dbgpt-data: + dbgpt-message: +x-dockge: + urls: + - http://local.citory.tech:10506 +networks: {} \ No newline at end of file diff --git a/nlp/compose.yaml b/nlp/compose.yaml index 9511932..98e63bd 100755 --- a/nlp/compose.yaml +++ b/nlp/compose.yaml @@ -13,5 +13,5 @@ services: count: all capabilities: - gpu - restart: always + restart: unless-stopped runtime: nvidia diff --git a/ollama/compose.yaml b/ollama/compose.yaml index f7f1545..468cd4c 100755 --- a/ollama/compose.yaml +++ b/ollama/compose.yaml @@ -18,7 +18,7 @@ services: runtime: nvidia volumes: - /home/deepgeek/data/data_base/ollama:/root/.ollama - restart: always + restart: unless-stopped command: ollama run gemma3:27b x-dockge: urls: diff --git a/open-webui/compose.yaml b/open-webui/compose.yaml index 7810a29..9339ac6 100755 --- a/open-webui/compose.yaml +++ b/open-webui/compose.yaml @@ -1,13 +1,13 @@ version: "3.8" services: open-webui: - image: docker.citory.tech/mirror/backplane/open-webui:0.5.20 + image: docker.citory.tech/mirror/backplane/open-webui:0 container_name: open-webui ports: - 10503:8080 environment: - - OPENAI_API_KEY='' # 替换为您的 OpenAI API 密钥 - - OPENAI_API_BASE_URL=http://host.docker.internal:10580/v1 # 替换为您的 VLLM IP 和端口 + - OPENAI_API_KEY='' + - OPENAI_API_BASE_URL=http://host.docker.internal:10580/v1 - WHISPER_MODEL=large - ENABLE_OLLAMA_API=true - OLLAMA_BASE_URL=http://host.docker.internal:10580 @@ -27,3 +27,4 @@ services: x-dockge: urls: - http://local.citory.tech:10503 +networks: {} diff --git a/owlsam/compose.yaml b/owlsam/compose.yaml index 6c02459..5c914b7 100755 --- a/owlsam/compose.yaml +++ b/owlsam/compose.yaml @@ -14,7 +14,7 @@ services: count: all capabilities: - gpu - restart: always + restart: unless-stopped runtime: nvidia x-dockge: urls: diff --git a/ragflow/.env b/ragflow/.env new file mode 100644 index 0000000..4d0d93a --- /dev/null +++ b/ragflow/.env @@ -0,0 +1,148 @@ +# The type of doc engine to use. +# Available options: +# - `elasticsearch` (default) +# - `infinity` (https://github.com/infiniflow/infinity) +DOC_ENGINE=${DOC_ENGINE:-elasticsearch} + +# ------------------------------ +# docker env var for specifying vector db type at startup +# (based on the vector db type, the corresponding docker +# compose profile will be used) +# ------------------------------ +COMPOSE_PROFILES=${DOC_ENGINE} + +# The version of Elasticsearch. +STACK_VERSION=8.11.3 + +# The hostname where the Elasticsearch service is exposed +ES_HOST=es01 + +# The port used to expose the Elasticsearch service to the host machine, +# allowing EXTERNAL access to the service running inside the Docker container. +ES_PORT=1200 + +# The password for Elasticsearch. +ELASTIC_PASSWORD=infini_rag_flow + +# The port used to expose the Kibana service to the host machine, +# allowing EXTERNAL access to the service running inside the Docker container. +KIBANA_PORT=6601 +KIBANA_USER=rag_flow +KIBANA_PASSWORD=infini_rag_flow + +# The maximum amount of the memory, in bytes, that a specific Docker container can use while running. +# Update it according to the available memory in the host machine. +MEM_LIMIT=8073741824 + +# The hostname where the Infinity service is exposed +INFINITY_HOST=infinity + +# Port to expose Infinity API to the host +INFINITY_THRIFT_PORT=23817 +INFINITY_HTTP_PORT=23820 +INFINITY_PSQL_PORT=5432 + +# The password for MySQL. +MYSQL_PASSWORD=infini_rag_flow +# The hostname where the MySQL service is exposed +MYSQL_HOST=mysql +# The database of the MySQL service to use +MYSQL_DBNAME=rag_flow +# The port used to expose the MySQL service to the host machine, +# allowing EXTERNAL access to the MySQL database running inside the Docker container. +MYSQL_PORT=5455 + +# The hostname where the MinIO service is exposed +MINIO_HOST=minio +# The port used to expose the MinIO console interface to the host machine, +# allowing EXTERNAL access to the web-based console running inside the Docker container. +MINIO_CONSOLE_PORT=9001 +# The port used to expose the MinIO API service to the host machine, +# allowing EXTERNAL access to the MinIO object storage service running inside the Docker container. +MINIO_PORT=9000 +# The username for MinIO. +# When updated, you must revise the `minio.user` entry in service_conf.yaml accordingly. +MINIO_USER=rag_flow +# The password for MinIO. +# When updated, you must revise the `minio.password` entry in service_conf.yaml accordingly. +MINIO_PASSWORD=infini_rag_flow + +# The hostname where the Redis service is exposed +REDIS_HOST=redis +# The port used to expose the Redis service to the host machine, +# allowing EXTERNAL access to the Redis service running inside the Docker container. +REDIS_PORT=6380 +# The password for Redis. +REDIS_PASSWORD=infini_rag_flow + +# The port used to expose RAGFlow's HTTP API service to the host machine, +# allowing EXTERNAL access to the service running inside the Docker container. +SVR_HTTP_PORT=9380 +HTTP_PORT=10502 +# The RAGFlow Docker image to download. +# Defaults to the v0.17.2-slim edition, which is the RAGFlow Docker image without embedding models. +RAGFLOW_IMAGE=docker.citory.tech/mirror/infiniflow/ragflow:v0.17.2 +# +# To download the RAGFlow Docker image with embedding models, uncomment the following line instead: +# RAGFLOW_IMAGE=infiniflow/ragflow:v0.17.2 +# +# The Docker image of the v0.17.2 edition includes: +# - Built-in embedding models: +# - BAAI/bge-large-zh-v1.5 +# - BAAI/bge-reranker-v2-m3 +# - maidalun1020/bce-embedding-base_v1 +# - maidalun1020/bce-reranker-base_v1 +# - Embedding models that will be downloaded once you select them in the RAGFlow UI: +# - BAAI/bge-base-en-v1.5 +# - BAAI/bge-large-en-v1.5 +# - BAAI/bge-small-en-v1.5 +# - BAAI/bge-small-zh-v1.5 +# - jinaai/jina-embeddings-v2-base-en +# - jinaai/jina-embeddings-v2-small-en +# - nomic-ai/nomic-embed-text-v1.5 +# - sentence-transformers/all-MiniLM-L6-v2 +# +# + + +# If you cannot download the RAGFlow Docker image: +# +# - For the `nightly-slim` edition, uncomment either of the following: +# RAGFLOW_IMAGE=swr.cn-north-4.myhuaweicloud.com/infiniflow/ragflow:nightly-slim +# RAGFLOW_IMAGE=registry.cn-hangzhou.aliyuncs.com/infiniflow/ragflow:nightly-slim +# +# - For the `nightly` edition, uncomment either of the following: +# RAGFLOW_IMAGE=swr.cn-north-4.myhuaweicloud.com/infiniflow/ragflow:nightly +# RAGFLOW_IMAGE=registry.cn-hangzhou.aliyuncs.com/infiniflow/ragflow:nightly + +# The local time zone. +TIMEZONE='Asia/Shanghai' + +# Uncomment the following line if you have limited access to huggingface.co: +HF_ENDPOINT=https://hf-mirror.com + +# Optimizations for MacOS +# Uncomment the following line if your operating system is MacOS: +# MACOS=1 + +# The maximum file size for each uploaded file, in bytes. +# You can uncomment this line and update the value if you wish to change the 1024M file size limit +# MAX_CONTENT_LENGTH=1073741824 +# After making the change, ensure you update `client_max_body_size` in nginx/nginx.conf correspondingly. + +# The log level for the RAGFlow's owned packages and imported packages. +# Available level: +# - `DEBUG` +# - `INFO` (default) +# - `WARNING` +# - `ERROR` +# For example, following line changes the log level of `ragflow.es_conn` to `DEBUG`: +# LOG_LEVELS=ragflow.es_conn=DEBUG + +# aliyun OSS configuration +# STORAGE_IMPL=OSS +# ACCESS_KEY=xxx +# SECRET_KEY=eee +# ENDPOINT=http://oss-cn-hangzhou.aliyuncs.com +# REGION=cn-hangzhou +# BUCKET=ragflow65536 diff --git a/ragflow/compose.yaml b/ragflow/compose.yaml new file mode 100644 index 0000000..886f9d2 --- /dev/null +++ b/ragflow/compose.yaml @@ -0,0 +1,181 @@ +services: + es01: + container_name: ragflow-es-01 + profiles: + - elasticsearch + image: docker.citory.tech/mirror/elasticsearch:${STACK_VERSION} + volumes: + - /home/deepgeek/data/data_local/server/ragflow/ragflow-docker/data/esdata01:/usr/share/elasticsearch/data + ports: + - ${ES_PORT}:9200 + env_file: .env + environment: + - node.name=es01 + - ELASTIC_PASSWORD=${ELASTIC_PASSWORD} + - bootstrap.memory_lock=false + - discovery.type=single-node + - xpack.security.enabled=true + - xpack.security.http.ssl.enabled=false + - xpack.security.transport.ssl.enabled=false + - cluster.routing.allocation.disk.watermark.low=5gb + - cluster.routing.allocation.disk.watermark.high=3gb + - cluster.routing.allocation.disk.watermark.flood_stage=2gb + - TZ=${TIMEZONE} + mem_limit: ${MEM_LIMIT} + ulimits: + memlock: + soft: -1 + hard: -1 + healthcheck: + test: ["CMD-SHELL", "curl http://localhost:9200"] + interval: 10s + timeout: 10s + retries: 120 + networks: + - ragflow + restart: on-failure + + infinity: + container_name: ragflow-infinity + profiles: + - infinity + image: docker.citory.tech/mirror/infiniflow/infinity:v0.6.0-dev3 + volumes: + - /home/deepgeek/data/data_local/server/ragflow/ragflow-docker/data/infinity_data:/var/infinity + - /home/deepgeek/data/data_local/server/ragflow/ragflow-docker/infinity_conf.toml:/infinity_conf.toml + command: ["-f", "/infinity_conf.toml"] + ports: + - ${INFINITY_THRIFT_PORT}:23817 + - ${INFINITY_HTTP_PORT}:23820 + - ${INFINITY_PSQL_PORT}:5432 + env_file: .env + environment: + - TZ=${TIMEZONE} + mem_limit: ${MEM_LIMIT} + ulimits: + nofile: + soft: 500000 + hard: 500000 + networks: + - ragflow + healthcheck: + test: ["CMD", "curl", "http://localhost:23820/admin/node/current"] + interval: 10s + timeout: 10s + retries: 120 + restart: on-failure + + + mysql: + # mysql:5.7 linux/arm64 image is unavailable. + image: docker.citory.tech/mirror/mysql:8.0.39 + container_name: ragflow-mysql + env_file: .env + environment: + - MYSQL_ROOT_PASSWORD=${MYSQL_PASSWORD} + - TZ=${TIMEZONE} + command: + --max_connections=1000 + --character-set-server=utf8mb4 + --collation-server=utf8mb4_unicode_ci + --default-authentication-plugin=mysql_native_password + --tls_version="TLSv1.2,TLSv1.3" + --init-file /data/application/init.sql + ports: + - ${MYSQL_PORT}:3306 + volumes: + - /home/deepgeek/data/data_local/server/ragflow/ragflow-docker/data/mysql_data:/var/lib/mysql + - /home/deepgeek/data/data_local/server/ragflow/ragflow-docker/init.sql:/data/application/init.sql + networks: + - ragflow + healthcheck: + test: ["CMD", "mysqladmin" ,"ping", "-uroot", "-p${MYSQL_PASSWORD}"] + interval: 10s + timeout: 10s + retries: 3 + restart: on-failure + + minio: + image: docker.citory.tech/mirror/minio/minio:RELEASE.2023-12-20T01-00-02Z + container_name: ragflow-minio + command: server --console-address ":9001" /data + ports: + - ${MINIO_PORT}:9000 + - ${MINIO_CONSOLE_PORT}:9001 + env_file: .env + environment: + - MINIO_ROOT_USER=${MINIO_USER} + - MINIO_ROOT_PASSWORD=${MINIO_PASSWORD} + - TZ=${TIMEZONE} + volumes: + - /home/deepgeek/data/data_local/server/ragflow/ragflow-docker/data/minio_data:/data + networks: + - ragflow + restart: on-failure + + redis: + # swr.cn-north-4.myhuaweicloud.com/ddn-k8s/docker.io/valkey/valkey:8 + image: docker.citory.tech/mirror/valkey/valkey:8 + container_name: ragflow-redis + command: redis-server --requirepass ${REDIS_PASSWORD} --maxmemory 128mb --maxmemory-policy allkeys-lru + env_file: .env + ports: + - ${REDIS_PORT}:6379 + volumes: + - /home/deepgeek/data/data_local/server/ragflow/ragflow-docker/data/redis_data:/data + networks: + - ragflow + restart: on-failure + + ragflow: + depends_on: + mysql: + condition: service_healthy + image: ${RAGFLOW_IMAGE} + container_name: ragflow-server + ports: + - ${SVR_HTTP_PORT}:9380 + - ${HTTP_PORT}:80 + # - 443:443 + volumes: + - /home/deepgeek/data/data_local/server/ragflow/ragflow-docker/ragflow-logs:/ragflow/logs + - /home/deepgeek/data/data_local/server/ragflow/ragflow-docker/nginx/ragflow.conf:/etc/nginx/conf.d/ragflow.conf + - /home/deepgeek/data/data_local/server/ragflow/ragflow-docker/nginx/proxy.conf:/etc/nginx/proxy.conf + - /home/deepgeek/data/data_local/server/ragflow/ragflow-docker/nginx/nginx.conf:/etc/nginx/nginx.conf + env_file: .env + environment: + - TZ=${TIMEZONE} + - HF_ENDPOINT=${HF_ENDPOINT} + - MACOS=${MACOS} + networks: + - ragflow + restart: on-failure + # https://docs.docker.com/engine/daemon/prometheus/#create-a-prometheus-configuration + # If you're using Docker Desktop, the --add-host flag is optional. This flag makes sure that the host's internal IP gets exposed to the Prometheus container. + extra_hosts: + - "host.docker.internal:host-gateway" + # executor: + # depends_on: + # mysql: + # condition: service_healthy + # image: ${RAGFLOW_IMAGE} + # container_name: ragflow-executor + # volumes: + # - ./ragflow-logs:/ragflow/logs + # - ./nginx/ragflow.conf:/etc/nginx/conf.d/ragflow.conf + # env_file: .env + # environment: + # - TZ=${TIMEZONE} + # - HF_ENDPOINT=${HF_ENDPOINT} + # - MACOS=${MACOS} + # entrypoint: "/ragflow/entrypoint_task_executor.sh 1 3" + # networks: + # - ragflow + # restart: on-failure + # # https://docs.docker.com/engine/daemon/prometheus/#create-a-prometheus-configuration + # # If you're using Docker Desktop, the --add-host flag is optional. This flag makes sure that the host's internal IP gets exposed to the Prometheus container. + # extra_hosts: + # - "host.docker.internal:host-gateway" +networks: + ragflow: + driver: bridge diff --git a/stable-diffusion-webui/compose.yaml b/stable-diffusion-webui/compose.yaml index 5d5c944..6419ce9 100755 --- a/stable-diffusion-webui/compose.yaml +++ b/stable-diffusion-webui/compose.yaml @@ -6,11 +6,11 @@ services: ports: - 10588:7860 volumes: - - /home/deepgeek/data/data_local/server/stable-diffusion-webui-docker/data:/data - - /home/deepgeek/data/data_local/server/stable-diffusion-webui-docker/data/config/auto/localizations:/stable-diffusion-webui/localizations - - /home/deepgeek/data/data_local/server/stable-diffusion-webui-docker/output:/output - - /home/deepgeek/data/data_local/server/sd-models/checkpoints:/data/models/Stable-diffusion - - /home/deepgeek/data/data_local/server/sd-models/loras:/data/models/Lora + - /home/deepgeek/data/data_base/stable-diffusion-webui-docker/data:/data + - /home/deepgeek/data/data_base/stable-diffusion-webui-docker/data/config/auto/localizations:/stable-diffusion-webui/localizations + - /home/deepgeek/data/data_base/stable-diffusion-webui-docker/output:/output + - /home/deepgeek/data/data_base/sd-models/checkpoints:/data/models/Stable-diffusion + - /home/deepgeek/data/data_base/sd-models/loras:/data/models/Lora tty: true environment: - CUDA_LAUNCH_BLOCKING=1 @@ -26,7 +26,7 @@ services: - compute - utility - gpu - restart: always + restart: unless-stopped runtime: nvidia x-dockge: urls: diff --git a/toolbox-ocr/compose.yaml b/toolbox-ocr/compose.yaml index d1a5c66..906c02f 100755 --- a/toolbox-ocr/compose.yaml +++ b/toolbox-ocr/compose.yaml @@ -13,7 +13,7 @@ services: count: all capabilities: - gpu - restart: always + restart: unless-stopped runtime: nvidia x-dockge: urls: diff --git a/vllm-deepseek-r1-32b/compose.yaml b/vllm-deepseek-r1-32b/compose.yaml old mode 100644 new mode 100755 index 5d217b4..41646c5 --- a/vllm-deepseek-r1-32b/compose.yaml +++ b/vllm-deepseek-r1-32b/compose.yaml @@ -17,13 +17,13 @@ services: runtime: nvidia volumes: - /home/deepgeek/data/data_local/server/vllm/models:/models - restart: always - command: --served-model-name deepseek-r1:32b越狱 --model + restart: unless-stopped + command: --served-model-name deepseek-r1:32b --model /models/huihui-ai/DeepSeek-R1-Distill-Qwen-32B-abliterated --trust-remote-code --host 0.0.0.0 --port 8080 --max-model-len 8192 --tensor-parallel-size 2 --gpu_memory_utilization 0.9 --enforce-eager - --dtype auto --swap-space 8 --enable-auto-tool-choice - --tool-call-parser pythonic + --dtype auto --swap-space 8 --enable-auto-tool-choice --tool-call-parser + pythonic x-dockge: urls: - http://local.citory.tech:10580 diff --git a/vllm-deepseek-r1-70b/compose.yaml b/vllm-deepseek-r1-70b/compose.yaml old mode 100644 new mode 100755 index 1928afc..1eff7c3 --- a/vllm-deepseek-r1-70b/compose.yaml +++ b/vllm-deepseek-r1-70b/compose.yaml @@ -18,12 +18,12 @@ services: runtime: nvidia volumes: - /home/deepgeek/data/data_local/server/vllm/models:/models - restart: always - command: --served-model-name deepseek-r1:70b越狱 --model - /models/Fasiany/DeepSeek-R1-Distill-Llama-70B-abliterated-GPTQ-int4 - --trust-remote-code --host 0.0.0.0 --port 8080 --enable-auto-tool-choice --tool-call-parser pythonic --max-model-len 8192 - --tensor-parallel-size 2 --gpu_memory_utilization 0.9 --enforce-eager - --dtype auto --swap-space 8 + restart: unless-stopped + command: --served-model-name deepseek-r1:70b + --model /models/Valdemardi/DeepSeek-R1-Distill-Llama-70B-AWQ + --trust-remote-code --host 0.0.0.0 --port 8080 --enable-auto-tool-choice + --tool-call-parser pythonic --max-model-len 8192 --tensor-parallel-size 2 + --gpu_memory_utilization 0.9 --enforce-eager --dtype auto --swap-space 8 x-dockge: urls: - http://local.citory.tech:10580 diff --git a/vllm-qwen25-72b-instruct-awq/compose.yaml b/vllm-qwen25-72b-instruct-awq/compose.yaml old mode 100644 new mode 100755 index 50112bd..977c854 --- a/vllm-qwen25-72b-instruct-awq/compose.yaml +++ b/vllm-qwen25-72b-instruct-awq/compose.yaml @@ -18,12 +18,12 @@ services: runtime: nvidia volumes: - /home/deepgeek/data/data_local/server/vllm/models:/models - restart: always - command: --served-model-name Qwen2.5-72B-Instruct-AWQ --model /models/Qwen/Qwen2.5-72B-Instruct-AWQ - --trust-remote-code --host 0.0.0.0 --port 8080 --max-model-len 8192 - --tensor-parallel-size 2 --gpu_memory_utilization 0.9 --enforce-eager - --dtype auto --swap-space 8 --enable-auto-tool-choice --tool-call-parser - llama3_json + restart: unless-stopped + command: --served-model-name Qwen2.5:72b --model + /models/Qwen/Qwen2.5-72B-Instruct-AWQ --trust-remote-code --host 0.0.0.0 + --port 8080 --max-model-len 8192 --tensor-parallel-size 2 + --gpu_memory_utilization 0.9 --enforce-eager --dtype auto --swap-space 8 + --enable-auto-tool-choice --tool-call-parser llama3_json x-dockge: urls: - http://local.citory.tech:10580 diff --git a/vllm-qwen25-coder-32b-instruct/compose.yaml b/vllm-qwen25-coder-32b-instruct/compose.yaml old mode 100644 new mode 100755 index ca45ef4..e052cce --- a/vllm-qwen25-coder-32b-instruct/compose.yaml +++ b/vllm-qwen25-coder-32b-instruct/compose.yaml @@ -18,8 +18,8 @@ services: runtime: nvidia volumes: - /home/deepgeek/data/data_local/server/vllm/models:/models - restart: always - command: --served-model-name Qwen2.5-Coder-32B-Instruct --model + restart: unless-stopped + command: --served-model-name Qwen2.5-Coder:32b --model /models/Qwen/Qwen2.5-Coder-32B-Instruct --trust-remote-code --host 0.0.0.0 --port 8080 --max-model-len 8192 --tensor-parallel-size 2 --gpu_memory_utilization 0.9 --enforce-eager --dtype auto --swap-space 8 diff --git a/vllm-qwen25-vl-32b-instruct/compose.yaml b/vllm-qwen25-vl-32b-instruct/compose.yaml old mode 100644 new mode 100755 index 490b805..f176e32 --- a/vllm-qwen25-vl-32b-instruct/compose.yaml +++ b/vllm-qwen25-vl-32b-instruct/compose.yaml @@ -18,8 +18,8 @@ services: runtime: nvidia volumes: - /home/deepgeek/data/data_local/server/vllm/models:/models - restart: always - command: --served-model-name Qwen2.5-VL-32B-Instruct --model + restart: unless-stopped + command: --served-model-name Qwen2.5-VL:32b --model /models/Qwen/Qwen2.5-VL-32B-Instruct --trust-remote-code --host 0.0.0.0 --port 8080 --max-model-len 8192 --tensor-parallel-size 2 --gpu_memory_utilization 0.9 --enforce-eager --dtype auto --swap-space 8 diff --git a/vllm-qwq-32b/compose.yaml b/vllm-qwq-32b/compose.yaml old mode 100644 new mode 100755 index a54720c..c84fa77 --- a/vllm-qwq-32b/compose.yaml +++ b/vllm-qwq-32b/compose.yaml @@ -18,12 +18,12 @@ services: runtime: nvidia volumes: - /home/deepgeek/data/data_local/server/vllm/models:/models - restart: always - command: --served-model-name QwQ-32B --model /models/Qwen/QwQ-32B + restart: unless-stopped + command: --served-model-name QwQ:32b --model /models/Qwen/QwQ-32B --trust-remote-code --host 0.0.0.0 --port 8080 --max-model-len 8192 --tensor-parallel-size 2 --gpu_memory_utilization 0.9 --enforce-eager - --dtype auto --swap-space 8 --enable-auto-tool-choice - --tool-call-parser llama3_json + --dtype auto --swap-space 8 --enable-auto-tool-choice --tool-call-parser + llama3_json x-dockge: urls: - http://local.citory.tech:10580