feat(*):增加了一些模型
This commit is contained in:
		| @@ -35,7 +35,7 @@ JWT_SECRET="Mnx6P2gPXDz1FngbX3Vmn9SB4T2EVeE4JDkrqM2biA0o6nrWxDNE34QTXzxSR7ToKpTn | ||||
|  | ||||
| # LLM_PROVIDER='localai' | ||||
| LOCAL_AI_BASE_PATH='http://host.docker.internal:10580/v1' | ||||
| LOCAL_AI_MODEL_PREF='DeepSeek-R1-Distill-Llama-70B-AWQ' | ||||
| LOCAL_AI_MODEL_PREF='deepseek-r1:671b' | ||||
| LOCAL_AI_MODEL_TOKEN_LIMIT=8192 | ||||
| LOCAL_AI_API_KEY="O8Is3NSYnp5fICWFbhkbwpLWgvMLkdCSuXR5ZggLmgwTKNPEWsjx1NqUxkyU7wLX" | ||||
|  | ||||
|   | ||||
| @@ -13,7 +13,7 @@ services: | ||||
|               count: all | ||||
|               capabilities: | ||||
|                 - gpu | ||||
|     restart: always | ||||
|     restart: unless-stopped | ||||
|     command: python webui.py | ||||
|     runtime: nvidia | ||||
| x-dockge: | ||||
|   | ||||
| @@ -1,14 +1,14 @@ | ||||
| version: "3.8" | ||||
| services: | ||||
|   comfyui: | ||||
|     image: docker.citory.tech/mirror/yanwk/comfyui-boot:cu121 | ||||
|     image: docker.citory.tech/mirror/yanwk/comfyui-boot:cu124-cn | ||||
|     container_name: comfyui | ||||
|     ports: | ||||
|       - 10587:8188 | ||||
|     volumes: | ||||
|       - /home/deepgeek/data/data_local/server/comfyui/storage:/home/runner | ||||
|       - /home/deepgeek/data/data_local/server/sd-models/checkpoints:/home/runner/ComfyUI/models/checkpoints | ||||
|       - /home/deepgeek/data/data_local/server/sd-models/loras:/home/runner/ComfyUI/models/loras | ||||
|       - /home/deepgeek/data/data_base/comfyui/storage/ComfyUI:/root/ComfyUI | ||||
|       - /home/deepgeek/data/data_base/sd-models/checkpoints:/root/ComfyUI/models/checkpoints | ||||
|       - /home/deepgeek/data/data_base/sd-models/loras:/root/ComfyUI/models/loras | ||||
|     deploy: | ||||
|       resources: | ||||
|         reservations: | ||||
| @@ -19,7 +19,7 @@ services: | ||||
|                 - compute | ||||
|                 - utility | ||||
|                 - gpu | ||||
|     restart: always | ||||
|     restart: unless-stopped | ||||
|     runtime: nvidia | ||||
| x-dockge: | ||||
|   urls: | ||||
|   | ||||
							
								
								
									
										1
									
								
								dbgpt/.env
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										1
									
								
								dbgpt/.env
									
									
									
									
									
										Normal file
									
								
							| @@ -0,0 +1 @@ | ||||
| # VARIABLE=value #comment | ||||
							
								
								
									
										23
									
								
								dbgpt/compose.yaml
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										23
									
								
								dbgpt/compose.yaml
									
									
									
									
									
										Normal file
									
								
							| @@ -0,0 +1,23 @@ | ||||
| services: | ||||
|   dbgpt: | ||||
|     image: docker.citory.tech/mirror/eosphorosai/dbgpt-openai:latest | ||||
|     container_name: dbgpt | ||||
|     environment: | ||||
|       - SILICONFLOW_API_KEY=sk-tlsrjadpcrvebqdmptabfrxznnuiimsawowtpnzbokpanfcx | ||||
|     volumes: | ||||
|       - dbgpt-data:/app/pilot/data | ||||
|       - dbgpt-message:/app/pilot/message | ||||
|     ports: | ||||
|       - 10506:5670/tcp | ||||
|     # webserver may be failed, it must wait all sqls in /docker-entrypoint-initdb.d execute finish. | ||||
|     restart: unless-stopped | ||||
|     ipc: host | ||||
|     tty: true | ||||
|     stdin_open: true | ||||
| volumes: | ||||
|   dbgpt-data: | ||||
|   dbgpt-message: | ||||
| x-dockge: | ||||
|   urls: | ||||
|     - http://local.citory.tech:10506 | ||||
| networks: {} | ||||
| @@ -13,5 +13,5 @@ services: | ||||
|               count: all | ||||
|               capabilities: | ||||
|                 - gpu | ||||
|     restart: always | ||||
|     restart: unless-stopped | ||||
|     runtime: nvidia | ||||
|   | ||||
| @@ -18,7 +18,7 @@ services: | ||||
|     runtime: nvidia | ||||
|     volumes: | ||||
|       - /home/deepgeek/data/data_base/ollama:/root/.ollama | ||||
|     restart: always | ||||
|     restart: unless-stopped | ||||
|     command: ollama run gemma3:27b | ||||
| x-dockge: | ||||
|   urls: | ||||
|   | ||||
| @@ -1,13 +1,13 @@ | ||||
| version: "3.8" | ||||
| services: | ||||
|   open-webui: | ||||
|     image: docker.citory.tech/mirror/backplane/open-webui:0.5.20 | ||||
|     image: docker.citory.tech/mirror/backplane/open-webui:0 | ||||
|     container_name: open-webui | ||||
|     ports: | ||||
|       - 10503:8080 | ||||
|     environment: | ||||
|       - OPENAI_API_KEY='' # 替换为您的 OpenAI API 密钥 | ||||
|       - OPENAI_API_BASE_URL=http://host.docker.internal:10580/v1 # 替换为您的 VLLM IP 和端口 | ||||
|       - OPENAI_API_KEY='' | ||||
|       - OPENAI_API_BASE_URL=http://host.docker.internal:10580/v1 | ||||
|       - WHISPER_MODEL=large | ||||
|       - ENABLE_OLLAMA_API=true | ||||
|       - OLLAMA_BASE_URL=http://host.docker.internal:10580 | ||||
| @@ -27,3 +27,4 @@ services: | ||||
| x-dockge: | ||||
|   urls: | ||||
|     - http://local.citory.tech:10503 | ||||
| networks: {} | ||||
|   | ||||
| @@ -14,7 +14,7 @@ services: | ||||
|               count: all | ||||
|               capabilities: | ||||
|                 - gpu | ||||
|     restart: always | ||||
|     restart: unless-stopped | ||||
|     runtime: nvidia | ||||
| x-dockge: | ||||
|   urls: | ||||
|   | ||||
							
								
								
									
										148
									
								
								ragflow/.env
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										148
									
								
								ragflow/.env
									
									
									
									
									
										Normal file
									
								
							| @@ -0,0 +1,148 @@ | ||||
| # The type of doc engine to use. | ||||
| # Available options: | ||||
| # - `elasticsearch` (default)  | ||||
| # - `infinity` (https://github.com/infiniflow/infinity) | ||||
| DOC_ENGINE=${DOC_ENGINE:-elasticsearch} | ||||
|  | ||||
| # ------------------------------ | ||||
| # docker env var for specifying vector db type at startup | ||||
| # (based on the vector db type, the corresponding docker | ||||
| # compose profile will be used) | ||||
| # ------------------------------ | ||||
| COMPOSE_PROFILES=${DOC_ENGINE} | ||||
|  | ||||
| # The version of Elasticsearch. | ||||
| STACK_VERSION=8.11.3 | ||||
|  | ||||
| # The hostname where the Elasticsearch service is exposed | ||||
| ES_HOST=es01 | ||||
|  | ||||
| # The port used to expose the Elasticsearch service to the host machine,  | ||||
| # allowing EXTERNAL access to the service running inside the Docker container. | ||||
| ES_PORT=1200 | ||||
|  | ||||
| # The password for Elasticsearch.  | ||||
| ELASTIC_PASSWORD=infini_rag_flow | ||||
|  | ||||
| # The port used to expose the Kibana service to the host machine,  | ||||
| # allowing EXTERNAL access to the service running inside the Docker container. | ||||
| KIBANA_PORT=6601 | ||||
| KIBANA_USER=rag_flow | ||||
| KIBANA_PASSWORD=infini_rag_flow | ||||
|  | ||||
| # The maximum amount of the memory, in bytes, that a specific Docker container can use while running. | ||||
| # Update it according to the available memory in the host machine. | ||||
| MEM_LIMIT=8073741824 | ||||
|  | ||||
| # The hostname where the Infinity service is exposed | ||||
| INFINITY_HOST=infinity | ||||
|  | ||||
| # Port to expose Infinity API to the host | ||||
| INFINITY_THRIFT_PORT=23817 | ||||
| INFINITY_HTTP_PORT=23820 | ||||
| INFINITY_PSQL_PORT=5432 | ||||
|  | ||||
| # The password for MySQL.  | ||||
| MYSQL_PASSWORD=infini_rag_flow | ||||
| # The hostname where the MySQL service is exposed | ||||
| MYSQL_HOST=mysql | ||||
| # The database of the MySQL service to use | ||||
| MYSQL_DBNAME=rag_flow | ||||
| # The port used to expose the MySQL service to the host machine,  | ||||
| # allowing EXTERNAL access to the MySQL database running inside the Docker container.  | ||||
| MYSQL_PORT=5455 | ||||
|  | ||||
| # The hostname where the MinIO service is exposed | ||||
| MINIO_HOST=minio | ||||
| # The port used to expose the MinIO console interface to the host machine,  | ||||
| # allowing EXTERNAL access to the web-based console running inside the Docker container.  | ||||
| MINIO_CONSOLE_PORT=9001 | ||||
| # The port used to expose the MinIO API service to the host machine,  | ||||
| # allowing EXTERNAL access to the MinIO object storage service running inside the Docker container.  | ||||
| MINIO_PORT=9000 | ||||
| # The username for MinIO.  | ||||
| # When updated, you must revise the `minio.user` entry in service_conf.yaml accordingly. | ||||
| MINIO_USER=rag_flow | ||||
| # The password for MinIO.  | ||||
| # When updated, you must revise the `minio.password` entry in service_conf.yaml accordingly. | ||||
| MINIO_PASSWORD=infini_rag_flow | ||||
|  | ||||
| # The hostname where the Redis service is exposed | ||||
| REDIS_HOST=redis | ||||
| # The port used to expose the Redis service to the host machine,  | ||||
| # allowing EXTERNAL access to the Redis service running inside the Docker container. | ||||
| REDIS_PORT=6380 | ||||
| # The password for Redis. | ||||
| REDIS_PASSWORD=infini_rag_flow | ||||
|  | ||||
| # The port used to expose RAGFlow's HTTP API service to the host machine,  | ||||
| # allowing EXTERNAL access to the service running inside the Docker container. | ||||
| SVR_HTTP_PORT=9380 | ||||
| HTTP_PORT=10502 | ||||
| # The RAGFlow Docker image to download. | ||||
| # Defaults to the v0.17.2-slim edition, which is the RAGFlow Docker image without embedding models. | ||||
| RAGFLOW_IMAGE=docker.citory.tech/mirror/infiniflow/ragflow:v0.17.2 | ||||
| # | ||||
| # To download the RAGFlow Docker image with embedding models, uncomment the following line instead: | ||||
| # RAGFLOW_IMAGE=infiniflow/ragflow:v0.17.2 | ||||
| #  | ||||
| # The Docker image of the v0.17.2 edition includes: | ||||
| # - Built-in embedding models: | ||||
| #   - BAAI/bge-large-zh-v1.5 | ||||
| #   - BAAI/bge-reranker-v2-m3 | ||||
| #   - maidalun1020/bce-embedding-base_v1 | ||||
| #   - maidalun1020/bce-reranker-base_v1 | ||||
| # - Embedding models that will be downloaded once you select them in the RAGFlow UI: | ||||
| #   - BAAI/bge-base-en-v1.5 | ||||
| #   - BAAI/bge-large-en-v1.5 | ||||
| #   - BAAI/bge-small-en-v1.5 | ||||
| #   - BAAI/bge-small-zh-v1.5 | ||||
| #   - jinaai/jina-embeddings-v2-base-en | ||||
| #   - jinaai/jina-embeddings-v2-small-en | ||||
| #   - nomic-ai/nomic-embed-text-v1.5 | ||||
| #   - sentence-transformers/all-MiniLM-L6-v2 | ||||
| # | ||||
| #  | ||||
|  | ||||
|  | ||||
| # If you cannot download the RAGFlow Docker image: | ||||
| # | ||||
| # - For the `nightly-slim` edition, uncomment either of the following: | ||||
| # RAGFLOW_IMAGE=swr.cn-north-4.myhuaweicloud.com/infiniflow/ragflow:nightly-slim | ||||
| # RAGFLOW_IMAGE=registry.cn-hangzhou.aliyuncs.com/infiniflow/ragflow:nightly-slim | ||||
| # | ||||
| # - For the `nightly` edition, uncomment either of the following: | ||||
| # RAGFLOW_IMAGE=swr.cn-north-4.myhuaweicloud.com/infiniflow/ragflow:nightly | ||||
| # RAGFLOW_IMAGE=registry.cn-hangzhou.aliyuncs.com/infiniflow/ragflow:nightly | ||||
|  | ||||
| # The local time zone. | ||||
| TIMEZONE='Asia/Shanghai' | ||||
|  | ||||
| # Uncomment the following line if you have limited access to huggingface.co: | ||||
| HF_ENDPOINT=https://hf-mirror.com | ||||
|  | ||||
| # Optimizations for MacOS | ||||
| # Uncomment the following line if your operating system is MacOS: | ||||
| # MACOS=1 | ||||
|  | ||||
| # The maximum file size for each uploaded file, in bytes. | ||||
| # You can uncomment this line and update the value if you wish to change the 1024M file size limit | ||||
| # MAX_CONTENT_LENGTH=1073741824 | ||||
| # After making the change, ensure you update `client_max_body_size` in nginx/nginx.conf correspondingly. | ||||
|  | ||||
| # The log level for the RAGFlow's owned packages and imported packages. | ||||
| # Available level: | ||||
| # - `DEBUG` | ||||
| # - `INFO` (default) | ||||
| # - `WARNING` | ||||
| # - `ERROR` | ||||
| # For example, following line changes the log level of `ragflow.es_conn` to `DEBUG`: | ||||
| # LOG_LEVELS=ragflow.es_conn=DEBUG | ||||
|  | ||||
| # aliyun OSS configuration | ||||
| # STORAGE_IMPL=OSS | ||||
| # ACCESS_KEY=xxx | ||||
| # SECRET_KEY=eee | ||||
| # ENDPOINT=http://oss-cn-hangzhou.aliyuncs.com | ||||
| # REGION=cn-hangzhou | ||||
| # BUCKET=ragflow65536 | ||||
							
								
								
									
										181
									
								
								ragflow/compose.yaml
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										181
									
								
								ragflow/compose.yaml
									
									
									
									
									
										Normal file
									
								
							| @@ -0,0 +1,181 @@ | ||||
| services: | ||||
|   es01: | ||||
|     container_name: ragflow-es-01 | ||||
|     profiles: | ||||
|       - elasticsearch | ||||
|     image: docker.citory.tech/mirror/elasticsearch:${STACK_VERSION} | ||||
|     volumes: | ||||
|       - /home/deepgeek/data/data_local/server/ragflow/ragflow-docker/data/esdata01:/usr/share/elasticsearch/data | ||||
|     ports: | ||||
|       - ${ES_PORT}:9200 | ||||
|     env_file: .env | ||||
|     environment: | ||||
|       - node.name=es01 | ||||
|       - ELASTIC_PASSWORD=${ELASTIC_PASSWORD} | ||||
|       - bootstrap.memory_lock=false | ||||
|       - discovery.type=single-node | ||||
|       - xpack.security.enabled=true | ||||
|       - xpack.security.http.ssl.enabled=false | ||||
|       - xpack.security.transport.ssl.enabled=false | ||||
|       - cluster.routing.allocation.disk.watermark.low=5gb | ||||
|       - cluster.routing.allocation.disk.watermark.high=3gb | ||||
|       - cluster.routing.allocation.disk.watermark.flood_stage=2gb | ||||
|       - TZ=${TIMEZONE} | ||||
|     mem_limit: ${MEM_LIMIT} | ||||
|     ulimits: | ||||
|       memlock: | ||||
|         soft: -1 | ||||
|         hard: -1 | ||||
|     healthcheck: | ||||
|       test: ["CMD-SHELL", "curl http://localhost:9200"] | ||||
|       interval: 10s | ||||
|       timeout: 10s | ||||
|       retries: 120 | ||||
|     networks: | ||||
|       - ragflow | ||||
|     restart: on-failure | ||||
|  | ||||
|   infinity: | ||||
|     container_name: ragflow-infinity | ||||
|     profiles: | ||||
|       - infinity | ||||
|     image: docker.citory.tech/mirror/infiniflow/infinity:v0.6.0-dev3 | ||||
|     volumes: | ||||
|       - /home/deepgeek/data/data_local/server/ragflow/ragflow-docker/data/infinity_data:/var/infinity | ||||
|       - /home/deepgeek/data/data_local/server/ragflow/ragflow-docker/infinity_conf.toml:/infinity_conf.toml | ||||
|     command: ["-f", "/infinity_conf.toml"] | ||||
|     ports: | ||||
|       - ${INFINITY_THRIFT_PORT}:23817 | ||||
|       - ${INFINITY_HTTP_PORT}:23820 | ||||
|       - ${INFINITY_PSQL_PORT}:5432 | ||||
|     env_file: .env | ||||
|     environment: | ||||
|       - TZ=${TIMEZONE} | ||||
|     mem_limit: ${MEM_LIMIT} | ||||
|     ulimits: | ||||
|       nofile: | ||||
|         soft: 500000 | ||||
|         hard: 500000 | ||||
|     networks: | ||||
|       - ragflow | ||||
|     healthcheck: | ||||
|       test: ["CMD", "curl", "http://localhost:23820/admin/node/current"] | ||||
|       interval: 10s | ||||
|       timeout: 10s | ||||
|       retries: 120 | ||||
|     restart: on-failure | ||||
|  | ||||
|  | ||||
|   mysql: | ||||
|     # mysql:5.7 linux/arm64 image is unavailable. | ||||
|     image: docker.citory.tech/mirror/mysql:8.0.39 | ||||
|     container_name: ragflow-mysql | ||||
|     env_file: .env | ||||
|     environment: | ||||
|       - MYSQL_ROOT_PASSWORD=${MYSQL_PASSWORD} | ||||
|       - TZ=${TIMEZONE} | ||||
|     command: | ||||
|       --max_connections=1000 | ||||
|       --character-set-server=utf8mb4 | ||||
|       --collation-server=utf8mb4_unicode_ci | ||||
|       --default-authentication-plugin=mysql_native_password | ||||
|       --tls_version="TLSv1.2,TLSv1.3" | ||||
|       --init-file /data/application/init.sql | ||||
|     ports: | ||||
|       - ${MYSQL_PORT}:3306 | ||||
|     volumes: | ||||
|       - /home/deepgeek/data/data_local/server/ragflow/ragflow-docker/data/mysql_data:/var/lib/mysql | ||||
|       - /home/deepgeek/data/data_local/server/ragflow/ragflow-docker/init.sql:/data/application/init.sql | ||||
|     networks: | ||||
|       - ragflow | ||||
|     healthcheck: | ||||
|       test: ["CMD", "mysqladmin" ,"ping", "-uroot", "-p${MYSQL_PASSWORD}"] | ||||
|       interval: 10s | ||||
|       timeout: 10s | ||||
|       retries: 3 | ||||
|     restart: on-failure | ||||
|  | ||||
|   minio: | ||||
|     image: docker.citory.tech/mirror/minio/minio:RELEASE.2023-12-20T01-00-02Z | ||||
|     container_name: ragflow-minio | ||||
|     command: server --console-address ":9001" /data | ||||
|     ports: | ||||
|       - ${MINIO_PORT}:9000 | ||||
|       - ${MINIO_CONSOLE_PORT}:9001 | ||||
|     env_file: .env | ||||
|     environment: | ||||
|       - MINIO_ROOT_USER=${MINIO_USER} | ||||
|       - MINIO_ROOT_PASSWORD=${MINIO_PASSWORD} | ||||
|       - TZ=${TIMEZONE} | ||||
|     volumes: | ||||
|       - /home/deepgeek/data/data_local/server/ragflow/ragflow-docker/data/minio_data:/data | ||||
|     networks: | ||||
|       - ragflow | ||||
|     restart: on-failure | ||||
|  | ||||
|   redis: | ||||
|     # swr.cn-north-4.myhuaweicloud.com/ddn-k8s/docker.io/valkey/valkey:8 | ||||
|     image: docker.citory.tech/mirror/valkey/valkey:8 | ||||
|     container_name: ragflow-redis | ||||
|     command: redis-server --requirepass ${REDIS_PASSWORD} --maxmemory 128mb --maxmemory-policy allkeys-lru | ||||
|     env_file: .env | ||||
|     ports: | ||||
|       - ${REDIS_PORT}:6379 | ||||
|     volumes: | ||||
|       - /home/deepgeek/data/data_local/server/ragflow/ragflow-docker/data/redis_data:/data | ||||
|     networks: | ||||
|       - ragflow | ||||
|     restart: on-failure | ||||
|      | ||||
|   ragflow: | ||||
|     depends_on: | ||||
|       mysql: | ||||
|         condition: service_healthy | ||||
|     image: ${RAGFLOW_IMAGE} | ||||
|     container_name: ragflow-server | ||||
|     ports: | ||||
|       - ${SVR_HTTP_PORT}:9380 | ||||
|       - ${HTTP_PORT}:80 | ||||
|       # - 443:443 | ||||
|     volumes: | ||||
|       - /home/deepgeek/data/data_local/server/ragflow/ragflow-docker/ragflow-logs:/ragflow/logs | ||||
|       - /home/deepgeek/data/data_local/server/ragflow/ragflow-docker/nginx/ragflow.conf:/etc/nginx/conf.d/ragflow.conf | ||||
|       - /home/deepgeek/data/data_local/server/ragflow/ragflow-docker/nginx/proxy.conf:/etc/nginx/proxy.conf | ||||
|       - /home/deepgeek/data/data_local/server/ragflow/ragflow-docker/nginx/nginx.conf:/etc/nginx/nginx.conf | ||||
|     env_file: .env | ||||
|     environment: | ||||
|       - TZ=${TIMEZONE} | ||||
|       - HF_ENDPOINT=${HF_ENDPOINT} | ||||
|       - MACOS=${MACOS} | ||||
|     networks: | ||||
|       - ragflow | ||||
|     restart: on-failure | ||||
|     # https://docs.docker.com/engine/daemon/prometheus/#create-a-prometheus-configuration | ||||
|     # If you're using Docker Desktop, the --add-host flag is optional. This flag makes sure that the host's internal IP gets exposed to the Prometheus container. | ||||
|     extra_hosts: | ||||
|       - "host.docker.internal:host-gateway" | ||||
|   # executor: | ||||
|   #   depends_on: | ||||
|   #     mysql: | ||||
|   #       condition: service_healthy | ||||
|   #   image: ${RAGFLOW_IMAGE} | ||||
|   #   container_name: ragflow-executor | ||||
|   #   volumes: | ||||
|   #     - ./ragflow-logs:/ragflow/logs | ||||
|   #     - ./nginx/ragflow.conf:/etc/nginx/conf.d/ragflow.conf | ||||
|   #   env_file: .env | ||||
|   #   environment: | ||||
|   #     - TZ=${TIMEZONE} | ||||
|   #     - HF_ENDPOINT=${HF_ENDPOINT} | ||||
|   #     - MACOS=${MACOS} | ||||
|   #   entrypoint: "/ragflow/entrypoint_task_executor.sh 1 3" | ||||
|   #   networks: | ||||
|   #     - ragflow | ||||
|   #   restart: on-failure | ||||
|   #   # https://docs.docker.com/engine/daemon/prometheus/#create-a-prometheus-configuration | ||||
|   #   # If you're using Docker Desktop, the --add-host flag is optional. This flag makes sure that the host's internal IP gets exposed to the Prometheus container. | ||||
|   #   extra_hosts: | ||||
|   #     - "host.docker.internal:host-gateway" | ||||
| networks: | ||||
|   ragflow: | ||||
|     driver: bridge | ||||
| @@ -6,11 +6,11 @@ services: | ||||
|     ports: | ||||
|       - 10588:7860 | ||||
|     volumes: | ||||
|       - /home/deepgeek/data/data_local/server/stable-diffusion-webui-docker/data:/data | ||||
|       - /home/deepgeek/data/data_local/server/stable-diffusion-webui-docker/data/config/auto/localizations:/stable-diffusion-webui/localizations | ||||
|       - /home/deepgeek/data/data_local/server/stable-diffusion-webui-docker/output:/output | ||||
|       - /home/deepgeek/data/data_local/server/sd-models/checkpoints:/data/models/Stable-diffusion | ||||
|       - /home/deepgeek/data/data_local/server/sd-models/loras:/data/models/Lora | ||||
|       - /home/deepgeek/data/data_base/stable-diffusion-webui-docker/data:/data | ||||
|       - /home/deepgeek/data/data_base/stable-diffusion-webui-docker/data/config/auto/localizations:/stable-diffusion-webui/localizations | ||||
|       - /home/deepgeek/data/data_base/stable-diffusion-webui-docker/output:/output | ||||
|       - /home/deepgeek/data/data_base/sd-models/checkpoints:/data/models/Stable-diffusion | ||||
|       - /home/deepgeek/data/data_base/sd-models/loras:/data/models/Lora | ||||
|     tty: true | ||||
|     environment: | ||||
|       - CUDA_LAUNCH_BLOCKING=1 | ||||
| @@ -26,7 +26,7 @@ services: | ||||
|                 - compute | ||||
|                 - utility | ||||
|                 - gpu | ||||
|     restart: always | ||||
|     restart: unless-stopped | ||||
|     runtime: nvidia | ||||
| x-dockge: | ||||
|   urls: | ||||
|   | ||||
| @@ -13,7 +13,7 @@ services: | ||||
|               count: all | ||||
|               capabilities: | ||||
|                 - gpu | ||||
|     restart: always | ||||
|     restart: unless-stopped | ||||
|     runtime: nvidia | ||||
| x-dockge: | ||||
|   urls: | ||||
|   | ||||
							
								
								
									
										8
									
								
								vllm-deepseek-r1-32b/compose.yaml
									
									
									
									
									
										
										
										Normal file → Executable file
									
								
							
							
						
						
									
										8
									
								
								vllm-deepseek-r1-32b/compose.yaml
									
									
									
									
									
										
										
										Normal file → Executable file
									
								
							| @@ -17,13 +17,13 @@ services: | ||||
|     runtime: nvidia | ||||
|     volumes: | ||||
|       - /home/deepgeek/data/data_local/server/vllm/models:/models | ||||
|     restart: always | ||||
|     command: --served-model-name deepseek-r1:32b越狱 --model | ||||
|     restart: unless-stopped | ||||
|     command: --served-model-name deepseek-r1:32b --model | ||||
|       /models/huihui-ai/DeepSeek-R1-Distill-Qwen-32B-abliterated | ||||
|       --trust-remote-code --host 0.0.0.0 --port 8080 --max-model-len 8192 | ||||
|       --tensor-parallel-size 2 --gpu_memory_utilization 0.9 --enforce-eager | ||||
|       --dtype auto --swap-space 8 --enable-auto-tool-choice | ||||
|       --tool-call-parser pythonic | ||||
|       --dtype auto --swap-space 8 --enable-auto-tool-choice --tool-call-parser | ||||
|       pythonic | ||||
| x-dockge: | ||||
|   urls: | ||||
|     - http://local.citory.tech:10580 | ||||
|   | ||||
							
								
								
									
										12
									
								
								vllm-deepseek-r1-70b/compose.yaml
									
									
									
									
									
										
										
										Normal file → Executable file
									
								
							
							
						
						
									
										12
									
								
								vllm-deepseek-r1-70b/compose.yaml
									
									
									
									
									
										
										
										Normal file → Executable file
									
								
							| @@ -18,12 +18,12 @@ services: | ||||
|     runtime: nvidia | ||||
|     volumes: | ||||
|       - /home/deepgeek/data/data_local/server/vllm/models:/models | ||||
|     restart: always | ||||
|     command: --served-model-name deepseek-r1:70b越狱 --model | ||||
|       /models/Fasiany/DeepSeek-R1-Distill-Llama-70B-abliterated-GPTQ-int4 | ||||
|       --trust-remote-code --host 0.0.0.0 --port 8080 --enable-auto-tool-choice --tool-call-parser pythonic --max-model-len 8192 | ||||
|       --tensor-parallel-size 2 --gpu_memory_utilization 0.9 --enforce-eager | ||||
|       --dtype auto --swap-space 8  | ||||
|     restart: unless-stopped | ||||
|     command: --served-model-name deepseek-r1:70b  | ||||
|       --model /models/Valdemardi/DeepSeek-R1-Distill-Llama-70B-AWQ | ||||
|       --trust-remote-code --host 0.0.0.0 --port 8080 --enable-auto-tool-choice | ||||
|       --tool-call-parser pythonic --max-model-len 8192 --tensor-parallel-size 2 | ||||
|       --gpu_memory_utilization 0.9 --enforce-eager --dtype auto --swap-space 8 | ||||
| x-dockge: | ||||
|   urls: | ||||
|     - http://local.citory.tech:10580 | ||||
|   | ||||
							
								
								
									
										12
									
								
								vllm-qwen25-72b-instruct-awq/compose.yaml
									
									
									
									
									
										
										
										Normal file → Executable file
									
								
							
							
						
						
									
										12
									
								
								vllm-qwen25-72b-instruct-awq/compose.yaml
									
									
									
									
									
										
										
										Normal file → Executable file
									
								
							| @@ -18,12 +18,12 @@ services: | ||||
|     runtime: nvidia | ||||
|     volumes: | ||||
|       - /home/deepgeek/data/data_local/server/vllm/models:/models | ||||
|     restart: always | ||||
|     command: --served-model-name Qwen2.5-72B-Instruct-AWQ --model /models/Qwen/Qwen2.5-72B-Instruct-AWQ | ||||
|       --trust-remote-code --host 0.0.0.0 --port 8080 --max-model-len 8192 | ||||
|       --tensor-parallel-size 2 --gpu_memory_utilization 0.9 --enforce-eager | ||||
|       --dtype auto --swap-space 8 --enable-auto-tool-choice --tool-call-parser | ||||
|       llama3_json | ||||
|     restart: unless-stopped | ||||
|     command: --served-model-name Qwen2.5:72b --model | ||||
|       /models/Qwen/Qwen2.5-72B-Instruct-AWQ --trust-remote-code --host 0.0.0.0 | ||||
|       --port 8080 --max-model-len 8192 --tensor-parallel-size 2 | ||||
|       --gpu_memory_utilization 0.9 --enforce-eager --dtype auto --swap-space 8 | ||||
|       --enable-auto-tool-choice --tool-call-parser llama3_json | ||||
| x-dockge: | ||||
|   urls: | ||||
|     - http://local.citory.tech:10580 | ||||
|   | ||||
							
								
								
									
										4
									
								
								vllm-qwen25-coder-32b-instruct/compose.yaml
									
									
									
									
									
										
										
										Normal file → Executable file
									
								
							
							
						
						
									
										4
									
								
								vllm-qwen25-coder-32b-instruct/compose.yaml
									
									
									
									
									
										
										
										Normal file → Executable file
									
								
							| @@ -18,8 +18,8 @@ services: | ||||
|     runtime: nvidia | ||||
|     volumes: | ||||
|       - /home/deepgeek/data/data_local/server/vllm/models:/models | ||||
|     restart: always | ||||
|     command: --served-model-name Qwen2.5-Coder-32B-Instruct --model | ||||
|     restart: unless-stopped | ||||
|     command: --served-model-name Qwen2.5-Coder:32b --model | ||||
|       /models/Qwen/Qwen2.5-Coder-32B-Instruct --trust-remote-code --host 0.0.0.0 | ||||
|       --port 8080 --max-model-len 8192 --tensor-parallel-size 2 | ||||
|       --gpu_memory_utilization 0.9 --enforce-eager --dtype auto --swap-space 8 | ||||
|   | ||||
							
								
								
									
										4
									
								
								vllm-qwen25-vl-32b-instruct/compose.yaml
									
									
									
									
									
										
										
										Normal file → Executable file
									
								
							
							
						
						
									
										4
									
								
								vllm-qwen25-vl-32b-instruct/compose.yaml
									
									
									
									
									
										
										
										Normal file → Executable file
									
								
							| @@ -18,8 +18,8 @@ services: | ||||
|     runtime: nvidia | ||||
|     volumes: | ||||
|       - /home/deepgeek/data/data_local/server/vllm/models:/models | ||||
|     restart: always | ||||
|     command: --served-model-name Qwen2.5-VL-32B-Instruct --model | ||||
|     restart: unless-stopped | ||||
|     command: --served-model-name Qwen2.5-VL:32b --model | ||||
|       /models/Qwen/Qwen2.5-VL-32B-Instruct --trust-remote-code --host 0.0.0.0 | ||||
|       --port 8080 --max-model-len 8192 --tensor-parallel-size 2 | ||||
|       --gpu_memory_utilization 0.9 --enforce-eager --dtype auto --swap-space 8 | ||||
|   | ||||
							
								
								
									
										8
									
								
								vllm-qwq-32b/compose.yaml
									
									
									
									
									
										
										
										Normal file → Executable file
									
								
							
							
						
						
									
										8
									
								
								vllm-qwq-32b/compose.yaml
									
									
									
									
									
										
										
										Normal file → Executable file
									
								
							| @@ -18,12 +18,12 @@ services: | ||||
|     runtime: nvidia | ||||
|     volumes: | ||||
|       - /home/deepgeek/data/data_local/server/vllm/models:/models | ||||
|     restart: always | ||||
|     command: --served-model-name QwQ-32B --model /models/Qwen/QwQ-32B | ||||
|     restart: unless-stopped | ||||
|     command: --served-model-name QwQ:32b --model /models/Qwen/QwQ-32B | ||||
|       --trust-remote-code --host 0.0.0.0 --port 8080 --max-model-len 8192 | ||||
|       --tensor-parallel-size 2 --gpu_memory_utilization 0.9 --enforce-eager | ||||
|       --dtype auto --swap-space 8 --enable-auto-tool-choice | ||||
|       --tool-call-parser llama3_json | ||||
|       --dtype auto --swap-space 8 --enable-auto-tool-choice --tool-call-parser | ||||
|       llama3_json | ||||
| x-dockge: | ||||
|   urls: | ||||
|     - http://local.citory.tech:10580 | ||||
|   | ||||
		Reference in New Issue
	
	Block a user