feat(*):增加了一些模型
This commit is contained in:
		| @@ -35,7 +35,7 @@ JWT_SECRET="Mnx6P2gPXDz1FngbX3Vmn9SB4T2EVeE4JDkrqM2biA0o6nrWxDNE34QTXzxSR7ToKpTn | |||||||
|  |  | ||||||
| # LLM_PROVIDER='localai' | # LLM_PROVIDER='localai' | ||||||
| LOCAL_AI_BASE_PATH='http://host.docker.internal:10580/v1' | LOCAL_AI_BASE_PATH='http://host.docker.internal:10580/v1' | ||||||
| LOCAL_AI_MODEL_PREF='DeepSeek-R1-Distill-Llama-70B-AWQ' | LOCAL_AI_MODEL_PREF='deepseek-r1:671b' | ||||||
| LOCAL_AI_MODEL_TOKEN_LIMIT=8192 | LOCAL_AI_MODEL_TOKEN_LIMIT=8192 | ||||||
| LOCAL_AI_API_KEY="O8Is3NSYnp5fICWFbhkbwpLWgvMLkdCSuXR5ZggLmgwTKNPEWsjx1NqUxkyU7wLX" | LOCAL_AI_API_KEY="O8Is3NSYnp5fICWFbhkbwpLWgvMLkdCSuXR5ZggLmgwTKNPEWsjx1NqUxkyU7wLX" | ||||||
|  |  | ||||||
|   | |||||||
| @@ -13,7 +13,7 @@ services: | |||||||
|               count: all |               count: all | ||||||
|               capabilities: |               capabilities: | ||||||
|                 - gpu |                 - gpu | ||||||
|     restart: always |     restart: unless-stopped | ||||||
|     command: python webui.py |     command: python webui.py | ||||||
|     runtime: nvidia |     runtime: nvidia | ||||||
| x-dockge: | x-dockge: | ||||||
|   | |||||||
| @@ -1,14 +1,14 @@ | |||||||
| version: "3.8" | version: "3.8" | ||||||
| services: | services: | ||||||
|   comfyui: |   comfyui: | ||||||
|     image: docker.citory.tech/mirror/yanwk/comfyui-boot:cu121 |     image: docker.citory.tech/mirror/yanwk/comfyui-boot:cu124-cn | ||||||
|     container_name: comfyui |     container_name: comfyui | ||||||
|     ports: |     ports: | ||||||
|       - 10587:8188 |       - 10587:8188 | ||||||
|     volumes: |     volumes: | ||||||
|       - /home/deepgeek/data/data_local/server/comfyui/storage:/home/runner |       - /home/deepgeek/data/data_base/comfyui/storage/ComfyUI:/root/ComfyUI | ||||||
|       - /home/deepgeek/data/data_local/server/sd-models/checkpoints:/home/runner/ComfyUI/models/checkpoints |       - /home/deepgeek/data/data_base/sd-models/checkpoints:/root/ComfyUI/models/checkpoints | ||||||
|       - /home/deepgeek/data/data_local/server/sd-models/loras:/home/runner/ComfyUI/models/loras |       - /home/deepgeek/data/data_base/sd-models/loras:/root/ComfyUI/models/loras | ||||||
|     deploy: |     deploy: | ||||||
|       resources: |       resources: | ||||||
|         reservations: |         reservations: | ||||||
| @@ -19,7 +19,7 @@ services: | |||||||
|                 - compute |                 - compute | ||||||
|                 - utility |                 - utility | ||||||
|                 - gpu |                 - gpu | ||||||
|     restart: always |     restart: unless-stopped | ||||||
|     runtime: nvidia |     runtime: nvidia | ||||||
| x-dockge: | x-dockge: | ||||||
|   urls: |   urls: | ||||||
|   | |||||||
							
								
								
									
										1
									
								
								dbgpt/.env
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										1
									
								
								dbgpt/.env
									
									
									
									
									
										Normal file
									
								
							| @@ -0,0 +1 @@ | |||||||
|  | # VARIABLE=value #comment | ||||||
							
								
								
									
										23
									
								
								dbgpt/compose.yaml
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										23
									
								
								dbgpt/compose.yaml
									
									
									
									
									
										Normal file
									
								
							| @@ -0,0 +1,23 @@ | |||||||
|  | services: | ||||||
|  |   dbgpt: | ||||||
|  |     image: docker.citory.tech/mirror/eosphorosai/dbgpt-openai:latest | ||||||
|  |     container_name: dbgpt | ||||||
|  |     environment: | ||||||
|  |       - SILICONFLOW_API_KEY=sk-tlsrjadpcrvebqdmptabfrxznnuiimsawowtpnzbokpanfcx | ||||||
|  |     volumes: | ||||||
|  |       - dbgpt-data:/app/pilot/data | ||||||
|  |       - dbgpt-message:/app/pilot/message | ||||||
|  |     ports: | ||||||
|  |       - 10506:5670/tcp | ||||||
|  |     # webserver may be failed, it must wait all sqls in /docker-entrypoint-initdb.d execute finish. | ||||||
|  |     restart: unless-stopped | ||||||
|  |     ipc: host | ||||||
|  |     tty: true | ||||||
|  |     stdin_open: true | ||||||
|  | volumes: | ||||||
|  |   dbgpt-data: | ||||||
|  |   dbgpt-message: | ||||||
|  | x-dockge: | ||||||
|  |   urls: | ||||||
|  |     - http://local.citory.tech:10506 | ||||||
|  | networks: {} | ||||||
| @@ -13,5 +13,5 @@ services: | |||||||
|               count: all |               count: all | ||||||
|               capabilities: |               capabilities: | ||||||
|                 - gpu |                 - gpu | ||||||
|     restart: always |     restart: unless-stopped | ||||||
|     runtime: nvidia |     runtime: nvidia | ||||||
|   | |||||||
| @@ -18,7 +18,7 @@ services: | |||||||
|     runtime: nvidia |     runtime: nvidia | ||||||
|     volumes: |     volumes: | ||||||
|       - /home/deepgeek/data/data_base/ollama:/root/.ollama |       - /home/deepgeek/data/data_base/ollama:/root/.ollama | ||||||
|     restart: always |     restart: unless-stopped | ||||||
|     command: ollama run gemma3:27b |     command: ollama run gemma3:27b | ||||||
| x-dockge: | x-dockge: | ||||||
|   urls: |   urls: | ||||||
|   | |||||||
| @@ -1,13 +1,13 @@ | |||||||
| version: "3.8" | version: "3.8" | ||||||
| services: | services: | ||||||
|   open-webui: |   open-webui: | ||||||
|     image: docker.citory.tech/mirror/backplane/open-webui:0.5.20 |     image: docker.citory.tech/mirror/backplane/open-webui:0 | ||||||
|     container_name: open-webui |     container_name: open-webui | ||||||
|     ports: |     ports: | ||||||
|       - 10503:8080 |       - 10503:8080 | ||||||
|     environment: |     environment: | ||||||
|       - OPENAI_API_KEY='' # 替换为您的 OpenAI API 密钥 |       - OPENAI_API_KEY='' | ||||||
|       - OPENAI_API_BASE_URL=http://host.docker.internal:10580/v1 # 替换为您的 VLLM IP 和端口 |       - OPENAI_API_BASE_URL=http://host.docker.internal:10580/v1 | ||||||
|       - WHISPER_MODEL=large |       - WHISPER_MODEL=large | ||||||
|       - ENABLE_OLLAMA_API=true |       - ENABLE_OLLAMA_API=true | ||||||
|       - OLLAMA_BASE_URL=http://host.docker.internal:10580 |       - OLLAMA_BASE_URL=http://host.docker.internal:10580 | ||||||
| @@ -27,3 +27,4 @@ services: | |||||||
| x-dockge: | x-dockge: | ||||||
|   urls: |   urls: | ||||||
|     - http://local.citory.tech:10503 |     - http://local.citory.tech:10503 | ||||||
|  | networks: {} | ||||||
|   | |||||||
| @@ -14,7 +14,7 @@ services: | |||||||
|               count: all |               count: all | ||||||
|               capabilities: |               capabilities: | ||||||
|                 - gpu |                 - gpu | ||||||
|     restart: always |     restart: unless-stopped | ||||||
|     runtime: nvidia |     runtime: nvidia | ||||||
| x-dockge: | x-dockge: | ||||||
|   urls: |   urls: | ||||||
|   | |||||||
							
								
								
									
										148
									
								
								ragflow/.env
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										148
									
								
								ragflow/.env
									
									
									
									
									
										Normal file
									
								
							| @@ -0,0 +1,148 @@ | |||||||
|  | # The type of doc engine to use. | ||||||
|  | # Available options: | ||||||
|  | # - `elasticsearch` (default)  | ||||||
|  | # - `infinity` (https://github.com/infiniflow/infinity) | ||||||
|  | DOC_ENGINE=${DOC_ENGINE:-elasticsearch} | ||||||
|  |  | ||||||
|  | # ------------------------------ | ||||||
|  | # docker env var for specifying vector db type at startup | ||||||
|  | # (based on the vector db type, the corresponding docker | ||||||
|  | # compose profile will be used) | ||||||
|  | # ------------------------------ | ||||||
|  | COMPOSE_PROFILES=${DOC_ENGINE} | ||||||
|  |  | ||||||
|  | # The version of Elasticsearch. | ||||||
|  | STACK_VERSION=8.11.3 | ||||||
|  |  | ||||||
|  | # The hostname where the Elasticsearch service is exposed | ||||||
|  | ES_HOST=es01 | ||||||
|  |  | ||||||
|  | # The port used to expose the Elasticsearch service to the host machine,  | ||||||
|  | # allowing EXTERNAL access to the service running inside the Docker container. | ||||||
|  | ES_PORT=1200 | ||||||
|  |  | ||||||
|  | # The password for Elasticsearch.  | ||||||
|  | ELASTIC_PASSWORD=infini_rag_flow | ||||||
|  |  | ||||||
|  | # The port used to expose the Kibana service to the host machine,  | ||||||
|  | # allowing EXTERNAL access to the service running inside the Docker container. | ||||||
|  | KIBANA_PORT=6601 | ||||||
|  | KIBANA_USER=rag_flow | ||||||
|  | KIBANA_PASSWORD=infini_rag_flow | ||||||
|  |  | ||||||
|  | # The maximum amount of the memory, in bytes, that a specific Docker container can use while running. | ||||||
|  | # Update it according to the available memory in the host machine. | ||||||
|  | MEM_LIMIT=8073741824 | ||||||
|  |  | ||||||
|  | # The hostname where the Infinity service is exposed | ||||||
|  | INFINITY_HOST=infinity | ||||||
|  |  | ||||||
|  | # Port to expose Infinity API to the host | ||||||
|  | INFINITY_THRIFT_PORT=23817 | ||||||
|  | INFINITY_HTTP_PORT=23820 | ||||||
|  | INFINITY_PSQL_PORT=5432 | ||||||
|  |  | ||||||
|  | # The password for MySQL.  | ||||||
|  | MYSQL_PASSWORD=infini_rag_flow | ||||||
|  | # The hostname where the MySQL service is exposed | ||||||
|  | MYSQL_HOST=mysql | ||||||
|  | # The database of the MySQL service to use | ||||||
|  | MYSQL_DBNAME=rag_flow | ||||||
|  | # The port used to expose the MySQL service to the host machine,  | ||||||
|  | # allowing EXTERNAL access to the MySQL database running inside the Docker container.  | ||||||
|  | MYSQL_PORT=5455 | ||||||
|  |  | ||||||
|  | # The hostname where the MinIO service is exposed | ||||||
|  | MINIO_HOST=minio | ||||||
|  | # The port used to expose the MinIO console interface to the host machine,  | ||||||
|  | # allowing EXTERNAL access to the web-based console running inside the Docker container.  | ||||||
|  | MINIO_CONSOLE_PORT=9001 | ||||||
|  | # The port used to expose the MinIO API service to the host machine,  | ||||||
|  | # allowing EXTERNAL access to the MinIO object storage service running inside the Docker container.  | ||||||
|  | MINIO_PORT=9000 | ||||||
|  | # The username for MinIO.  | ||||||
|  | # When updated, you must revise the `minio.user` entry in service_conf.yaml accordingly. | ||||||
|  | MINIO_USER=rag_flow | ||||||
|  | # The password for MinIO.  | ||||||
|  | # When updated, you must revise the `minio.password` entry in service_conf.yaml accordingly. | ||||||
|  | MINIO_PASSWORD=infini_rag_flow | ||||||
|  |  | ||||||
|  | # The hostname where the Redis service is exposed | ||||||
|  | REDIS_HOST=redis | ||||||
|  | # The port used to expose the Redis service to the host machine,  | ||||||
|  | # allowing EXTERNAL access to the Redis service running inside the Docker container. | ||||||
|  | REDIS_PORT=6380 | ||||||
|  | # The password for Redis. | ||||||
|  | REDIS_PASSWORD=infini_rag_flow | ||||||
|  |  | ||||||
|  | # The port used to expose RAGFlow's HTTP API service to the host machine,  | ||||||
|  | # allowing EXTERNAL access to the service running inside the Docker container. | ||||||
|  | SVR_HTTP_PORT=9380 | ||||||
|  | HTTP_PORT=10502 | ||||||
|  | # The RAGFlow Docker image to download. | ||||||
|  | # Defaults to the v0.17.2-slim edition, which is the RAGFlow Docker image without embedding models. | ||||||
|  | RAGFLOW_IMAGE=docker.citory.tech/mirror/infiniflow/ragflow:v0.17.2 | ||||||
|  | # | ||||||
|  | # To download the RAGFlow Docker image with embedding models, uncomment the following line instead: | ||||||
|  | # RAGFLOW_IMAGE=infiniflow/ragflow:v0.17.2 | ||||||
|  | #  | ||||||
|  | # The Docker image of the v0.17.2 edition includes: | ||||||
|  | # - Built-in embedding models: | ||||||
|  | #   - BAAI/bge-large-zh-v1.5 | ||||||
|  | #   - BAAI/bge-reranker-v2-m3 | ||||||
|  | #   - maidalun1020/bce-embedding-base_v1 | ||||||
|  | #   - maidalun1020/bce-reranker-base_v1 | ||||||
|  | # - Embedding models that will be downloaded once you select them in the RAGFlow UI: | ||||||
|  | #   - BAAI/bge-base-en-v1.5 | ||||||
|  | #   - BAAI/bge-large-en-v1.5 | ||||||
|  | #   - BAAI/bge-small-en-v1.5 | ||||||
|  | #   - BAAI/bge-small-zh-v1.5 | ||||||
|  | #   - jinaai/jina-embeddings-v2-base-en | ||||||
|  | #   - jinaai/jina-embeddings-v2-small-en | ||||||
|  | #   - nomic-ai/nomic-embed-text-v1.5 | ||||||
|  | #   - sentence-transformers/all-MiniLM-L6-v2 | ||||||
|  | # | ||||||
|  | #  | ||||||
|  |  | ||||||
|  |  | ||||||
|  | # If you cannot download the RAGFlow Docker image: | ||||||
|  | # | ||||||
|  | # - For the `nightly-slim` edition, uncomment either of the following: | ||||||
|  | # RAGFLOW_IMAGE=swr.cn-north-4.myhuaweicloud.com/infiniflow/ragflow:nightly-slim | ||||||
|  | # RAGFLOW_IMAGE=registry.cn-hangzhou.aliyuncs.com/infiniflow/ragflow:nightly-slim | ||||||
|  | # | ||||||
|  | # - For the `nightly` edition, uncomment either of the following: | ||||||
|  | # RAGFLOW_IMAGE=swr.cn-north-4.myhuaweicloud.com/infiniflow/ragflow:nightly | ||||||
|  | # RAGFLOW_IMAGE=registry.cn-hangzhou.aliyuncs.com/infiniflow/ragflow:nightly | ||||||
|  |  | ||||||
|  | # The local time zone. | ||||||
|  | TIMEZONE='Asia/Shanghai' | ||||||
|  |  | ||||||
|  | # Uncomment the following line if you have limited access to huggingface.co: | ||||||
|  | HF_ENDPOINT=https://hf-mirror.com | ||||||
|  |  | ||||||
|  | # Optimizations for MacOS | ||||||
|  | # Uncomment the following line if your operating system is MacOS: | ||||||
|  | # MACOS=1 | ||||||
|  |  | ||||||
|  | # The maximum file size for each uploaded file, in bytes. | ||||||
|  | # You can uncomment this line and update the value if you wish to change the 1024M file size limit | ||||||
|  | # MAX_CONTENT_LENGTH=1073741824 | ||||||
|  | # After making the change, ensure you update `client_max_body_size` in nginx/nginx.conf correspondingly. | ||||||
|  |  | ||||||
|  | # The log level for the RAGFlow's owned packages and imported packages. | ||||||
|  | # Available level: | ||||||
|  | # - `DEBUG` | ||||||
|  | # - `INFO` (default) | ||||||
|  | # - `WARNING` | ||||||
|  | # - `ERROR` | ||||||
|  | # For example, following line changes the log level of `ragflow.es_conn` to `DEBUG`: | ||||||
|  | # LOG_LEVELS=ragflow.es_conn=DEBUG | ||||||
|  |  | ||||||
|  | # aliyun OSS configuration | ||||||
|  | # STORAGE_IMPL=OSS | ||||||
|  | # ACCESS_KEY=xxx | ||||||
|  | # SECRET_KEY=eee | ||||||
|  | # ENDPOINT=http://oss-cn-hangzhou.aliyuncs.com | ||||||
|  | # REGION=cn-hangzhou | ||||||
|  | # BUCKET=ragflow65536 | ||||||
							
								
								
									
										181
									
								
								ragflow/compose.yaml
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										181
									
								
								ragflow/compose.yaml
									
									
									
									
									
										Normal file
									
								
							| @@ -0,0 +1,181 @@ | |||||||
|  | services: | ||||||
|  |   es01: | ||||||
|  |     container_name: ragflow-es-01 | ||||||
|  |     profiles: | ||||||
|  |       - elasticsearch | ||||||
|  |     image: docker.citory.tech/mirror/elasticsearch:${STACK_VERSION} | ||||||
|  |     volumes: | ||||||
|  |       - /home/deepgeek/data/data_local/server/ragflow/ragflow-docker/data/esdata01:/usr/share/elasticsearch/data | ||||||
|  |     ports: | ||||||
|  |       - ${ES_PORT}:9200 | ||||||
|  |     env_file: .env | ||||||
|  |     environment: | ||||||
|  |       - node.name=es01 | ||||||
|  |       - ELASTIC_PASSWORD=${ELASTIC_PASSWORD} | ||||||
|  |       - bootstrap.memory_lock=false | ||||||
|  |       - discovery.type=single-node | ||||||
|  |       - xpack.security.enabled=true | ||||||
|  |       - xpack.security.http.ssl.enabled=false | ||||||
|  |       - xpack.security.transport.ssl.enabled=false | ||||||
|  |       - cluster.routing.allocation.disk.watermark.low=5gb | ||||||
|  |       - cluster.routing.allocation.disk.watermark.high=3gb | ||||||
|  |       - cluster.routing.allocation.disk.watermark.flood_stage=2gb | ||||||
|  |       - TZ=${TIMEZONE} | ||||||
|  |     mem_limit: ${MEM_LIMIT} | ||||||
|  |     ulimits: | ||||||
|  |       memlock: | ||||||
|  |         soft: -1 | ||||||
|  |         hard: -1 | ||||||
|  |     healthcheck: | ||||||
|  |       test: ["CMD-SHELL", "curl http://localhost:9200"] | ||||||
|  |       interval: 10s | ||||||
|  |       timeout: 10s | ||||||
|  |       retries: 120 | ||||||
|  |     networks: | ||||||
|  |       - ragflow | ||||||
|  |     restart: on-failure | ||||||
|  |  | ||||||
|  |   infinity: | ||||||
|  |     container_name: ragflow-infinity | ||||||
|  |     profiles: | ||||||
|  |       - infinity | ||||||
|  |     image: docker.citory.tech/mirror/infiniflow/infinity:v0.6.0-dev3 | ||||||
|  |     volumes: | ||||||
|  |       - /home/deepgeek/data/data_local/server/ragflow/ragflow-docker/data/infinity_data:/var/infinity | ||||||
|  |       - /home/deepgeek/data/data_local/server/ragflow/ragflow-docker/infinity_conf.toml:/infinity_conf.toml | ||||||
|  |     command: ["-f", "/infinity_conf.toml"] | ||||||
|  |     ports: | ||||||
|  |       - ${INFINITY_THRIFT_PORT}:23817 | ||||||
|  |       - ${INFINITY_HTTP_PORT}:23820 | ||||||
|  |       - ${INFINITY_PSQL_PORT}:5432 | ||||||
|  |     env_file: .env | ||||||
|  |     environment: | ||||||
|  |       - TZ=${TIMEZONE} | ||||||
|  |     mem_limit: ${MEM_LIMIT} | ||||||
|  |     ulimits: | ||||||
|  |       nofile: | ||||||
|  |         soft: 500000 | ||||||
|  |         hard: 500000 | ||||||
|  |     networks: | ||||||
|  |       - ragflow | ||||||
|  |     healthcheck: | ||||||
|  |       test: ["CMD", "curl", "http://localhost:23820/admin/node/current"] | ||||||
|  |       interval: 10s | ||||||
|  |       timeout: 10s | ||||||
|  |       retries: 120 | ||||||
|  |     restart: on-failure | ||||||
|  |  | ||||||
|  |  | ||||||
|  |   mysql: | ||||||
|  |     # mysql:5.7 linux/arm64 image is unavailable. | ||||||
|  |     image: docker.citory.tech/mirror/mysql:8.0.39 | ||||||
|  |     container_name: ragflow-mysql | ||||||
|  |     env_file: .env | ||||||
|  |     environment: | ||||||
|  |       - MYSQL_ROOT_PASSWORD=${MYSQL_PASSWORD} | ||||||
|  |       - TZ=${TIMEZONE} | ||||||
|  |     command: | ||||||
|  |       --max_connections=1000 | ||||||
|  |       --character-set-server=utf8mb4 | ||||||
|  |       --collation-server=utf8mb4_unicode_ci | ||||||
|  |       --default-authentication-plugin=mysql_native_password | ||||||
|  |       --tls_version="TLSv1.2,TLSv1.3" | ||||||
|  |       --init-file /data/application/init.sql | ||||||
|  |     ports: | ||||||
|  |       - ${MYSQL_PORT}:3306 | ||||||
|  |     volumes: | ||||||
|  |       - /home/deepgeek/data/data_local/server/ragflow/ragflow-docker/data/mysql_data:/var/lib/mysql | ||||||
|  |       - /home/deepgeek/data/data_local/server/ragflow/ragflow-docker/init.sql:/data/application/init.sql | ||||||
|  |     networks: | ||||||
|  |       - ragflow | ||||||
|  |     healthcheck: | ||||||
|  |       test: ["CMD", "mysqladmin" ,"ping", "-uroot", "-p${MYSQL_PASSWORD}"] | ||||||
|  |       interval: 10s | ||||||
|  |       timeout: 10s | ||||||
|  |       retries: 3 | ||||||
|  |     restart: on-failure | ||||||
|  |  | ||||||
|  |   minio: | ||||||
|  |     image: docker.citory.tech/mirror/minio/minio:RELEASE.2023-12-20T01-00-02Z | ||||||
|  |     container_name: ragflow-minio | ||||||
|  |     command: server --console-address ":9001" /data | ||||||
|  |     ports: | ||||||
|  |       - ${MINIO_PORT}:9000 | ||||||
|  |       - ${MINIO_CONSOLE_PORT}:9001 | ||||||
|  |     env_file: .env | ||||||
|  |     environment: | ||||||
|  |       - MINIO_ROOT_USER=${MINIO_USER} | ||||||
|  |       - MINIO_ROOT_PASSWORD=${MINIO_PASSWORD} | ||||||
|  |       - TZ=${TIMEZONE} | ||||||
|  |     volumes: | ||||||
|  |       - /home/deepgeek/data/data_local/server/ragflow/ragflow-docker/data/minio_data:/data | ||||||
|  |     networks: | ||||||
|  |       - ragflow | ||||||
|  |     restart: on-failure | ||||||
|  |  | ||||||
|  |   redis: | ||||||
|  |     # swr.cn-north-4.myhuaweicloud.com/ddn-k8s/docker.io/valkey/valkey:8 | ||||||
|  |     image: docker.citory.tech/mirror/valkey/valkey:8 | ||||||
|  |     container_name: ragflow-redis | ||||||
|  |     command: redis-server --requirepass ${REDIS_PASSWORD} --maxmemory 128mb --maxmemory-policy allkeys-lru | ||||||
|  |     env_file: .env | ||||||
|  |     ports: | ||||||
|  |       - ${REDIS_PORT}:6379 | ||||||
|  |     volumes: | ||||||
|  |       - /home/deepgeek/data/data_local/server/ragflow/ragflow-docker/data/redis_data:/data | ||||||
|  |     networks: | ||||||
|  |       - ragflow | ||||||
|  |     restart: on-failure | ||||||
|  |      | ||||||
|  |   ragflow: | ||||||
|  |     depends_on: | ||||||
|  |       mysql: | ||||||
|  |         condition: service_healthy | ||||||
|  |     image: ${RAGFLOW_IMAGE} | ||||||
|  |     container_name: ragflow-server | ||||||
|  |     ports: | ||||||
|  |       - ${SVR_HTTP_PORT}:9380 | ||||||
|  |       - ${HTTP_PORT}:80 | ||||||
|  |       # - 443:443 | ||||||
|  |     volumes: | ||||||
|  |       - /home/deepgeek/data/data_local/server/ragflow/ragflow-docker/ragflow-logs:/ragflow/logs | ||||||
|  |       - /home/deepgeek/data/data_local/server/ragflow/ragflow-docker/nginx/ragflow.conf:/etc/nginx/conf.d/ragflow.conf | ||||||
|  |       - /home/deepgeek/data/data_local/server/ragflow/ragflow-docker/nginx/proxy.conf:/etc/nginx/proxy.conf | ||||||
|  |       - /home/deepgeek/data/data_local/server/ragflow/ragflow-docker/nginx/nginx.conf:/etc/nginx/nginx.conf | ||||||
|  |     env_file: .env | ||||||
|  |     environment: | ||||||
|  |       - TZ=${TIMEZONE} | ||||||
|  |       - HF_ENDPOINT=${HF_ENDPOINT} | ||||||
|  |       - MACOS=${MACOS} | ||||||
|  |     networks: | ||||||
|  |       - ragflow | ||||||
|  |     restart: on-failure | ||||||
|  |     # https://docs.docker.com/engine/daemon/prometheus/#create-a-prometheus-configuration | ||||||
|  |     # If you're using Docker Desktop, the --add-host flag is optional. This flag makes sure that the host's internal IP gets exposed to the Prometheus container. | ||||||
|  |     extra_hosts: | ||||||
|  |       - "host.docker.internal:host-gateway" | ||||||
|  |   # executor: | ||||||
|  |   #   depends_on: | ||||||
|  |   #     mysql: | ||||||
|  |   #       condition: service_healthy | ||||||
|  |   #   image: ${RAGFLOW_IMAGE} | ||||||
|  |   #   container_name: ragflow-executor | ||||||
|  |   #   volumes: | ||||||
|  |   #     - ./ragflow-logs:/ragflow/logs | ||||||
|  |   #     - ./nginx/ragflow.conf:/etc/nginx/conf.d/ragflow.conf | ||||||
|  |   #   env_file: .env | ||||||
|  |   #   environment: | ||||||
|  |   #     - TZ=${TIMEZONE} | ||||||
|  |   #     - HF_ENDPOINT=${HF_ENDPOINT} | ||||||
|  |   #     - MACOS=${MACOS} | ||||||
|  |   #   entrypoint: "/ragflow/entrypoint_task_executor.sh 1 3" | ||||||
|  |   #   networks: | ||||||
|  |   #     - ragflow | ||||||
|  |   #   restart: on-failure | ||||||
|  |   #   # https://docs.docker.com/engine/daemon/prometheus/#create-a-prometheus-configuration | ||||||
|  |   #   # If you're using Docker Desktop, the --add-host flag is optional. This flag makes sure that the host's internal IP gets exposed to the Prometheus container. | ||||||
|  |   #   extra_hosts: | ||||||
|  |   #     - "host.docker.internal:host-gateway" | ||||||
|  | networks: | ||||||
|  |   ragflow: | ||||||
|  |     driver: bridge | ||||||
| @@ -6,11 +6,11 @@ services: | |||||||
|     ports: |     ports: | ||||||
|       - 10588:7860 |       - 10588:7860 | ||||||
|     volumes: |     volumes: | ||||||
|       - /home/deepgeek/data/data_local/server/stable-diffusion-webui-docker/data:/data |       - /home/deepgeek/data/data_base/stable-diffusion-webui-docker/data:/data | ||||||
|       - /home/deepgeek/data/data_local/server/stable-diffusion-webui-docker/data/config/auto/localizations:/stable-diffusion-webui/localizations |       - /home/deepgeek/data/data_base/stable-diffusion-webui-docker/data/config/auto/localizations:/stable-diffusion-webui/localizations | ||||||
|       - /home/deepgeek/data/data_local/server/stable-diffusion-webui-docker/output:/output |       - /home/deepgeek/data/data_base/stable-diffusion-webui-docker/output:/output | ||||||
|       - /home/deepgeek/data/data_local/server/sd-models/checkpoints:/data/models/Stable-diffusion |       - /home/deepgeek/data/data_base/sd-models/checkpoints:/data/models/Stable-diffusion | ||||||
|       - /home/deepgeek/data/data_local/server/sd-models/loras:/data/models/Lora |       - /home/deepgeek/data/data_base/sd-models/loras:/data/models/Lora | ||||||
|     tty: true |     tty: true | ||||||
|     environment: |     environment: | ||||||
|       - CUDA_LAUNCH_BLOCKING=1 |       - CUDA_LAUNCH_BLOCKING=1 | ||||||
| @@ -26,7 +26,7 @@ services: | |||||||
|                 - compute |                 - compute | ||||||
|                 - utility |                 - utility | ||||||
|                 - gpu |                 - gpu | ||||||
|     restart: always |     restart: unless-stopped | ||||||
|     runtime: nvidia |     runtime: nvidia | ||||||
| x-dockge: | x-dockge: | ||||||
|   urls: |   urls: | ||||||
|   | |||||||
| @@ -13,7 +13,7 @@ services: | |||||||
|               count: all |               count: all | ||||||
|               capabilities: |               capabilities: | ||||||
|                 - gpu |                 - gpu | ||||||
|     restart: always |     restart: unless-stopped | ||||||
|     runtime: nvidia |     runtime: nvidia | ||||||
| x-dockge: | x-dockge: | ||||||
|   urls: |   urls: | ||||||
|   | |||||||
							
								
								
									
										8
									
								
								vllm-deepseek-r1-32b/compose.yaml
									
									
									
									
									
										
										
										Normal file → Executable file
									
								
							
							
						
						
									
										8
									
								
								vllm-deepseek-r1-32b/compose.yaml
									
									
									
									
									
										
										
										Normal file → Executable file
									
								
							| @@ -17,13 +17,13 @@ services: | |||||||
|     runtime: nvidia |     runtime: nvidia | ||||||
|     volumes: |     volumes: | ||||||
|       - /home/deepgeek/data/data_local/server/vllm/models:/models |       - /home/deepgeek/data/data_local/server/vllm/models:/models | ||||||
|     restart: always |     restart: unless-stopped | ||||||
|     command: --served-model-name deepseek-r1:32b越狱 --model |     command: --served-model-name deepseek-r1:32b --model | ||||||
|       /models/huihui-ai/DeepSeek-R1-Distill-Qwen-32B-abliterated |       /models/huihui-ai/DeepSeek-R1-Distill-Qwen-32B-abliterated | ||||||
|       --trust-remote-code --host 0.0.0.0 --port 8080 --max-model-len 8192 |       --trust-remote-code --host 0.0.0.0 --port 8080 --max-model-len 8192 | ||||||
|       --tensor-parallel-size 2 --gpu_memory_utilization 0.9 --enforce-eager |       --tensor-parallel-size 2 --gpu_memory_utilization 0.9 --enforce-eager | ||||||
|       --dtype auto --swap-space 8 --enable-auto-tool-choice |       --dtype auto --swap-space 8 --enable-auto-tool-choice --tool-call-parser | ||||||
|       --tool-call-parser pythonic |       pythonic | ||||||
| x-dockge: | x-dockge: | ||||||
|   urls: |   urls: | ||||||
|     - http://local.citory.tech:10580 |     - http://local.citory.tech:10580 | ||||||
|   | |||||||
							
								
								
									
										12
									
								
								vllm-deepseek-r1-70b/compose.yaml
									
									
									
									
									
										
										
										Normal file → Executable file
									
								
							
							
						
						
									
										12
									
								
								vllm-deepseek-r1-70b/compose.yaml
									
									
									
									
									
										
										
										Normal file → Executable file
									
								
							| @@ -18,12 +18,12 @@ services: | |||||||
|     runtime: nvidia |     runtime: nvidia | ||||||
|     volumes: |     volumes: | ||||||
|       - /home/deepgeek/data/data_local/server/vllm/models:/models |       - /home/deepgeek/data/data_local/server/vllm/models:/models | ||||||
|     restart: always |     restart: unless-stopped | ||||||
|     command: --served-model-name deepseek-r1:70b越狱 --model |     command: --served-model-name deepseek-r1:70b  | ||||||
|       /models/Fasiany/DeepSeek-R1-Distill-Llama-70B-abliterated-GPTQ-int4 |       --model /models/Valdemardi/DeepSeek-R1-Distill-Llama-70B-AWQ | ||||||
|       --trust-remote-code --host 0.0.0.0 --port 8080 --enable-auto-tool-choice --tool-call-parser pythonic --max-model-len 8192 |       --trust-remote-code --host 0.0.0.0 --port 8080 --enable-auto-tool-choice | ||||||
|       --tensor-parallel-size 2 --gpu_memory_utilization 0.9 --enforce-eager |       --tool-call-parser pythonic --max-model-len 8192 --tensor-parallel-size 2 | ||||||
|       --dtype auto --swap-space 8  |       --gpu_memory_utilization 0.9 --enforce-eager --dtype auto --swap-space 8 | ||||||
| x-dockge: | x-dockge: | ||||||
|   urls: |   urls: | ||||||
|     - http://local.citory.tech:10580 |     - http://local.citory.tech:10580 | ||||||
|   | |||||||
							
								
								
									
										12
									
								
								vllm-qwen25-72b-instruct-awq/compose.yaml
									
									
									
									
									
										
										
										Normal file → Executable file
									
								
							
							
						
						
									
										12
									
								
								vllm-qwen25-72b-instruct-awq/compose.yaml
									
									
									
									
									
										
										
										Normal file → Executable file
									
								
							| @@ -18,12 +18,12 @@ services: | |||||||
|     runtime: nvidia |     runtime: nvidia | ||||||
|     volumes: |     volumes: | ||||||
|       - /home/deepgeek/data/data_local/server/vllm/models:/models |       - /home/deepgeek/data/data_local/server/vllm/models:/models | ||||||
|     restart: always |     restart: unless-stopped | ||||||
|     command: --served-model-name Qwen2.5-72B-Instruct-AWQ --model /models/Qwen/Qwen2.5-72B-Instruct-AWQ |     command: --served-model-name Qwen2.5:72b --model | ||||||
|       --trust-remote-code --host 0.0.0.0 --port 8080 --max-model-len 8192 |       /models/Qwen/Qwen2.5-72B-Instruct-AWQ --trust-remote-code --host 0.0.0.0 | ||||||
|       --tensor-parallel-size 2 --gpu_memory_utilization 0.9 --enforce-eager |       --port 8080 --max-model-len 8192 --tensor-parallel-size 2 | ||||||
|       --dtype auto --swap-space 8 --enable-auto-tool-choice --tool-call-parser |       --gpu_memory_utilization 0.9 --enforce-eager --dtype auto --swap-space 8 | ||||||
|       llama3_json |       --enable-auto-tool-choice --tool-call-parser llama3_json | ||||||
| x-dockge: | x-dockge: | ||||||
|   urls: |   urls: | ||||||
|     - http://local.citory.tech:10580 |     - http://local.citory.tech:10580 | ||||||
|   | |||||||
							
								
								
									
										4
									
								
								vllm-qwen25-coder-32b-instruct/compose.yaml
									
									
									
									
									
										
										
										Normal file → Executable file
									
								
							
							
						
						
									
										4
									
								
								vllm-qwen25-coder-32b-instruct/compose.yaml
									
									
									
									
									
										
										
										Normal file → Executable file
									
								
							| @@ -18,8 +18,8 @@ services: | |||||||
|     runtime: nvidia |     runtime: nvidia | ||||||
|     volumes: |     volumes: | ||||||
|       - /home/deepgeek/data/data_local/server/vllm/models:/models |       - /home/deepgeek/data/data_local/server/vllm/models:/models | ||||||
|     restart: always |     restart: unless-stopped | ||||||
|     command: --served-model-name Qwen2.5-Coder-32B-Instruct --model |     command: --served-model-name Qwen2.5-Coder:32b --model | ||||||
|       /models/Qwen/Qwen2.5-Coder-32B-Instruct --trust-remote-code --host 0.0.0.0 |       /models/Qwen/Qwen2.5-Coder-32B-Instruct --trust-remote-code --host 0.0.0.0 | ||||||
|       --port 8080 --max-model-len 8192 --tensor-parallel-size 2 |       --port 8080 --max-model-len 8192 --tensor-parallel-size 2 | ||||||
|       --gpu_memory_utilization 0.9 --enforce-eager --dtype auto --swap-space 8 |       --gpu_memory_utilization 0.9 --enforce-eager --dtype auto --swap-space 8 | ||||||
|   | |||||||
							
								
								
									
										4
									
								
								vllm-qwen25-vl-32b-instruct/compose.yaml
									
									
									
									
									
										
										
										Normal file → Executable file
									
								
							
							
						
						
									
										4
									
								
								vllm-qwen25-vl-32b-instruct/compose.yaml
									
									
									
									
									
										
										
										Normal file → Executable file
									
								
							| @@ -18,8 +18,8 @@ services: | |||||||
|     runtime: nvidia |     runtime: nvidia | ||||||
|     volumes: |     volumes: | ||||||
|       - /home/deepgeek/data/data_local/server/vllm/models:/models |       - /home/deepgeek/data/data_local/server/vllm/models:/models | ||||||
|     restart: always |     restart: unless-stopped | ||||||
|     command: --served-model-name Qwen2.5-VL-32B-Instruct --model |     command: --served-model-name Qwen2.5-VL:32b --model | ||||||
|       /models/Qwen/Qwen2.5-VL-32B-Instruct --trust-remote-code --host 0.0.0.0 |       /models/Qwen/Qwen2.5-VL-32B-Instruct --trust-remote-code --host 0.0.0.0 | ||||||
|       --port 8080 --max-model-len 8192 --tensor-parallel-size 2 |       --port 8080 --max-model-len 8192 --tensor-parallel-size 2 | ||||||
|       --gpu_memory_utilization 0.9 --enforce-eager --dtype auto --swap-space 8 |       --gpu_memory_utilization 0.9 --enforce-eager --dtype auto --swap-space 8 | ||||||
|   | |||||||
							
								
								
									
										8
									
								
								vllm-qwq-32b/compose.yaml
									
									
									
									
									
										
										
										Normal file → Executable file
									
								
							
							
						
						
									
										8
									
								
								vllm-qwq-32b/compose.yaml
									
									
									
									
									
										
										
										Normal file → Executable file
									
								
							| @@ -18,12 +18,12 @@ services: | |||||||
|     runtime: nvidia |     runtime: nvidia | ||||||
|     volumes: |     volumes: | ||||||
|       - /home/deepgeek/data/data_local/server/vllm/models:/models |       - /home/deepgeek/data/data_local/server/vllm/models:/models | ||||||
|     restart: always |     restart: unless-stopped | ||||||
|     command: --served-model-name QwQ-32B --model /models/Qwen/QwQ-32B |     command: --served-model-name QwQ:32b --model /models/Qwen/QwQ-32B | ||||||
|       --trust-remote-code --host 0.0.0.0 --port 8080 --max-model-len 8192 |       --trust-remote-code --host 0.0.0.0 --port 8080 --max-model-len 8192 | ||||||
|       --tensor-parallel-size 2 --gpu_memory_utilization 0.9 --enforce-eager |       --tensor-parallel-size 2 --gpu_memory_utilization 0.9 --enforce-eager | ||||||
|       --dtype auto --swap-space 8 --enable-auto-tool-choice |       --dtype auto --swap-space 8 --enable-auto-tool-choice --tool-call-parser | ||||||
|       --tool-call-parser llama3_json |       llama3_json | ||||||
| x-dockge: | x-dockge: | ||||||
|   urls: |   urls: | ||||||
|     - http://local.citory.tech:10580 |     - http://local.citory.tech:10580 | ||||||
|   | |||||||
		Reference in New Issue
	
	Block a user