init(*):初始化;使用sun-panel作为前端;使用dockge管理dockercompose
329
anythinllm/.env
Normal file
@ -0,0 +1,329 @@
|
||||
|
||||
SERVER_PORT=3001
|
||||
STORAGE_DIR="/app/server/storage"
|
||||
UID='1000'
|
||||
GID='1000'
|
||||
# SIG_KEY='passphrase' # Please generate random string at least 32 chars long.
|
||||
# SIG_SALT='salt' # Please generate random string at least 32 chars long.
|
||||
JWT_SECRET="Mnx6P2gPXDz1FngbX3Vmn9SB4T2EVeE4JDkrqM2biA0o6nrWxDNE34QTXzxSR7ToKpTnjU6Qk2rGc3UGc3C03XL0w3gsXoUkA7kje1A82f2V2bTchu4N64uPkljZmF1x" # Only needed if AUTH_TOKEN is set. Please generate random string at least 12 chars long.
|
||||
|
||||
###########################################
|
||||
######## LLM API SElECTION ################
|
||||
###########################################
|
||||
# LLM_PROVIDER='openai'
|
||||
# OPEN_AI_KEY=
|
||||
# OPEN_MODEL_PREF='gpt-4o'
|
||||
|
||||
# LLM_PROVIDER='gemini'
|
||||
# GEMINI_API_KEY=
|
||||
# GEMINI_LLM_MODEL_PREF='gemini-pro'
|
||||
|
||||
# LLM_PROVIDER='azure'
|
||||
# AZURE_OPENAI_ENDPOINT=
|
||||
# AZURE_OPENAI_KEY=
|
||||
# OPEN_MODEL_PREF='my-gpt35-deployment' # This is the "deployment" on Azure you want to use. Not the base model.
|
||||
# EMBEDDING_MODEL_PREF='embedder-model' # This is the "deployment" on Azure you want to use for embeddings. Not the base model. Valid base model is text-embedding-ada-002
|
||||
|
||||
# LLM_PROVIDER='anthropic'
|
||||
# ANTHROPIC_API_KEY=sk-ant-xxxx
|
||||
# ANTHROPIC_MODEL_PREF='claude-2'
|
||||
|
||||
# LLM_PROVIDER='lmstudio'
|
||||
# LMSTUDIO_BASE_PATH='http://your-server:1234/v1'
|
||||
# LMSTUDIO_MODEL_PREF='Loaded from Chat UI' # this is a bug in LMStudio 0.2.17
|
||||
# LMSTUDIO_MODEL_TOKEN_LIMIT=4096
|
||||
|
||||
# LLM_PROVIDER='localai'
|
||||
LOCAL_AI_BASE_PATH='http://host.docker.internal:10580/v1'
|
||||
LOCAL_AI_MODEL_PREF='DeepSeek-R1-Distill-Llama-70B-AWQ'
|
||||
LOCAL_AI_MODEL_TOKEN_LIMIT=8192
|
||||
LOCAL_AI_API_KEY="O8Is3NSYnp5fICWFbhkbwpLWgvMLkdCSuXR5ZggLmgwTKNPEWsjx1NqUxkyU7wLX"
|
||||
|
||||
# LLM_PROVIDER='ollama'
|
||||
# OLLAMA_BASE_PATH='http://host.docker.internal:11434'
|
||||
# OLLAMA_MODEL_PREF='llama2'
|
||||
# OLLAMA_MODEL_TOKEN_LIMIT=4096
|
||||
# OLLAMA_AUTH_TOKEN='your-ollama-auth-token-here (optional, only for ollama running behind auth - Bearer token)'
|
||||
|
||||
# LLM_PROVIDER='togetherai'
|
||||
# TOGETHER_AI_API_KEY='my-together-ai-key'
|
||||
# TOGETHER_AI_MODEL_PREF='mistralai/Mixtral-8x7B-Instruct-v0.1'
|
||||
|
||||
# LLM_PROVIDER='mistral'
|
||||
# MISTRAL_API_KEY='example-mistral-ai-api-key'
|
||||
# MISTRAL_MODEL_PREF='mistral-tiny'
|
||||
|
||||
# LLM_PROVIDER='perplexity'
|
||||
# PERPLEXITY_API_KEY='my-perplexity-key'
|
||||
# PERPLEXITY_MODEL_PREF='codellama-34b-instruct'
|
||||
|
||||
# LLM_PROVIDER='openrouter'
|
||||
# OPENROUTER_API_KEY='my-openrouter-key'
|
||||
# OPENROUTER_MODEL_PREF='openrouter/auto'
|
||||
|
||||
# LLM_PROVIDER='huggingface'
|
||||
# HUGGING_FACE_LLM_ENDPOINT=https://uuid-here.us-east-1.aws.endpoints.huggingface.cloud
|
||||
# HUGGING_FACE_LLM_API_KEY=hf_xxxxxx
|
||||
# HUGGING_FACE_LLM_TOKEN_LIMIT=8000
|
||||
|
||||
# LLM_PROVIDER='groq'
|
||||
# GROQ_API_KEY=gsk_abcxyz
|
||||
# GROQ_MODEL_PREF=llama3-8b-8192
|
||||
|
||||
# LLM_PROVIDER='koboldcpp'
|
||||
# KOBOLD_CPP_BASE_PATH='http://127.0.0.1:5000/v1'
|
||||
# KOBOLD_CPP_MODEL_PREF='koboldcpp/codellama-7b-instruct.Q4_K_S'
|
||||
# KOBOLD_CPP_MODEL_TOKEN_LIMIT=4096
|
||||
|
||||
# LLM_PROVIDER='textgenwebui'
|
||||
# TEXT_GEN_WEB_UI_BASE_PATH='http://127.0.0.1:5000/v1'
|
||||
# TEXT_GEN_WEB_UI_TOKEN_LIMIT=4096
|
||||
# TEXT_GEN_WEB_UI_API_KEY='sk-123abc'
|
||||
|
||||
# LLM_PROVIDER='generic-openai'
|
||||
# GENERIC_OPEN_AI_BASE_PATH='http://proxy.url.openai.com/v1'
|
||||
# GENERIC_OPEN_AI_MODEL_PREF='gpt-3.5-turbo'
|
||||
# GENERIC_OPEN_AI_MODEL_TOKEN_LIMIT=4096
|
||||
# GENERIC_OPEN_AI_API_KEY=sk-123abc
|
||||
|
||||
# LLM_PROVIDER='litellm'
|
||||
# LITE_LLM_MODEL_PREF='gpt-3.5-turbo'
|
||||
# LITE_LLM_MODEL_TOKEN_LIMIT=4096
|
||||
# LITE_LLM_BASE_PATH='http://127.0.0.1:4000'
|
||||
# LITE_LLM_API_KEY='sk-123abc'
|
||||
|
||||
# LLM_PROVIDER='novita'
|
||||
# NOVITA_LLM_API_KEY='your-novita-api-key-here' check on https://novita.ai/settings/key-management
|
||||
# NOVITA_LLM_MODEL_PREF='deepseek/deepseek-r1'
|
||||
|
||||
# LLM_PROVIDER='cohere'
|
||||
# COHERE_API_KEY=
|
||||
# COHERE_MODEL_PREF='command-r'
|
||||
|
||||
# LLM_PROVIDER='bedrock'
|
||||
# AWS_BEDROCK_LLM_ACCESS_KEY_ID=
|
||||
# AWS_BEDROCK_LLM_ACCESS_KEY=
|
||||
# AWS_BEDROCK_LLM_REGION=us-west-2
|
||||
# AWS_BEDROCK_LLM_MODEL_PREFERENCE=meta.llama3-1-8b-instruct-v1:0
|
||||
# AWS_BEDROCK_LLM_MODEL_TOKEN_LIMIT=8191
|
||||
|
||||
# LLM_PROVIDER='fireworksai'
|
||||
# FIREWORKS_AI_LLM_API_KEY='my-fireworks-ai-key'
|
||||
# FIREWORKS_AI_LLM_MODEL_PREF='accounts/fireworks/models/llama-v3p1-8b-instruct'
|
||||
|
||||
# LLM_PROVIDER='apipie'
|
||||
# APIPIE_LLM_API_KEY='sk-123abc'
|
||||
# APIPIE_LLM_MODEL_PREF='openrouter/llama-3.1-8b-instruct'
|
||||
|
||||
# LLM_PROVIDER='xai'
|
||||
# XAI_LLM_API_KEY='xai-your-api-key-here'
|
||||
# XAI_LLM_MODEL_PREF='grok-beta'
|
||||
|
||||
# LLM_PROVIDER='nvidia-nim'
|
||||
# NVIDIA_NIM_LLM_BASE_PATH='http://127.0.0.1:8000'
|
||||
# NVIDIA_NIM_LLM_MODEL_PREF='meta/llama-3.2-3b-instruct'
|
||||
|
||||
# LLM_PROVIDER='deepseek'
|
||||
# DEEPSEEK_API_KEY='your-deepseek-api-key-here'
|
||||
# DEEPSEEK_MODEL_PREF='deepseek-chat'
|
||||
|
||||
# LLM_PROVIDER='ppio'
|
||||
# PPIO_API_KEY='your-ppio-api-key-here'
|
||||
# PPIO_MODEL_PREF=deepseek/deepseek-v3/community
|
||||
|
||||
###########################################
|
||||
######## Embedding API SElECTION ##########
|
||||
###########################################
|
||||
# Only used if you are using an LLM that does not natively support embedding (openai or Azure)
|
||||
# EMBEDDING_ENGINE='openai'
|
||||
# OPEN_AI_KEY=sk-xxxx
|
||||
# EMBEDDING_MODEL_PREF='text-embedding-ada-002'
|
||||
|
||||
# EMBEDDING_ENGINE='azure'
|
||||
# AZURE_OPENAI_ENDPOINT=
|
||||
# AZURE_OPENAI_KEY=
|
||||
# EMBEDDING_MODEL_PREF='my-embedder-model' # This is the "deployment" on Azure you want to use for embeddings. Not the base model. Valid base model is text-embedding-ada-002
|
||||
|
||||
# EMBEDDING_ENGINE='localai'
|
||||
# EMBEDDING_BASE_PATH='http://localhost:8080/v1'
|
||||
# EMBEDDING_MODEL_PREF='text-embedding-ada-002'
|
||||
# EMBEDDING_MODEL_MAX_CHUNK_LENGTH=1000 # The max chunk size in chars a string to embed can be
|
||||
|
||||
# EMBEDDING_ENGINE='ollama'
|
||||
# EMBEDDING_BASE_PATH='http://host.docker.internal:11434'
|
||||
# EMBEDDING_MODEL_PREF='nomic-embed-text:latest'
|
||||
# EMBEDDING_MODEL_MAX_CHUNK_LENGTH=8192
|
||||
|
||||
# EMBEDDING_ENGINE='lmstudio'
|
||||
# EMBEDDING_BASE_PATH='https://host.docker.internal:1234/v1'
|
||||
# EMBEDDING_MODEL_PREF='nomic-ai/nomic-embed-text-v1.5-GGUF/nomic-embed-text-v1.5.Q4_0.gguf'
|
||||
# EMBEDDING_MODEL_MAX_CHUNK_LENGTH=8192
|
||||
|
||||
# EMBEDDING_ENGINE='cohere'
|
||||
# COHERE_API_KEY=
|
||||
# EMBEDDING_MODEL_PREF='embed-english-v3.0'
|
||||
|
||||
# EMBEDDING_ENGINE='voyageai'
|
||||
# VOYAGEAI_API_KEY=
|
||||
# EMBEDDING_MODEL_PREF='voyage-large-2-instruct'
|
||||
|
||||
# EMBEDDING_ENGINE='litellm'
|
||||
# EMBEDDING_MODEL_PREF='text-embedding-ada-002'
|
||||
# EMBEDDING_MODEL_MAX_CHUNK_LENGTH=8192
|
||||
# LITE_LLM_BASE_PATH='http://127.0.0.1:4000'
|
||||
# LITE_LLM_API_KEY='sk-123abc'
|
||||
|
||||
# EMBEDDING_ENGINE='generic-openai'
|
||||
# EMBEDDING_MODEL_PREF='text-embedding-ada-002'
|
||||
# EMBEDDING_MODEL_MAX_CHUNK_LENGTH=8192
|
||||
# EMBEDDING_BASE_PATH='http://127.0.0.1:4000'
|
||||
# GENERIC_OPEN_AI_EMBEDDING_API_KEY='sk-123abc'
|
||||
# GENERIC_OPEN_AI_EMBEDDING_MAX_CONCURRENT_CHUNKS=500
|
||||
|
||||
# EMBEDDING_ENGINE='gemini'
|
||||
# GEMINI_EMBEDDING_API_KEY=
|
||||
# EMBEDDING_MODEL_PREF='text-embedding-004'
|
||||
|
||||
###########################################
|
||||
######## Vector Database Selection ########
|
||||
###########################################
|
||||
# Enable all below if you are using vector database: Chroma.
|
||||
# VECTOR_DB="chroma"
|
||||
# CHROMA_ENDPOINT='http://host.docker.internal:8000'
|
||||
# CHROMA_API_HEADER="X-Api-Key"
|
||||
# CHROMA_API_KEY="sk-123abc"
|
||||
|
||||
# Enable all below if you are using vector database: Pinecone.
|
||||
# VECTOR_DB="pinecone"
|
||||
# PINECONE_API_KEY=
|
||||
# PINECONE_INDEX=
|
||||
|
||||
# Enable all below if you are using vector database: LanceDB.
|
||||
# VECTOR_DB="lancedb"
|
||||
|
||||
# Enable all below if you are using vector database: Weaviate.
|
||||
# VECTOR_DB="weaviate"
|
||||
# WEAVIATE_ENDPOINT="http://localhost:8080"
|
||||
# WEAVIATE_API_KEY=
|
||||
|
||||
# Enable all below if you are using vector database: Qdrant.
|
||||
# VECTOR_DB="qdrant"
|
||||
# QDRANT_ENDPOINT="http://localhost:6333"
|
||||
# QDRANT_API_KEY=
|
||||
|
||||
# Enable all below if you are using vector database: Milvus.
|
||||
# VECTOR_DB="milvus"
|
||||
# MILVUS_ADDRESS="http://localhost:19530"
|
||||
# MILVUS_USERNAME=
|
||||
# MILVUS_PASSWORD=
|
||||
|
||||
# Enable all below if you are using vector database: Zilliz Cloud.
|
||||
# VECTOR_DB="zilliz"
|
||||
# ZILLIZ_ENDPOINT="https://sample.api.gcp-us-west1.zillizcloud.com"
|
||||
# ZILLIZ_API_TOKEN=api-token-here
|
||||
|
||||
# Enable all below if you are using vector database: Astra DB.
|
||||
# VECTOR_DB="astra"
|
||||
# ASTRA_DB_APPLICATION_TOKEN=
|
||||
# ASTRA_DB_ENDPOINT=
|
||||
|
||||
###########################################
|
||||
######## Audio Model Selection ############
|
||||
###########################################
|
||||
# (default) use built-in whisper-small model.
|
||||
# WHISPER_PROVIDER="local"
|
||||
|
||||
# use openai hosted whisper model.
|
||||
# WHISPER_PROVIDER="openai"
|
||||
# OPEN_AI_KEY=sk-xxxxxxxx
|
||||
|
||||
###########################################
|
||||
######## TTS/STT Model Selection ##########
|
||||
###########################################
|
||||
# TTS_PROVIDER="native"
|
||||
|
||||
# TTS_PROVIDER="openai"
|
||||
# TTS_OPEN_AI_KEY=sk-example
|
||||
# TTS_OPEN_AI_VOICE_MODEL=nova
|
||||
|
||||
# TTS_PROVIDER="generic-openai"
|
||||
# TTS_OPEN_AI_COMPATIBLE_KEY=sk-example
|
||||
# TTS_OPEN_AI_COMPATIBLE_VOICE_MODEL=nova
|
||||
# TTS_OPEN_AI_COMPATIBLE_ENDPOINT="https://api.openai.com/v1"
|
||||
|
||||
# TTS_PROVIDER="elevenlabs"
|
||||
# TTS_ELEVEN_LABS_KEY=
|
||||
# TTS_ELEVEN_LABS_VOICE_MODEL=21m00Tcm4TlvDq8ikWAM # Rachel
|
||||
|
||||
# CLOUD DEPLOYMENT VARIRABLES ONLY
|
||||
# AUTH_TOKEN="hunter2" # This is the password to your application if remote hosting.
|
||||
# DISABLE_TELEMETRY="false"
|
||||
|
||||
###########################################
|
||||
######## PASSWORD COMPLEXITY ##############
|
||||
###########################################
|
||||
# Enforce a password schema for your organization users.
|
||||
# Documentation on how to use https://github.com/kamronbatman/joi-password-complexity
|
||||
# Default is only 8 char minimum
|
||||
# PASSWORDMINCHAR=8
|
||||
# PASSWORDMAXCHAR=250
|
||||
# PASSWORDLOWERCASE=1
|
||||
# PASSWORDUPPERCASE=1
|
||||
# PASSWORDNUMERIC=1
|
||||
# PASSWORDSYMBOL=1
|
||||
# PASSWORDREQUIREMENTS=4
|
||||
|
||||
###########################################
|
||||
######## ENABLE HTTPS SERVER ##############
|
||||
###########################################
|
||||
# By enabling this and providing the path/filename for the key and cert,
|
||||
# the server will use HTTPS instead of HTTP.
|
||||
#ENABLE_HTTPS="true"
|
||||
#HTTPS_CERT_PATH="sslcert/cert.pem"
|
||||
#HTTPS_KEY_PATH="sslcert/key.pem"
|
||||
|
||||
###########################################
|
||||
######## AGENT SERVICE KEYS ###############
|
||||
###########################################
|
||||
|
||||
#------ SEARCH ENGINES -------
|
||||
#=============================
|
||||
#------ Google Search -------- https://programmablesearchengine.google.com/controlpanel/create
|
||||
# AGENT_GSE_KEY=
|
||||
# AGENT_GSE_CTX=
|
||||
|
||||
#------ SearchApi.io ----------- https://www.searchapi.io/
|
||||
# AGENT_SEARCHAPI_API_KEY=
|
||||
# AGENT_SEARCHAPI_ENGINE=google
|
||||
|
||||
#------ Serper.dev ----------- https://serper.dev/
|
||||
# AGENT_SERPER_DEV_KEY=
|
||||
|
||||
#------ Bing Search ----------- https://portal.azure.com/
|
||||
# AGENT_BING_SEARCH_API_KEY=
|
||||
|
||||
#------ Serply.io ----------- https://serply.io/
|
||||
# AGENT_SERPLY_API_KEY=
|
||||
|
||||
#------ SearXNG ----------- https://github.com/searxng/searxng
|
||||
# AGENT_SEARXNG_API_URL=
|
||||
|
||||
#------ Tavily ----------- https://www.tavily.com/
|
||||
# AGENT_TAVILY_API_KEY=
|
||||
|
||||
###########################################
|
||||
######## Other Configurations ############
|
||||
###########################################
|
||||
|
||||
# Disable viewing chat history from the UI and frontend APIs.
|
||||
# See https://docs.anythingllm.com/configuration#disable-view-chat-history for more information.
|
||||
# DISABLE_VIEW_CHAT_HISTORY=1
|
||||
|
||||
# Enable simple SSO passthrough to pre-authenticate users from a third party service.
|
||||
# See https://docs.anythingllm.com/configuration#simple-sso-passthrough for more information.
|
||||
# SIMPLE_SSO_ENABLED=1
|
||||
|
||||
# Specify the target languages for when using OCR to parse images and PDFs.
|
||||
# This is a comma separated list of language codes as a string. Unsupported languages will be ignored.
|
||||
# Default is English. See https://tesseract-ocr.github.io/tessdoc/Data-Files-in-different-versions.html for a list of valid language codes.
|
||||
# TARGET_OCR_LANG=eng,deu,ita,spa,fra,por,rus,nld,tur,hun,pol,ita,spa,fra,por,rus,nld,tur,hun,pol
|
17
anythinllm/compose.yaml
Normal file
@ -0,0 +1,17 @@
|
||||
version: "3.8"
|
||||
services:
|
||||
anythingllm:
|
||||
image: docker.citory.tech/mirror/mintplexlabs/anythingllm:latest
|
||||
container_name: anythingllm
|
||||
ports:
|
||||
- 10504:3001
|
||||
env_file:
|
||||
- .env
|
||||
cap_add:
|
||||
- SYS_ADMIN
|
||||
extra_hosts:
|
||||
- host.docker.internal:host-gateway
|
||||
restart: always
|
||||
x-dockge:
|
||||
urls:
|
||||
- http://local.citory.tech:10504
|
1
chattts/.env
Normal file
@ -0,0 +1 @@
|
||||
CUDA_VISIBLE_DEVICES=0,1
|
23
chattts/compose.yaml
Normal file
@ -0,0 +1,23 @@
|
||||
version: "3.8"
|
||||
services:
|
||||
lenml-chattts-forge:
|
||||
image: docker.citory.tech/public/lenml-chattts-forge
|
||||
container_name: lenml-chattts-forge
|
||||
ports:
|
||||
- 10583:7860
|
||||
env_file: .env
|
||||
deploy:
|
||||
resources:
|
||||
reservations:
|
||||
devices:
|
||||
- driver: nvidia
|
||||
count: all
|
||||
capabilities:
|
||||
- gpu
|
||||
restart: always
|
||||
command: python webui.py
|
||||
runtime: nvidia
|
||||
x-dockge:
|
||||
urls:
|
||||
- http://local.citory.tech:10583
|
||||
networks: {}
|
1
comfyui/.env
Normal file
@ -0,0 +1 @@
|
||||
# VARIABLE=value #comment
|
27
comfyui/compose.yaml
Normal file
@ -0,0 +1,27 @@
|
||||
version: "3.8"
|
||||
services:
|
||||
comfyui:
|
||||
image: docker.citory.tech/mirror/yanwk/comfyui-boot:cu121
|
||||
container_name: comfyui
|
||||
ports:
|
||||
- 10587:8188
|
||||
volumes:
|
||||
- /home/deepgeek/data/data_local/server/comfyui/storage:/home/runner
|
||||
- /home/deepgeek/data/data_local/server/sd-models/checkpoints:/home/runner/ComfyUI/models/checkpoints
|
||||
- /home/deepgeek/data/data_local/server/sd-models/loras:/home/runner/ComfyUI/models/loras
|
||||
deploy:
|
||||
resources:
|
||||
reservations:
|
||||
devices:
|
||||
- driver: nvidia
|
||||
count: all
|
||||
capabilities:
|
||||
- compute
|
||||
- utility
|
||||
- gpu
|
||||
restart: always
|
||||
runtime: nvidia
|
||||
x-dockge:
|
||||
urls:
|
||||
- http://local.citory.tech:10587
|
||||
networks: {}
|
1
gpustat-web/.env
Normal file
@ -0,0 +1 @@
|
||||
# VARIABLE=value #comment
|
13
gpustat-web/compose.yaml
Normal file
@ -0,0 +1,13 @@
|
||||
version: "3.8"
|
||||
services:
|
||||
gpustat-web:
|
||||
image: docker.citory.tech/public/gpustat-web:10599
|
||||
container_name: gpustat-web
|
||||
ports:
|
||||
- 10509:80
|
||||
extra_hosts:
|
||||
- host.docker.internal:host-gateway
|
||||
restart: always
|
||||
x-dockge:
|
||||
urls:
|
||||
- http://local.citory.tech:10509
|
5
kodbox/.env
Normal file
@ -0,0 +1,5 @@
|
||||
MYSQL_PASSWORD=SFMhkWEZdeb6jxtel0igQGToKyJ7bHwd
|
||||
MYSQL_DATABASE=kodbox
|
||||
MYSQL_USER=kodbox
|
||||
KODBOX_ADMIN_USER=deepgeek
|
||||
KODBOX_ADMIN_PASSWORD=DeepGeek2025
|
39
kodbox/compose.yaml
Normal file
@ -0,0 +1,39 @@
|
||||
version: "3.8"
|
||||
services:
|
||||
kodbox-db:
|
||||
image: docker.citory.tech/mirror/mariadb:10.6
|
||||
container_name: kodbox-db
|
||||
restart: always
|
||||
command: --transaction-isolation=READ-COMMITTED --log-bin=binlog --binlog-format=ROW
|
||||
volumes:
|
||||
- ./db:/var/lib/mysql #./db是数据库持久化目录,可以修改
|
||||
environment:
|
||||
- MYSQL_ROOT_PASSWORD=DeepGeek2025
|
||||
- MARIADB_AUTO_UPGRADE=1
|
||||
- MARIADB_DISABLE_UPGRADE_BACKUP=1
|
||||
env_file:
|
||||
- .env
|
||||
kodbox:
|
||||
image: docker.citory.tech/mirror/kodcloud/kodbox:latest
|
||||
container_name: kodbox
|
||||
restart: always
|
||||
ports:
|
||||
- 10510:80 #左边80是使用端口,可以修改
|
||||
volumes:
|
||||
- /home/deepgeek/data/data_base/kodbox/data:/var/www/html #./site是站点目录位置,可以修改
|
||||
environment:
|
||||
- MYSQL_HOST=kodbox-db
|
||||
- REDIS_HOST=kodbox-redis
|
||||
env_file:
|
||||
- .env
|
||||
depends_on:
|
||||
- kodbox-db
|
||||
- kodbox-redis
|
||||
kodbox-redis:
|
||||
image: docker.citory.tech/mirror/redis:alpine
|
||||
container_name: kodbox-redis
|
||||
restart: always
|
||||
x-dockge:
|
||||
urls:
|
||||
- http://local.citory.tech:10510
|
||||
networks: {}
|
17
nlp/compose.yaml
Normal file
@ -0,0 +1,17 @@
|
||||
version: "3.8"
|
||||
services:
|
||||
toolbox-nlp:
|
||||
image: docker.citory.tech/public/toolbox-nlp:1.0.1
|
||||
container_name: toolbox-nlp
|
||||
ports:
|
||||
- 10584:8080
|
||||
deploy:
|
||||
resources:
|
||||
reservations:
|
||||
devices:
|
||||
- driver: nvidia
|
||||
count: all
|
||||
capabilities:
|
||||
- gpu
|
||||
restart: always
|
||||
runtime: nvidia
|
1
ollama/.env
Normal file
@ -0,0 +1 @@
|
||||
# VARIABLE=value #comment
|
22
ollama/compose.yaml
Normal file
@ -0,0 +1,22 @@
|
||||
version: "3.9"
|
||||
services:
|
||||
ollama:
|
||||
image: docker.citory.tech/mirror/ollama/ollama:latest
|
||||
container_name: ollama
|
||||
ports:
|
||||
- 10581:11434
|
||||
runtime: nvidia
|
||||
volumes:
|
||||
- /home/deepgeek/data/data_local/server/ollama:/root/.ollama
|
||||
environment:
|
||||
NVIDIA_VISIBLE_DEVICES: all
|
||||
deploy:
|
||||
resources:
|
||||
reservations:
|
||||
devices:
|
||||
- capabilities:
|
||||
- gpu
|
||||
x-dockge:
|
||||
urls:
|
||||
- http://local.citory.tech:10581
|
||||
networks: {}
|
1
open-webui/.env
Normal file
@ -0,0 +1 @@
|
||||
# VARIABLE=value #comment
|
22
open-webui/compose.yaml
Normal file
@ -0,0 +1,22 @@
|
||||
version: "3.8"
|
||||
services:
|
||||
open-webui:
|
||||
image: docker.citory.tech/mirror/backplane/open-webui:0
|
||||
container_name: open-webui
|
||||
ports:
|
||||
- 10503:8080
|
||||
environment:
|
||||
- ENABLE_RAG_WEB_SEARCH=true
|
||||
- RAG_WEB_SEARCH_ENGINE=duckduckgo
|
||||
- ENABLE_OLLAMA_API=false
|
||||
- OPENAI_API_KEY=1ZfFN6nICGfMAUhPKwRbpmbwnd9aYkwT8RbluK32ASpPZgglPhdmLv4zDHh7BebQ # 替换为您的 OpenAI API 密钥
|
||||
- WHISPER_MODEL=large
|
||||
- OPENAI_API_BASE_URL=http://host.docker.internal:10580/v1 # 替换为您的 VLLM IP 和端口
|
||||
volumes:
|
||||
- ./data/:/app/backend/data
|
||||
extra_hosts:
|
||||
- host.docker.internal:host-gateway
|
||||
restart: always
|
||||
x-dockge:
|
||||
urls:
|
||||
- http://local.citory.tech:10503
|
1
owlsam/.env
Normal file
@ -0,0 +1 @@
|
||||
CUDA_VISIBLE_DEVICES=0,1
|
21
owlsam/compose.yaml
Normal file
@ -0,0 +1,21 @@
|
||||
version: "3.8"
|
||||
services:
|
||||
owlsam:
|
||||
image: docker.citory.tech/public/toolbox-owlsam:1.0.1
|
||||
container_name: owlsam
|
||||
ports:
|
||||
- 10582:8080
|
||||
env_file: .env
|
||||
deploy:
|
||||
resources:
|
||||
reservations:
|
||||
devices:
|
||||
- driver: nvidia
|
||||
count: all
|
||||
capabilities:
|
||||
- gpu
|
||||
restart: always
|
||||
runtime: nvidia
|
||||
x-dockge:
|
||||
urls:
|
||||
- http://local.citory.tech:10582
|
1
portainer/.env
Normal file
@ -0,0 +1 @@
|
||||
# VARIABLE=value #comment
|
15
portainer/compose.yaml
Normal file
@ -0,0 +1,15 @@
|
||||
version: "3.8"
|
||||
services:
|
||||
portainer:
|
||||
image: docker.citory.tech/mirror/6053537/portainer-ce:latest
|
||||
container_name: portainer
|
||||
ports:
|
||||
- 10507:9000
|
||||
volumes:
|
||||
- /var/run/docker.sock:/var/run/docker.sock
|
||||
- /var/lib/docker/volumes:/var/lib/docker/volumes
|
||||
- ./data:/data
|
||||
restart: always
|
||||
x-dockge:
|
||||
urls:
|
||||
- http://local.citory.tech:10507
|
1
stable-diffusion-webui/.env
Normal file
@ -0,0 +1 @@
|
||||
# VARIABLE=value #comment
|
34
stable-diffusion-webui/compose.yaml
Normal file
@ -0,0 +1,34 @@
|
||||
version: "3.9"
|
||||
services:
|
||||
stable-diffusion-webui:
|
||||
image: docker.citory.tech/public/stable-diffusion-webui:installed
|
||||
container_name: stable-diffusion-webui
|
||||
ports:
|
||||
- 10588:7860
|
||||
volumes:
|
||||
- /home/deepgeek/data/data_local/server/stable-diffusion-webui-docker/data:/data
|
||||
- /home/deepgeek/data/data_local/server/stable-diffusion-webui-docker/data/config/auto/localizations:/stable-diffusion-webui/localizations
|
||||
- /home/deepgeek/data/data_local/server/stable-diffusion-webui-docker/output:/output
|
||||
- /home/deepgeek/data/data_local/server/sd-models/checkpoints:/data/models/Stable-diffusion
|
||||
- /home/deepgeek/data/data_local/server/sd-models/loras:/data/models/Lora
|
||||
tty: true
|
||||
environment:
|
||||
- CUDA_LAUNCH_BLOCKING=1
|
||||
- COMMANDLINE_ARGS="--api"
|
||||
- COMMANDLINE_ARGS="--enable-insecure-extension-access"
|
||||
deploy:
|
||||
resources:
|
||||
reservations:
|
||||
devices:
|
||||
- driver: nvidia
|
||||
count: all
|
||||
capabilities:
|
||||
- compute
|
||||
- utility
|
||||
- gpu
|
||||
restart: always
|
||||
runtime: nvidia
|
||||
x-dockge:
|
||||
urls:
|
||||
- http://local.citory.tech:10588
|
||||
networks: {}
|
1
sun-panel/.env
Normal file
@ -0,0 +1 @@
|
||||
# VARIABLE=value #comment
|
18
sun-panel/compose.yaml
Normal file
@ -0,0 +1,18 @@
|
||||
version: "3.8"
|
||||
services:
|
||||
sun-panel:
|
||||
image: docker.citory.tech/mirror/hslr/sun-panel:latest
|
||||
container_name: sun-panel
|
||||
volumes:
|
||||
- ./conf:/app/conf
|
||||
- /var/run/docker.sock:/var/run/docker.sock # 挂载docker.sock
|
||||
- ./runtime:/app/runtime # 挂载日志目录
|
||||
- /mnt/data_base:/os # 硬盘挂载点(根据自己需求修改)
|
||||
- /:/system # 硬盘挂载点(根据自己需求修改)
|
||||
ports:
|
||||
- 10500:3002
|
||||
restart: always
|
||||
networks: {}
|
||||
x-dockge:
|
||||
urls:
|
||||
- http://local.citory.tech:10500
|
BIN
sun-panel/conf/custom/favicon.ico
Normal file
After Width: | Height: | Size: 18 KiB |
0
sun-panel/conf/custom/index.css
Normal file
0
sun-panel/conf/custom/index.js
Normal file
After Width: | Height: | Size: 16 KiB |
After Width: | Height: | Size: 7.3 KiB |
After Width: | Height: | Size: 180 KiB |
@ -0,0 +1 @@
|
||||
<svg height="1em" style="flex:none;line-height:1" viewBox="0 0 24 24" width="1em" xmlns="http://www.w3.org/2000/svg"><title>vLLM</title><path d="M0 4.973h9.324V23L0 4.973z" fill="#FDB515"></path><path d="M13.986 4.351L22.378 0l-6.216 23H9.324l4.662-18.649z" fill="#30A2FF"></path></svg>
|
After Width: | Height: | Size: 286 B |
After Width: | Height: | Size: 2.2 KiB |
After Width: | Height: | Size: 11 KiB |
After Width: | Height: | Size: 82 KiB |
After Width: | Height: | Size: 180 KiB |
After Width: | Height: | Size: 21 KiB |
@ -0,0 +1,14 @@
|
||||
<?xml version="1.0" encoding="UTF-8" standalone="no" ?>
|
||||
<!DOCTYPE svg PUBLIC "-//W3C//DTD SVG 1.1//EN" "http://www.w3.org/Graphics/SVG/1.1/DTD/svg11.dtd">
|
||||
<svg xmlns="http://www.w3.org/2000/svg" xmlns:xlink="http://www.w3.org/1999/xlink" version="1.1" width="640" height="640" viewBox="0 0 640 640" xml:space="preserve">
|
||||
<desc>Created with Fabric.js 5.3.0</desc>
|
||||
<defs>
|
||||
</defs>
|
||||
<g transform="matrix(0.9544918218 0 0 0.9544918218 320 325.5657767239)" id="0UAuLmXgnot4bJxVEVJCQ" >
|
||||
<linearGradient id="SVGID_136_0" gradientUnits="userSpaceOnUse" gradientTransform="matrix(1 0 0 1 -236.6470440833 -213.9441386034)" x1="259.78" y1="261.15" x2="463.85" y2="456.49">
|
||||
<stop offset="0%" style="stop-color:#74C2FF;stop-opacity: 1"/>
|
||||
<stop offset="100%" style="stop-color:rgb(134,230,169);stop-opacity: 1"/>
|
||||
</linearGradient>
|
||||
<path style="stroke: rgb(242,242,242); stroke-opacity: 0.51; stroke-width: 190; stroke-dasharray: none; stroke-linecap: butt; stroke-dashoffset: 0; stroke-linejoin: miter; stroke-miterlimit: 4; fill: url(#SVGID_136_0); fill-rule: nonzero; opacity: 1;" transform=" translate(0, 0)" d="M 131.8665 -139.04883 C 159.01022 -111.20969000000001 170.12421 -99.45396000000001 203.11849999999998 -51.72057000000001 C 236.1128 -3.9871800000000093 264.44147999999996 83.98416999999998 187.33995 144.05073 C 177.72728999999998 151.53955 166.73827 158.81189999999998 154.65932999999998 165.65812999999997 C 69.85514999999998 213.72433999999998 -68.67309000000003 240.78578 -161.79279 174.28328999999997 C -268.17583 98.30862999999997 -260.10282 -68.66557000000003 -144.35093 -170.50579000000005 C -28.599040000000002 -272.34602000000007 104.72278 -166.88797000000005 131.86649999999997 -139.04883000000004 z" stroke-linecap="round" />
|
||||
</g>
|
||||
</svg>
|
After Width: | Height: | Size: 1.7 KiB |
After Width: | Height: | Size: 252 KiB |
After Width: | Height: | Size: 252 KiB |
After Width: | Height: | Size: 5.5 KiB |
After Width: | Height: | Size: 27 KiB |
After Width: | Height: | Size: 5.5 KiB |
After Width: | Height: | Size: 16 KiB |
1
toolbox-ocr/.env
Normal file
@ -0,0 +1 @@
|
||||
# VARIABLE=value #comment
|
20
toolbox-ocr/compose.yaml
Normal file
@ -0,0 +1,20 @@
|
||||
version: "3.8"
|
||||
services:
|
||||
toolbox-ocr:
|
||||
image: docker.citory.tech/public/toolbox-ocr:1.0.0
|
||||
container_name: toolbox-ocr
|
||||
ports:
|
||||
- 10585:8080
|
||||
deploy:
|
||||
resources:
|
||||
reservations:
|
||||
devices:
|
||||
- driver: nvidia
|
||||
count: all
|
||||
capabilities:
|
||||
- gpu
|
||||
restart: always
|
||||
runtime: nvidia
|
||||
x-dockge:
|
||||
urls:
|
||||
- http://local.citory.tech:10585
|
32
vllm/compose.yaml
Normal file
@ -0,0 +1,32 @@
|
||||
version: "3.8"
|
||||
services:
|
||||
vllm:
|
||||
image: docker.citory.tech/mirror/vllm/vllm-openai:latest
|
||||
container_name: vllm
|
||||
ports:
|
||||
- 10580:8080
|
||||
tty: true
|
||||
environment:
|
||||
- CUDA_VISIBLE_DEVICES=0,1
|
||||
deploy:
|
||||
resources:
|
||||
reservations:
|
||||
devices:
|
||||
- driver: nvidia
|
||||
count: 2
|
||||
capabilities:
|
||||
- gpu
|
||||
ipc: host
|
||||
runtime: nvidia
|
||||
volumes:
|
||||
- /home/deepgeek/data/data_local/server/vllm/models:/models
|
||||
restart: always
|
||||
command: --served-model-name DeepSeek-R1-Distill-Llama-70B-AWQ --model
|
||||
/models/Valdemardi/DeepSeek-R1-Distill-Llama-70B-AWQ --trust-remote-code
|
||||
--host 0.0.0.0 --port 8080 --max-model-len 8192 --tensor-parallel-size 2
|
||||
--gpu_memory_utilization 0.96 --enforce-eager --dtype auto --swap-space 8
|
||||
--api-key
|
||||
"O8Is3NSYnp5fICWFbhkbwpLWgvMLkdCSuXR5ZggLmgwTKNPEWsjx1NqUxkyU7wLX"
|
||||
x-dockge:
|
||||
urls:
|
||||
- http://local.citory.tech:10580
|
1
watchtower/.env
Normal file
@ -0,0 +1 @@
|
||||
# VARIABLE=value #comment
|
9
watchtower/compose.yaml
Normal file
@ -0,0 +1,9 @@
|
||||
version: "3.8"
|
||||
services:
|
||||
watchtower:
|
||||
container_name: watchtower
|
||||
image: docker.citory.tech/mirror/containrrr/watchtower:latest
|
||||
volumes:
|
||||
- /var/run/docker.sock:/var/run/docker.sock
|
||||
command: --interval 3600 # 每小时检查一次[[1,10]]
|
||||
networks: {}
|