mirror of
https://github.com/arch3rPro/1Panel-Appstore.git
synced 2026-05-09 10:48:07 +08:00
feat: fix dify 1.6.0 installed failed
This commit is contained in:
@@ -39,6 +39,12 @@ APP_WEB_URL=
|
|||||||
# File preview or download Url prefix.
|
# File preview or download Url prefix.
|
||||||
# used to display File preview or download Url to the front-end or as Multi-model inputs;
|
# used to display File preview or download Url to the front-end or as Multi-model inputs;
|
||||||
# Url is signed and has expiration time.
|
# Url is signed and has expiration time.
|
||||||
|
# Setting FILES_URL is required for file processing plugins.
|
||||||
|
# - For https://example.com, use FILES_URL=https://example.com
|
||||||
|
# - For http://example.com, use FILES_URL=http://example.com
|
||||||
|
# Recommendation: use a dedicated domain (e.g., https://upload.example.com).
|
||||||
|
# Alternatively, use http://<your-ip>:5001 or http://api:5001,
|
||||||
|
# ensuring port 5001 is externally accessible (see docker-compose.yaml).
|
||||||
FILES_URL=
|
FILES_URL=
|
||||||
|
|
||||||
# ------------------------------
|
# ------------------------------
|
||||||
@@ -68,7 +74,11 @@ DEBUG=false
|
|||||||
# which is convenient for debugging.
|
# which is convenient for debugging.
|
||||||
FLASK_DEBUG=false
|
FLASK_DEBUG=false
|
||||||
|
|
||||||
# A secretkey that is used for securely signing the session cookie
|
# Enable request logging, which will log the request and response information.
|
||||||
|
# And the log level is DEBUG
|
||||||
|
ENABLE_REQUEST_LOGGING=False
|
||||||
|
|
||||||
|
# A secret key that is used for securely signing the session cookie
|
||||||
# and encrypting sensitive information on the database.
|
# and encrypting sensitive information on the database.
|
||||||
# You can generate a strong key using `openssl rand -base64 42`.
|
# You can generate a strong key using `openssl rand -base64 42`.
|
||||||
SECRET_KEY=sk-9f73s3ljTXVcMT3Blb3ljTqtsKiGHXVcMT3BlbkFJLK7U
|
SECRET_KEY=sk-9f73s3ljTXVcMT3Blb3ljTqtsKiGHXVcMT3BlbkFJLK7U
|
||||||
@@ -76,7 +86,7 @@ SECRET_KEY=sk-9f73s3ljTXVcMT3Blb3ljTqtsKiGHXVcMT3BlbkFJLK7U
|
|||||||
# Password for admin user initialization.
|
# Password for admin user initialization.
|
||||||
# If left unset, admin user will not be prompted for a password
|
# If left unset, admin user will not be prompted for a password
|
||||||
# when creating the initial admin account.
|
# when creating the initial admin account.
|
||||||
# The length of the password cannot exceed 30 charactors.
|
# The length of the password cannot exceed 30 characters.
|
||||||
INIT_PASSWORD=
|
INIT_PASSWORD=
|
||||||
|
|
||||||
# Deployment environment.
|
# Deployment environment.
|
||||||
@@ -174,6 +184,12 @@ CELERY_MIN_WORKERS=
|
|||||||
API_TOOL_DEFAULT_CONNECT_TIMEOUT=10
|
API_TOOL_DEFAULT_CONNECT_TIMEOUT=10
|
||||||
API_TOOL_DEFAULT_READ_TIMEOUT=60
|
API_TOOL_DEFAULT_READ_TIMEOUT=60
|
||||||
|
|
||||||
|
# -------------------------------
|
||||||
|
# Datasource Configuration
|
||||||
|
# --------------------------------
|
||||||
|
ENABLE_WEBSITE_JINAREADER=true
|
||||||
|
ENABLE_WEBSITE_FIRECRAWL=true
|
||||||
|
ENABLE_WEBSITE_WATERCRAWL=true
|
||||||
|
|
||||||
# ------------------------------
|
# ------------------------------
|
||||||
# Database Configuration
|
# Database Configuration
|
||||||
@@ -269,6 +285,7 @@ BROKER_USE_SSL=false
|
|||||||
# If you are using Redis Sentinel for high availability, configure the following settings.
|
# If you are using Redis Sentinel for high availability, configure the following settings.
|
||||||
CELERY_USE_SENTINEL=false
|
CELERY_USE_SENTINEL=false
|
||||||
CELERY_SENTINEL_MASTER_NAME=
|
CELERY_SENTINEL_MASTER_NAME=
|
||||||
|
CELERY_SENTINEL_PASSWORD=
|
||||||
CELERY_SENTINEL_SOCKET_TIMEOUT=0.1
|
CELERY_SENTINEL_SOCKET_TIMEOUT=0.1
|
||||||
|
|
||||||
# ------------------------------
|
# ------------------------------
|
||||||
@@ -344,7 +361,7 @@ TENCENT_COS_SCHEME=your-scheme
|
|||||||
|
|
||||||
# Oracle Storage Configuration
|
# Oracle Storage Configuration
|
||||||
#
|
#
|
||||||
OCI_ENDPOINT=https://objectstorage.us-ashburn-1.oraclecloud.com
|
OCI_ENDPOINT=https://your-object-storage-namespace.compat.objectstorage.us-ashburn-1.oraclecloud.com
|
||||||
OCI_BUCKET_NAME=your-bucket-name
|
OCI_BUCKET_NAME=your-bucket-name
|
||||||
OCI_ACCESS_KEY=your-access-key
|
OCI_ACCESS_KEY=your-access-key
|
||||||
OCI_SECRET_KEY=your-secret-key
|
OCI_SECRET_KEY=your-secret-key
|
||||||
@@ -383,7 +400,7 @@ SUPABASE_URL=your-server-url
|
|||||||
# ------------------------------
|
# ------------------------------
|
||||||
|
|
||||||
# The type of vector store to use.
|
# The type of vector store to use.
|
||||||
# Supported values are `weaviate`, `qdrant`, `milvus`, `myscale`, `relyt`, `pgvector`, `pgvecto-rs`, `chroma`, `opensearch`, `tidb_vector`, `oracle`, `tencent`, `elasticsearch`, `elasticsearch-ja`, `analyticdb`, `couchbase`, `vikingdb`, `oceanbase`.
|
# Supported values are `weaviate`, `qdrant`, `milvus`, `myscale`, `relyt`, `pgvector`, `pgvecto-rs`, `chroma`, `opensearch`, `oracle`, `tencent`, `elasticsearch`, `elasticsearch-ja`, `analyticdb`, `couchbase`, `vikingdb`, `oceanbase`, `opengauss`, `tablestore`,`vastbase`,`tidb`,`tidb_on_qdrant`,`baidu`,`lindorm`,`huawei_cloud`,`upstash`, `matrixone`.
|
||||||
VECTOR_STORE=weaviate
|
VECTOR_STORE=weaviate
|
||||||
|
|
||||||
# The Weaviate endpoint URL. Only available when VECTOR_STORE is `weaviate`.
|
# The Weaviate endpoint URL. Only available when VECTOR_STORE is `weaviate`.
|
||||||
@@ -396,14 +413,17 @@ QDRANT_API_KEY=difyai123456
|
|||||||
QDRANT_CLIENT_TIMEOUT=20
|
QDRANT_CLIENT_TIMEOUT=20
|
||||||
QDRANT_GRPC_ENABLED=false
|
QDRANT_GRPC_ENABLED=false
|
||||||
QDRANT_GRPC_PORT=6334
|
QDRANT_GRPC_PORT=6334
|
||||||
|
QDRANT_REPLICATION_FACTOR=1
|
||||||
|
|
||||||
# Milvus configuration Only available when VECTOR_STORE is `milvus`.
|
# Milvus configuration. Only available when VECTOR_STORE is `milvus`.
|
||||||
# The milvus uri.
|
# The milvus uri.
|
||||||
MILVUS_URI=http://127.0.0.1:19530
|
MILVUS_URI=http://host.docker.internal:19530
|
||||||
|
MILVUS_DATABASE=
|
||||||
MILVUS_TOKEN=
|
MILVUS_TOKEN=
|
||||||
MILVUS_USER=root
|
MILVUS_USER=
|
||||||
MILVUS_PASSWORD=Milvus
|
MILVUS_PASSWORD=
|
||||||
MILVUS_ENABLE_HYBRID_SEARCH=False
|
MILVUS_ENABLE_HYBRID_SEARCH=False
|
||||||
|
MILVUS_ANALYZER_PARAMS=
|
||||||
|
|
||||||
# MyScale configuration, only available when VECTOR_STORE is `myscale`
|
# MyScale configuration, only available when VECTOR_STORE is `myscale`
|
||||||
# For multi-language support, please set MYSCALE_FTS_PARAMS with referring to:
|
# For multi-language support, please set MYSCALE_FTS_PARAMS with referring to:
|
||||||
@@ -431,6 +451,17 @@ PGVECTOR_PASSWORD=difyai123456
|
|||||||
PGVECTOR_DATABASE=dify
|
PGVECTOR_DATABASE=dify
|
||||||
PGVECTOR_MIN_CONNECTION=1
|
PGVECTOR_MIN_CONNECTION=1
|
||||||
PGVECTOR_MAX_CONNECTION=5
|
PGVECTOR_MAX_CONNECTION=5
|
||||||
|
PGVECTOR_PG_BIGM=false
|
||||||
|
PGVECTOR_PG_BIGM_VERSION=1.2-20240606
|
||||||
|
|
||||||
|
# vastbase configurations, only available when VECTOR_STORE is `vastbase`
|
||||||
|
VASTBASE_HOST=vastbase
|
||||||
|
VASTBASE_PORT=5432
|
||||||
|
VASTBASE_USER=dify
|
||||||
|
VASTBASE_PASSWORD=Difyai123456
|
||||||
|
VASTBASE_DATABASE=dify
|
||||||
|
VASTBASE_MIN_CONNECTION=1
|
||||||
|
VASTBASE_MAX_CONNECTION=5
|
||||||
|
|
||||||
# pgvecto-rs configurations, only available when VECTOR_STORE is `pgvecto-rs`
|
# pgvecto-rs configurations, only available when VECTOR_STORE is `pgvecto-rs`
|
||||||
PGVECTO_RS_HOST=pgvecto-rs
|
PGVECTO_RS_HOST=pgvecto-rs
|
||||||
@@ -453,13 +484,20 @@ ANALYTICDB_PORT=5432
|
|||||||
ANALYTICDB_MIN_CONNECTION=1
|
ANALYTICDB_MIN_CONNECTION=1
|
||||||
ANALYTICDB_MAX_CONNECTION=5
|
ANALYTICDB_MAX_CONNECTION=5
|
||||||
|
|
||||||
# TiDB vector configurations, only available when VECTOR_STORE is `tidb`
|
# TiDB vector configurations, only available when VECTOR_STORE is `tidb_vector`
|
||||||
TIDB_VECTOR_HOST=tidb
|
TIDB_VECTOR_HOST=tidb
|
||||||
TIDB_VECTOR_PORT=4000
|
TIDB_VECTOR_PORT=4000
|
||||||
TIDB_VECTOR_USER=
|
TIDB_VECTOR_USER=
|
||||||
TIDB_VECTOR_PASSWORD=
|
TIDB_VECTOR_PASSWORD=
|
||||||
TIDB_VECTOR_DATABASE=dify
|
TIDB_VECTOR_DATABASE=dify
|
||||||
|
|
||||||
|
# Matrixone vector configurations.
|
||||||
|
MATRIXONE_HOST=matrixone
|
||||||
|
MATRIXONE_PORT=6001
|
||||||
|
MATRIXONE_USER=dump
|
||||||
|
MATRIXONE_PASSWORD=111
|
||||||
|
MATRIXONE_DATABASE=dify
|
||||||
|
|
||||||
# Tidb on qdrant configuration, only available when VECTOR_STORE is `tidb_on_qdrant`
|
# Tidb on qdrant configuration, only available when VECTOR_STORE is `tidb_on_qdrant`
|
||||||
TIDB_ON_QDRANT_URL=http://127.0.0.1
|
TIDB_ON_QDRANT_URL=http://127.0.0.1
|
||||||
TIDB_ON_QDRANT_API_KEY=dify
|
TIDB_ON_QDRANT_API_KEY=dify
|
||||||
@@ -483,11 +521,13 @@ CHROMA_AUTH_PROVIDER=chromadb.auth.token_authn.TokenAuthClientProvider
|
|||||||
CHROMA_AUTH_CREDENTIALS=
|
CHROMA_AUTH_CREDENTIALS=
|
||||||
|
|
||||||
# Oracle configuration, only available when VECTOR_STORE is `oracle`
|
# Oracle configuration, only available when VECTOR_STORE is `oracle`
|
||||||
ORACLE_HOST=oracle
|
|
||||||
ORACLE_PORT=1521
|
|
||||||
ORACLE_USER=dify
|
ORACLE_USER=dify
|
||||||
ORACLE_PASSWORD=dify
|
ORACLE_PASSWORD=dify
|
||||||
ORACLE_DATABASE=FREEPDB1
|
ORACLE_DSN=oracle:1521/FREEPDB1
|
||||||
|
ORACLE_CONFIG_DIR=/app/api/storage/wallet
|
||||||
|
ORACLE_WALLET_LOCATION=/app/api/storage/wallet
|
||||||
|
ORACLE_WALLET_PASSWORD=dify
|
||||||
|
ORACLE_IS_AUTONOMOUS=false
|
||||||
|
|
||||||
# relyt configurations, only available when VECTOR_STORE is `relyt`
|
# relyt configurations, only available when VECTOR_STORE is `relyt`
|
||||||
RELYT_HOST=db
|
RELYT_HOST=db
|
||||||
@@ -499,9 +539,14 @@ RELYT_DATABASE=postgres
|
|||||||
# open search configuration, only available when VECTOR_STORE is `opensearch`
|
# open search configuration, only available when VECTOR_STORE is `opensearch`
|
||||||
OPENSEARCH_HOST=opensearch
|
OPENSEARCH_HOST=opensearch
|
||||||
OPENSEARCH_PORT=9200
|
OPENSEARCH_PORT=9200
|
||||||
|
OPENSEARCH_SECURE=true
|
||||||
|
OPENSEARCH_VERIFY_CERTS=true
|
||||||
|
OPENSEARCH_AUTH_METHOD=basic
|
||||||
OPENSEARCH_USER=admin
|
OPENSEARCH_USER=admin
|
||||||
OPENSEARCH_PASSWORD=admin
|
OPENSEARCH_PASSWORD=admin
|
||||||
OPENSEARCH_SECURE=true
|
# If using AWS managed IAM, e.g. Managed Cluster or OpenSearch Serverless
|
||||||
|
OPENSEARCH_AWS_REGION=ap-southeast-1
|
||||||
|
OPENSEARCH_AWS_SERVICE=aoss
|
||||||
|
|
||||||
# tencent vector configurations, only available when VECTOR_STORE is `tencent`
|
# tencent vector configurations, only available when VECTOR_STORE is `tencent`
|
||||||
TENCENT_VECTOR_DB_URL=http://127.0.0.1
|
TENCENT_VECTOR_DB_URL=http://127.0.0.1
|
||||||
@@ -511,6 +556,7 @@ TENCENT_VECTOR_DB_USERNAME=dify
|
|||||||
TENCENT_VECTOR_DB_DATABASE=dify
|
TENCENT_VECTOR_DB_DATABASE=dify
|
||||||
TENCENT_VECTOR_DB_SHARD=1
|
TENCENT_VECTOR_DB_SHARD=1
|
||||||
TENCENT_VECTOR_DB_REPLICAS=2
|
TENCENT_VECTOR_DB_REPLICAS=2
|
||||||
|
TENCENT_VECTOR_DB_ENABLE_HYBRID_SEARCH=false
|
||||||
|
|
||||||
# ElasticSearch configuration, only available when VECTOR_STORE is `elasticsearch`
|
# ElasticSearch configuration, only available when VECTOR_STORE is `elasticsearch`
|
||||||
ELASTICSEARCH_HOST=0.0.0.0
|
ELASTICSEARCH_HOST=0.0.0.0
|
||||||
@@ -541,6 +587,7 @@ VIKINGDB_SOCKET_TIMEOUT=30
|
|||||||
LINDORM_URL=http://lindorm:30070
|
LINDORM_URL=http://lindorm:30070
|
||||||
LINDORM_USERNAME=lindorm
|
LINDORM_USERNAME=lindorm
|
||||||
LINDORM_PASSWORD=lindorm
|
LINDORM_PASSWORD=lindorm
|
||||||
|
LINDORM_QUERY_TIMEOUT=1
|
||||||
|
|
||||||
# OceanBase Vector configuration, only available when VECTOR_STORE is `oceanbase`
|
# OceanBase Vector configuration, only available when VECTOR_STORE is `oceanbase`
|
||||||
OCEANBASE_VECTOR_HOST=oceanbase
|
OCEANBASE_VECTOR_HOST=oceanbase
|
||||||
@@ -550,11 +597,34 @@ OCEANBASE_VECTOR_PASSWORD=difyai123456
|
|||||||
OCEANBASE_VECTOR_DATABASE=test
|
OCEANBASE_VECTOR_DATABASE=test
|
||||||
OCEANBASE_CLUSTER_NAME=difyai
|
OCEANBASE_CLUSTER_NAME=difyai
|
||||||
OCEANBASE_MEMORY_LIMIT=6G
|
OCEANBASE_MEMORY_LIMIT=6G
|
||||||
|
OCEANBASE_ENABLE_HYBRID_SEARCH=false
|
||||||
|
|
||||||
|
# opengauss configurations, only available when VECTOR_STORE is `opengauss`
|
||||||
|
OPENGAUSS_HOST=opengauss
|
||||||
|
OPENGAUSS_PORT=6600
|
||||||
|
OPENGAUSS_USER=postgres
|
||||||
|
OPENGAUSS_PASSWORD=Dify@123
|
||||||
|
OPENGAUSS_DATABASE=dify
|
||||||
|
OPENGAUSS_MIN_CONNECTION=1
|
||||||
|
OPENGAUSS_MAX_CONNECTION=5
|
||||||
|
OPENGAUSS_ENABLE_PQ=false
|
||||||
|
|
||||||
|
# huawei cloud search service vector configurations, only available when VECTOR_STORE is `huawei_cloud`
|
||||||
|
HUAWEI_CLOUD_HOSTS=https://127.0.0.1:9200
|
||||||
|
HUAWEI_CLOUD_USER=admin
|
||||||
|
HUAWEI_CLOUD_PASSWORD=admin
|
||||||
|
|
||||||
# Upstash Vector configuration, only available when VECTOR_STORE is `upstash`
|
# Upstash Vector configuration, only available when VECTOR_STORE is `upstash`
|
||||||
UPSTASH_VECTOR_URL=https://xxx-vector.upstash.io
|
UPSTASH_VECTOR_URL=https://xxx-vector.upstash.io
|
||||||
UPSTASH_VECTOR_TOKEN=dify
|
UPSTASH_VECTOR_TOKEN=dify
|
||||||
|
|
||||||
|
# TableStore Vector configuration
|
||||||
|
# (only used when VECTOR_STORE is tablestore)
|
||||||
|
TABLESTORE_ENDPOINT=https://instance-name.cn-hangzhou.ots.aliyuncs.com
|
||||||
|
TABLESTORE_INSTANCE_NAME=instance-name
|
||||||
|
TABLESTORE_ACCESS_KEY_ID=xxx
|
||||||
|
TABLESTORE_ACCESS_KEY_SECRET=xxx
|
||||||
|
|
||||||
# ------------------------------
|
# ------------------------------
|
||||||
# Knowledge Configuration
|
# Knowledge Configuration
|
||||||
# ------------------------------
|
# ------------------------------
|
||||||
@@ -593,6 +663,11 @@ PROMPT_GENERATION_MAX_TOKENS=512
|
|||||||
# Default: 1024 tokens.
|
# Default: 1024 tokens.
|
||||||
CODE_GENERATION_MAX_TOKENS=1024
|
CODE_GENERATION_MAX_TOKENS=1024
|
||||||
|
|
||||||
|
# Enable or disable plugin based token counting. If disabled, token counting will return 0.
|
||||||
|
# This can improve performance by skipping token counting operations.
|
||||||
|
# Default: false (disabled).
|
||||||
|
PLUGIN_BASED_TOKEN_COUNTING_ENABLED=false
|
||||||
|
|
||||||
# ------------------------------
|
# ------------------------------
|
||||||
# Multi-modal Configuration
|
# Multi-modal Configuration
|
||||||
# ------------------------------
|
# ------------------------------
|
||||||
@@ -652,10 +727,11 @@ NOTION_INTERNAL_SECRET=
|
|||||||
# Mail related configuration
|
# Mail related configuration
|
||||||
# ------------------------------
|
# ------------------------------
|
||||||
|
|
||||||
# Mail type, support: resend, smtp
|
# Mail type, support: resend, smtp, sendgrid
|
||||||
MAIL_TYPE=resend
|
MAIL_TYPE=resend
|
||||||
|
|
||||||
# Default send from email address, if not specified
|
# Default send from email address, if not specified
|
||||||
|
# If using SendGrid, use the 'from' field for authentication if necessary.
|
||||||
MAIL_DEFAULT_SEND_FROM=
|
MAIL_DEFAULT_SEND_FROM=
|
||||||
|
|
||||||
# API-Key for the Resend email provider, used when MAIL_TYPE is `resend`.
|
# API-Key for the Resend email provider, used when MAIL_TYPE is `resend`.
|
||||||
@@ -671,6 +747,9 @@ SMTP_PASSWORD=
|
|||||||
SMTP_USE_TLS=true
|
SMTP_USE_TLS=true
|
||||||
SMTP_OPPORTUNISTIC_TLS=false
|
SMTP_OPPORTUNISTIC_TLS=false
|
||||||
|
|
||||||
|
# Sendgid configuration
|
||||||
|
SENDGRID_API_KEY=
|
||||||
|
|
||||||
# ------------------------------
|
# ------------------------------
|
||||||
# Others Configuration
|
# Others Configuration
|
||||||
# ------------------------------
|
# ------------------------------
|
||||||
@@ -709,15 +788,37 @@ MAX_VARIABLE_SIZE=204800
|
|||||||
WORKFLOW_PARALLEL_DEPTH_LIMIT=3
|
WORKFLOW_PARALLEL_DEPTH_LIMIT=3
|
||||||
WORKFLOW_FILE_UPLOAD_LIMIT=10
|
WORKFLOW_FILE_UPLOAD_LIMIT=10
|
||||||
|
|
||||||
|
# Workflow storage configuration
|
||||||
|
# Options: rdbms, hybrid
|
||||||
|
# rdbms: Use only the relational database (default)
|
||||||
|
# hybrid: Save new data to object storage, read from both object storage and RDBMS
|
||||||
|
WORKFLOW_NODE_EXECUTION_STORAGE=rdbms
|
||||||
|
|
||||||
# HTTP request node in workflow configuration
|
# HTTP request node in workflow configuration
|
||||||
HTTP_REQUEST_NODE_MAX_BINARY_SIZE=10485760
|
HTTP_REQUEST_NODE_MAX_BINARY_SIZE=10485760
|
||||||
HTTP_REQUEST_NODE_MAX_TEXT_SIZE=1048576
|
HTTP_REQUEST_NODE_MAX_TEXT_SIZE=1048576
|
||||||
|
HTTP_REQUEST_NODE_SSL_VERIFY=True
|
||||||
|
|
||||||
|
# Respect X-* headers to redirect clients
|
||||||
|
RESPECT_XFORWARD_HEADERS_ENABLED=false
|
||||||
|
|
||||||
# SSRF Proxy server HTTP URL
|
# SSRF Proxy server HTTP URL
|
||||||
SSRF_PROXY_HTTP_URL=http://ssrf_proxy:3128
|
SSRF_PROXY_HTTP_URL=http://ssrf_proxy:3128
|
||||||
# SSRF Proxy server HTTPS URL
|
# SSRF Proxy server HTTPS URL
|
||||||
SSRF_PROXY_HTTPS_URL=http://ssrf_proxy:3128
|
SSRF_PROXY_HTTPS_URL=http://ssrf_proxy:3128
|
||||||
|
|
||||||
|
# Maximum loop count in the workflow
|
||||||
|
LOOP_NODE_MAX_COUNT=100
|
||||||
|
|
||||||
|
# The maximum number of tools that can be used in the agent.
|
||||||
|
MAX_TOOLS_NUM=10
|
||||||
|
|
||||||
|
# Maximum number of Parallelism branches in the workflow
|
||||||
|
MAX_PARALLEL_LIMIT=10
|
||||||
|
|
||||||
|
# The maximum number of iterations for agent setting
|
||||||
|
MAX_ITERATIONS_NUM=99
|
||||||
|
|
||||||
# ------------------------------
|
# ------------------------------
|
||||||
# Environment Variables for web Service
|
# Environment Variables for web Service
|
||||||
# ------------------------------
|
# ------------------------------
|
||||||
@@ -725,11 +826,15 @@ SSRF_PROXY_HTTPS_URL=http://ssrf_proxy:3128
|
|||||||
# The timeout for the text generation in millisecond
|
# The timeout for the text generation in millisecond
|
||||||
TEXT_GENERATION_TIMEOUT_MS=60000
|
TEXT_GENERATION_TIMEOUT_MS=60000
|
||||||
|
|
||||||
|
# Allow rendering unsafe URLs which have "data:" scheme.
|
||||||
|
ALLOW_UNSAFE_DATA_SCHEME=false
|
||||||
|
|
||||||
# ------------------------------
|
# ------------------------------
|
||||||
# Environment Variables for db Service
|
# Environment Variables for db Service
|
||||||
# ------------------------------
|
# ------------------------------
|
||||||
|
|
||||||
PGUSER=${DB_USERNAME}
|
# The name of the default postgres user.
|
||||||
|
POSTGRES_USER=${DB_USERNAME}
|
||||||
# The password for the default postgres user.
|
# The password for the default postgres user.
|
||||||
POSTGRES_PASSWORD=${DB_PASSWORD}
|
POSTGRES_PASSWORD=${DB_PASSWORD}
|
||||||
# The name of the default postgres database.
|
# The name of the default postgres database.
|
||||||
@@ -785,7 +890,7 @@ CHROMA_IS_PERSISTENT=TRUE
|
|||||||
|
|
||||||
# ------------------------------
|
# ------------------------------
|
||||||
# Environment Variables for Oracle Service
|
# Environment Variables for Oracle Service
|
||||||
# (only used when VECTOR_STORE is Oracle)
|
# (only used when VECTOR_STORE is oracle)
|
||||||
# ------------------------------
|
# ------------------------------
|
||||||
ORACLE_PWD=Dify123456
|
ORACLE_PWD=Dify123456
|
||||||
ORACLE_CHARACTERSET=AL32UTF8
|
ORACLE_CHARACTERSET=AL32UTF8
|
||||||
@@ -856,7 +961,7 @@ NGINX_SSL_PROTOCOLS=TLSv1.1 TLSv1.2 TLSv1.3
|
|||||||
|
|
||||||
# Nginx performance tuning
|
# Nginx performance tuning
|
||||||
NGINX_WORKER_PROCESSES=auto
|
NGINX_WORKER_PROCESSES=auto
|
||||||
NGINX_CLIENT_MAX_BODY_SIZE=15M
|
NGINX_CLIENT_MAX_BODY_SIZE=100M
|
||||||
NGINX_KEEPALIVE_TIMEOUT=65
|
NGINX_KEEPALIVE_TIMEOUT=65
|
||||||
|
|
||||||
# Proxy settings
|
# Proxy settings
|
||||||
@@ -954,6 +1059,7 @@ PLUGIN_DEBUGGING_PORT=5003
|
|||||||
EXPOSE_PLUGIN_DEBUGGING_HOST=localhost
|
EXPOSE_PLUGIN_DEBUGGING_HOST=localhost
|
||||||
EXPOSE_PLUGIN_DEBUGGING_PORT=5003
|
EXPOSE_PLUGIN_DEBUGGING_PORT=5003
|
||||||
|
|
||||||
|
# If this key is changed, DIFY_INNER_API_KEY in plugin_daemon service must also be updated or agent node will fail.
|
||||||
PLUGIN_DIFY_INNER_API_KEY=QaHbTe77CtuXmsfyhR7+vRjI/+XbV1AaFy691iy+kGDv2Jvy0/eAh8Y1
|
PLUGIN_DIFY_INNER_API_KEY=QaHbTe77CtuXmsfyhR7+vRjI/+XbV1AaFy691iy+kGDv2Jvy0/eAh8Y1
|
||||||
PLUGIN_DIFY_INNER_API_URL=http://api:5001
|
PLUGIN_DIFY_INNER_API_URL=http://api:5001
|
||||||
|
|
||||||
@@ -963,3 +1069,72 @@ MARKETPLACE_ENABLED=true
|
|||||||
MARKETPLACE_API_URL=https://marketplace.dify.ai
|
MARKETPLACE_API_URL=https://marketplace.dify.ai
|
||||||
|
|
||||||
FORCE_VERIFYING_SIGNATURE=true
|
FORCE_VERIFYING_SIGNATURE=true
|
||||||
|
|
||||||
|
PLUGIN_PYTHON_ENV_INIT_TIMEOUT=120
|
||||||
|
PLUGIN_MAX_EXECUTION_TIMEOUT=600
|
||||||
|
# PIP_MIRROR_URL=https://pypi.tuna.tsinghua.edu.cn/simple
|
||||||
|
PIP_MIRROR_URL=
|
||||||
|
|
||||||
|
# https://github.com/langgenius/dify-plugin-daemon/blob/main/.env.example
|
||||||
|
# Plugin storage type, local aws_s3 tencent_cos azure_blob aliyun_oss volcengine_tos
|
||||||
|
PLUGIN_STORAGE_TYPE=local
|
||||||
|
PLUGIN_STORAGE_LOCAL_ROOT=/app/storage
|
||||||
|
PLUGIN_WORKING_PATH=/app/storage/cwd
|
||||||
|
PLUGIN_INSTALLED_PATH=plugin
|
||||||
|
PLUGIN_PACKAGE_CACHE_PATH=plugin_packages
|
||||||
|
PLUGIN_MEDIA_CACHE_PATH=assets
|
||||||
|
# Plugin oss bucket
|
||||||
|
PLUGIN_STORAGE_OSS_BUCKET=
|
||||||
|
# Plugin oss s3 credentials
|
||||||
|
PLUGIN_S3_USE_AWS=false
|
||||||
|
PLUGIN_S3_USE_AWS_MANAGED_IAM=false
|
||||||
|
PLUGIN_S3_ENDPOINT=
|
||||||
|
PLUGIN_S3_USE_PATH_STYLE=false
|
||||||
|
PLUGIN_AWS_ACCESS_KEY=
|
||||||
|
PLUGIN_AWS_SECRET_KEY=
|
||||||
|
PLUGIN_AWS_REGION=
|
||||||
|
# Plugin oss azure blob
|
||||||
|
PLUGIN_AZURE_BLOB_STORAGE_CONTAINER_NAME=
|
||||||
|
PLUGIN_AZURE_BLOB_STORAGE_CONNECTION_STRING=
|
||||||
|
# Plugin oss tencent cos
|
||||||
|
PLUGIN_TENCENT_COS_SECRET_KEY=
|
||||||
|
PLUGIN_TENCENT_COS_SECRET_ID=
|
||||||
|
PLUGIN_TENCENT_COS_REGION=
|
||||||
|
# Plugin oss aliyun oss
|
||||||
|
PLUGIN_ALIYUN_OSS_REGION=
|
||||||
|
PLUGIN_ALIYUN_OSS_ENDPOINT=
|
||||||
|
PLUGIN_ALIYUN_OSS_ACCESS_KEY_ID=
|
||||||
|
PLUGIN_ALIYUN_OSS_ACCESS_KEY_SECRET=
|
||||||
|
PLUGIN_ALIYUN_OSS_AUTH_VERSION=v4
|
||||||
|
PLUGIN_ALIYUN_OSS_PATH=
|
||||||
|
# Plugin oss volcengine tos
|
||||||
|
PLUGIN_VOLCENGINE_TOS_ENDPOINT=
|
||||||
|
PLUGIN_VOLCENGINE_TOS_ACCESS_KEY=
|
||||||
|
PLUGIN_VOLCENGINE_TOS_SECRET_KEY=
|
||||||
|
PLUGIN_VOLCENGINE_TOS_REGION=
|
||||||
|
|
||||||
|
# ------------------------------
|
||||||
|
# OTLP Collector Configuration
|
||||||
|
# ------------------------------
|
||||||
|
ENABLE_OTEL=false
|
||||||
|
OTLP_BASE_ENDPOINT=http://localhost:4318
|
||||||
|
OTLP_API_KEY=
|
||||||
|
OTEL_EXPORTER_OTLP_PROTOCOL=
|
||||||
|
OTEL_EXPORTER_TYPE=otlp
|
||||||
|
OTEL_SAMPLING_RATE=0.1
|
||||||
|
OTEL_BATCH_EXPORT_SCHEDULE_DELAY=5000
|
||||||
|
OTEL_MAX_QUEUE_SIZE=2048
|
||||||
|
OTEL_MAX_EXPORT_BATCH_SIZE=512
|
||||||
|
OTEL_METRIC_EXPORT_INTERVAL=60000
|
||||||
|
OTEL_BATCH_EXPORT_TIMEOUT=10000
|
||||||
|
OTEL_METRIC_EXPORT_TIMEOUT=30000
|
||||||
|
|
||||||
|
# Prevent Clickjacking
|
||||||
|
ALLOW_EMBED=false
|
||||||
|
|
||||||
|
# Dataset queue monitor configuration
|
||||||
|
QUEUE_MONITOR_THRESHOLD=200
|
||||||
|
# You can configure multiple ones, separated by commas. eg: test1@dify.ai,test2@dify.ai
|
||||||
|
QUEUE_MONITOR_ALERT_EMAILS=
|
||||||
|
# Monitor interval in minutes, default is 30 minutes
|
||||||
|
QUEUE_MONITOR_INTERVAL=30
|
||||||
@@ -1,4 +0,0 @@
|
|||||||
FROM couchbase/server:latest AS stage_base
|
|
||||||
# FROM couchbase:latest AS stage_base
|
|
||||||
COPY init-cbserver.sh /opt/couchbase/init/
|
|
||||||
RUN chmod +x /opt/couchbase/init/init-cbserver.sh
|
|
||||||
@@ -1 +0,0 @@
|
|||||||
ALTER SYSTEM SET ob_vector_memory_limit_percentage = 30;
|
|
||||||
@@ -0,0 +1,4 @@
|
|||||||
|
FROM couchbase/server:latest AS stage_base
|
||||||
|
# FROM couchbase:latest AS stage_base
|
||||||
|
COPY init-cbserver.sh /opt/couchbase/init/
|
||||||
|
RUN chmod +x /opt/couchbase/init/init-cbserver.sh
|
||||||
@@ -1,12 +1,5 @@
|
|||||||
additionalProperties:
|
additionalProperties:
|
||||||
formFields:
|
formFields:
|
||||||
- default: "/home/dify"
|
|
||||||
edit: true
|
|
||||||
envKey: DIFY_ROOT_PATH
|
|
||||||
labelZh: 数据持久化路径
|
|
||||||
labelEn: Data persistence path
|
|
||||||
required: true
|
|
||||||
type: text
|
|
||||||
- default: 8080
|
- default: 8080
|
||||||
edit: true
|
edit: true
|
||||||
envKey: PANEL_APP_PORT_HTTP
|
envKey: PANEL_APP_PORT_HTTP
|
||||||
@@ -23,60 +16,3 @@ additionalProperties:
|
|||||||
required: true
|
required: true
|
||||||
rule: paramPort
|
rule: paramPort
|
||||||
type: number
|
type: number
|
||||||
- default: 5432
|
|
||||||
edit: true
|
|
||||||
envKey: EXPOSE_DB_PORT
|
|
||||||
labelZh: 数据库端口
|
|
||||||
labelEn: Database port
|
|
||||||
required: true
|
|
||||||
rule: paramPort
|
|
||||||
type: number
|
|
||||||
- default: 5003
|
|
||||||
edit: true
|
|
||||||
envKey: EXPOSE_PLUGIN_DEBUGGING_PORT
|
|
||||||
labelZh: 插件调试端口
|
|
||||||
labelEn: Plugin debugging port
|
|
||||||
required: true
|
|
||||||
rule: paramPort
|
|
||||||
type: number
|
|
||||||
- default: 19530
|
|
||||||
disabled: true
|
|
||||||
edit: true
|
|
||||||
envKey: MILVUS_STANDALONE_API_PORT
|
|
||||||
labelZh: Milvus 接口端口
|
|
||||||
labelEn: Milvus API port
|
|
||||||
required: true
|
|
||||||
rule: paramPort
|
|
||||||
type: number
|
|
||||||
- default: 9091
|
|
||||||
disabled: true
|
|
||||||
envKey: MILVUS_STANDALONE_SERVER_PORT
|
|
||||||
labelZh: Milvus 服务端口
|
|
||||||
labelEn: Milvus server port
|
|
||||||
required: true
|
|
||||||
rule: paramPort
|
|
||||||
type: number
|
|
||||||
- default: 8123
|
|
||||||
edit: true
|
|
||||||
envKey: MYSCALE_PORT
|
|
||||||
labelZh: MyScale 端口
|
|
||||||
labelEn: MyScale port
|
|
||||||
required: true
|
|
||||||
rule: paramPort
|
|
||||||
type: number
|
|
||||||
- default: 9200
|
|
||||||
edit: true
|
|
||||||
envKey: ELASTICSEARCH_PORT
|
|
||||||
labelZh: Elasticsearch 端口
|
|
||||||
labelEn: Elasticsearch port
|
|
||||||
required: true
|
|
||||||
rule: paramPort
|
|
||||||
type: number
|
|
||||||
- default: 5601
|
|
||||||
edit: true
|
|
||||||
envKey: KIBANA_PORT
|
|
||||||
labelZh: Kibana 端口
|
|
||||||
labelEn: Kibana port
|
|
||||||
required: true
|
|
||||||
rule: paramPort
|
|
||||||
type: number
|
|
||||||
|
|||||||
+295
-1649
File diff suppressed because it is too large
Load Diff
Vendored
+1140
File diff suppressed because it is too large
Load Diff
@@ -1 +0,0 @@
|
|||||||
ENV_FILE=.env
|
|
||||||
@@ -1 +0,0 @@
|
|||||||
TZ=Asia/Shanghai
|
|
||||||
+4
-1
@@ -39,7 +39,10 @@ server {
|
|||||||
proxy_pass http://web:3000;
|
proxy_pass http://web:3000;
|
||||||
include proxy.conf;
|
include proxy.conf;
|
||||||
}
|
}
|
||||||
|
location /mcp {
|
||||||
|
proxy_pass http://api:5001;
|
||||||
|
include proxy.conf;
|
||||||
|
}
|
||||||
# placeholder for acme challenge location
|
# placeholder for acme challenge location
|
||||||
${ACME_CHALLENGE_LOCATION}
|
${ACME_CHALLENGE_LOCATION}
|
||||||
|
|
||||||
+4
-1
@@ -1,5 +1,7 @@
|
|||||||
#!/bin/bash
|
#!/bin/bash
|
||||||
|
|
||||||
|
HTTPS_CONFIG=''
|
||||||
|
|
||||||
if [ "${NGINX_HTTPS_ENABLED}" = "true" ]; then
|
if [ "${NGINX_HTTPS_ENABLED}" = "true" ]; then
|
||||||
# Check if the certificate and key files for the specified domain exist
|
# Check if the certificate and key files for the specified domain exist
|
||||||
if [ -n "${CERTBOT_DOMAIN}" ] && \
|
if [ -n "${CERTBOT_DOMAIN}" ] && \
|
||||||
@@ -20,6 +22,7 @@ if [ "${NGINX_HTTPS_ENABLED}" = "true" ]; then
|
|||||||
# Substitute the HTTPS_CONFIG in the default.conf.template with content from https.conf.template
|
# Substitute the HTTPS_CONFIG in the default.conf.template with content from https.conf.template
|
||||||
envsubst '${HTTPS_CONFIG}' < /etc/nginx/conf.d/default.conf.template > /etc/nginx/conf.d/default.conf
|
envsubst '${HTTPS_CONFIG}' < /etc/nginx/conf.d/default.conf.template > /etc/nginx/conf.d/default.conf
|
||||||
fi
|
fi
|
||||||
|
export HTTPS_CONFIG
|
||||||
|
|
||||||
if [ "${NGINX_ENABLE_CERTBOT_CHALLENGE}" = "true" ]; then
|
if [ "${NGINX_ENABLE_CERTBOT_CHALLENGE}" = "true" ]; then
|
||||||
ACME_CHALLENGE_LOCATION='location /.well-known/acme-challenge/ { root /var/www/html; }'
|
ACME_CHALLENGE_LOCATION='location /.well-known/acme-challenge/ { root /var/www/html; }'
|
||||||
@@ -33,7 +36,7 @@ env_vars=$(printenv | cut -d= -f1 | sed 's/^/$/g' | paste -sd, -)
|
|||||||
envsubst "$env_vars" < /etc/nginx/nginx.conf.template > /etc/nginx/nginx.conf
|
envsubst "$env_vars" < /etc/nginx/nginx.conf.template > /etc/nginx/nginx.conf
|
||||||
envsubst "$env_vars" < /etc/nginx/proxy.conf.template > /etc/nginx/proxy.conf
|
envsubst "$env_vars" < /etc/nginx/proxy.conf.template > /etc/nginx/proxy.conf
|
||||||
|
|
||||||
envsubst < /etc/nginx/conf.d/default.conf.template > /etc/nginx/conf.d/default.conf
|
envsubst "$env_vars" < /etc/nginx/conf.d/default.conf.template > /etc/nginx/conf.d/default.conf
|
||||||
|
|
||||||
# Start Nginx using the default entrypoint
|
# Start Nginx using the default entrypoint
|
||||||
exec nginx -g 'daemon off;'
|
exec nginx -g 'daemon off;'
|
||||||
@@ -0,0 +1,24 @@
|
|||||||
|
#!/bin/bash
|
||||||
|
|
||||||
|
PG_MAJOR=16
|
||||||
|
|
||||||
|
if [ "${PG_BIGM}" = "true" ]; then
|
||||||
|
# install pg_bigm
|
||||||
|
apt-get update
|
||||||
|
apt-get install -y curl make gcc postgresql-server-dev-${PG_MAJOR}
|
||||||
|
|
||||||
|
curl -LO https://github.com/pgbigm/pg_bigm/archive/refs/tags/v${PG_BIGM_VERSION}.tar.gz
|
||||||
|
tar xf v${PG_BIGM_VERSION}.tar.gz
|
||||||
|
cd pg_bigm-${PG_BIGM_VERSION} || exit 1
|
||||||
|
make USE_PGXS=1 PG_CONFIG=/usr/bin/pg_config
|
||||||
|
make USE_PGXS=1 PG_CONFIG=/usr/bin/pg_config install
|
||||||
|
|
||||||
|
cd - || exit 1
|
||||||
|
rm -rf v${PG_BIGM_VERSION}.tar.gz pg_bigm-${PG_BIGM_VERSION}
|
||||||
|
|
||||||
|
# enable pg_bigm
|
||||||
|
sed -i -e 's/^#\s*shared_preload_libraries.*/shared_preload_libraries = '\''pg_bigm'\''/' /var/lib/postgresql/data/pgdata/postgresql.conf
|
||||||
|
fi
|
||||||
|
|
||||||
|
# Run the original entrypoint script
|
||||||
|
exec /usr/local/bin/docker-entrypoint.sh postgres
|
||||||
@@ -1,36 +0,0 @@
|
|||||||
#!/bin/bash
|
|
||||||
|
|
||||||
if [ -f .env ]; then
|
|
||||||
source .env
|
|
||||||
|
|
||||||
# setup-1 add default values
|
|
||||||
CURRENT_DIR=$(pwd)
|
|
||||||
sed -i '/^ENV_FILE=/d' .env
|
|
||||||
sed -i '/^GLOBAL_ENV_FILE=/d' .env
|
|
||||||
echo "ENV_FILE=${CURRENT_DIR}/.env" >> .env
|
|
||||||
echo "GLOBAL_ENV_FILE=${CURRENT_DIR}/envs/global.env" >> .env
|
|
||||||
echo "APP_ENV_FILE=${CURRENT_DIR}/envs/dify.env" >> .env
|
|
||||||
|
|
||||||
# setup-2 update dir permissions
|
|
||||||
mkdir -p "$DIFY_ROOT_PATH"
|
|
||||||
|
|
||||||
cp -r conf/. "$DIFY_ROOT_PATH/"
|
|
||||||
|
|
||||||
# setup-3 sync environment variables
|
|
||||||
env_source="envs/dify.env"
|
|
||||||
if [ -f "$env_source" ]; then
|
|
||||||
while IFS='=' read -r key value; do
|
|
||||||
if [[ -z "$key" || "$key" =~ ^# ]]; then
|
|
||||||
continue
|
|
||||||
fi
|
|
||||||
if ! grep -q "^$key=" .env; then
|
|
||||||
echo "$key=$value" >> .env
|
|
||||||
fi
|
|
||||||
done < "$env_source"
|
|
||||||
fi
|
|
||||||
|
|
||||||
echo "Check Finish."
|
|
||||||
|
|
||||||
else
|
|
||||||
echo "Error: .env file not found."
|
|
||||||
fi
|
|
||||||
@@ -1,10 +0,0 @@
|
|||||||
#!/bin/bash
|
|
||||||
|
|
||||||
if [ -f .env ]; then
|
|
||||||
source .env
|
|
||||||
|
|
||||||
echo "Check Finish."
|
|
||||||
|
|
||||||
else
|
|
||||||
echo "Error: .env file not found."
|
|
||||||
fi
|
|
||||||
@@ -1,47 +0,0 @@
|
|||||||
#!/bin/bash
|
|
||||||
|
|
||||||
if [ -f .env ]; then
|
|
||||||
source .env
|
|
||||||
|
|
||||||
# setup-1 add default values
|
|
||||||
CURRENT_DIR=$(pwd)
|
|
||||||
sed -i '/^ENV_FILE=/d' .env
|
|
||||||
sed -i '/^GLOBAL_ENV_FILE=/d' .env
|
|
||||||
echo "ENV_FILE=${CURRENT_DIR}/.env" >> .env
|
|
||||||
echo "GLOBAL_ENV_FILE=${CURRENT_DIR}/envs/global.env" >> .env
|
|
||||||
echo "APP_ENV_FILE=${CURRENT_DIR}/envs/dify.env" >> .env
|
|
||||||
|
|
||||||
# setup-2 update dir permissions
|
|
||||||
mkdir -p "$DIFY_ROOT_PATH"
|
|
||||||
|
|
||||||
if [ -d "conf" ]; then
|
|
||||||
find conf -type f | while read -r file; do
|
|
||||||
dest="$DIFY_ROOT_PATH/${file#conf/}"
|
|
||||||
if [ ! -e "$dest" ]; then
|
|
||||||
mkdir -p "$(dirname "$dest")"
|
|
||||||
cp "$file" "$dest"
|
|
||||||
fi
|
|
||||||
done
|
|
||||||
echo "Conf files copied to $DIFY_ROOT_PATH."
|
|
||||||
else
|
|
||||||
echo "Warning: conf directory not found."
|
|
||||||
fi
|
|
||||||
|
|
||||||
# setup-3 sync environment variables
|
|
||||||
env_source="envs/dify.env"
|
|
||||||
if [ -f "$env_source" ]; then
|
|
||||||
while IFS='=' read -r key value; do
|
|
||||||
if [[ -z "$key" || "$key" =~ ^# ]]; then
|
|
||||||
continue
|
|
||||||
fi
|
|
||||||
if ! grep -q "^$key=" .env; then
|
|
||||||
echo "$key=$value" >> .env
|
|
||||||
fi
|
|
||||||
done < "$env_source"
|
|
||||||
fi
|
|
||||||
|
|
||||||
echo "Check Finish."
|
|
||||||
|
|
||||||
else
|
|
||||||
echo "Error: .env file not found."
|
|
||||||
fi
|
|
||||||
+5
@@ -19,6 +19,8 @@ acl Safe_ports port 488 # gss-http
|
|||||||
acl Safe_ports port 591 # filemaker
|
acl Safe_ports port 591 # filemaker
|
||||||
acl Safe_ports port 777 # multiling http
|
acl Safe_ports port 777 # multiling http
|
||||||
acl CONNECT method CONNECT
|
acl CONNECT method CONNECT
|
||||||
|
acl allowed_domains dstdomain .marketplace.dify.ai
|
||||||
|
http_access allow allowed_domains
|
||||||
http_access deny !Safe_ports
|
http_access deny !Safe_ports
|
||||||
http_access deny CONNECT !SSL_ports
|
http_access deny CONNECT !SSL_ports
|
||||||
http_access allow localhost manager
|
http_access allow localhost manager
|
||||||
@@ -49,3 +51,6 @@ http_port ${REVERSE_PROXY_PORT} accel vhost
|
|||||||
cache_peer ${SANDBOX_HOST} parent ${SANDBOX_PORT} 0 no-query originserver
|
cache_peer ${SANDBOX_HOST} parent ${SANDBOX_PORT} 0 no-query originserver
|
||||||
acl src_all src all
|
acl src_all src all
|
||||||
http_access allow src_all
|
http_access allow src_all
|
||||||
|
|
||||||
|
# Unless the option's size is increased, an error will occur when uploading more than two files.
|
||||||
|
client_request_buffer_max_size 100 MB
|
||||||
+1
-1
@@ -5,6 +5,6 @@ create user dify identified by dify DEFAULT TABLESPACE users quota unlimited on
|
|||||||
grant DB_DEVELOPER_ROLE to dify;
|
grant DB_DEVELOPER_ROLE to dify;
|
||||||
|
|
||||||
BEGIN
|
BEGIN
|
||||||
CTX_DDL.CREATE_PREFERENCE('my_chinese_vgram_lexer','CHINESE_VGRAM_LEXER');
|
CTX_DDL.CREATE_PREFERENCE('dify.world_lexer','WORLD_LEXER');
|
||||||
END;
|
END;
|
||||||
/
|
/
|
||||||
@@ -0,0 +1 @@
|
|||||||
|
ALTER SYSTEM SET ob_vector_memory_limit_percentage = 30;
|
||||||
Reference in New Issue
Block a user