milvus/tests/_helm/values/e2e/distributed
yanliang567 bfdd6adb7b
test: Upgrade helm chart version and resource limit (#41881)
related issue: https://github.com/milvus-io/milvus/issues/41819
pr: #41820

---------

Signed-off-by: yanliang567 <yanliang.qiao@zilliz.com>
2025-05-16 13:24:23 +08:00

216 lines
4.4 KiB
Plaintext

affinity:
nodeAffinity:
preferredDuringSchedulingIgnoredDuringExecution:
- preference:
matchExpressions:
- key: node-role.kubernetes.io/e2e
operator: Exists
weight: 1
cluster:
enabled: true
proxy:
resources:
limits:
cpu: "1"
memory: 4Gi
requests:
cpu: "0.3"
memory: 256Mi
dataNode:
resources:
limits:
cpu: "2"
memory: 8Gi
requests:
cpu: "0.5"
memory: 500Mi
indexNode:
disk:
enabled: true
resources:
limits:
cpu: "2"
memory: 8Gi
requests:
cpu: "0.5"
memory: 500Mi
queryNode:
disk:
enabled: true
resources:
limits:
cpu: "2"
memory: 4Gi
requests:
cpu: "1"
memory: 2Gi
mixCoordinator:
resources:
limits:
cpu: "1"
memory: 4Gi
requests:
cpu: "0.2"
memory: 256Mi
service:
type: ClusterIP
log:
level: debug
extraConfigFiles:
user.yaml: |+
dataCoord:
gc:
interval: 1800
missingTolerance: 1800
dropTolerance: 1800
metrics:
serviceMonitor:
enabled: true
etcd:
affinity:
nodeAffinity:
preferredDuringSchedulingIgnoredDuringExecution:
- preference:
matchExpressions:
- key: node-role.kubernetes.io/e2e
operator: Exists
weight: 1
metrics:
enabled: true
podMonitor:
enabled: true
replicaCount: 1
resources:
requests:
cpu: "0.2"
memory: 256Mi
limits:
cpu: "1"
memory: 4Gi
tolerations:
- effect: NoSchedule
key: node-role.kubernetes.io/e2e
operator: Exists
image:
all:
pullPolicy: Always
repository: harbor.milvus.io/milvus/milvus
tag: PR-35426-20240812-46dadb120
minio:
affinity:
nodeAffinity:
preferredDuringSchedulingIgnoredDuringExecution:
- preference:
matchExpressions:
- key: node-role.kubernetes.io/e2e
operator: Exists
weight: 1
mode: standalone
resources:
requests:
cpu: "0.2"
memory: 512Mi
limits:
cpu: "1"
memory: 4Gi
tolerations:
- effect: NoSchedule
key: node-role.kubernetes.io/e2e
operator: Exists
pulsarv3:
enabled: true
bookkeeper:
affinity:
nodeAffinity:
preferredDuringSchedulingIgnoredDuringExecution:
- preference:
matchExpressions:
- key: node-role.kubernetes.io/e2e
operator: Exists
weight: 1
resources:
requests:
cpu: "0.1"
memory: 256Mi
limits:
cpu: "0.5"
memory: 2Gi
tolerations:
- effect: NoSchedule
key: node-role.kubernetes.io/e2e
operator: Exists
broker:
affinity:
nodeAffinity:
preferredDuringSchedulingIgnoredDuringExecution:
- preference:
matchExpressions:
- key: node-role.kubernetes.io/e2e
operator: Exists
weight: 1
replicaCount: 2
resources:
requests:
cpu: "0.1"
memory: 256Mi
limits:
cpu: "0.5"
memory: 4Gi
tolerations:
- effect: NoSchedule
key: node-role.kubernetes.io/e2e
operator: Exists
components:
autorecovery: false
proxy:
affinity:
nodeAffinity:
preferredDuringSchedulingIgnoredDuringExecution:
- preference:
matchExpressions:
- key: node-role.kubernetes.io/e2e
operator: Exists
weight: 1
resources:
requests:
cpu: "0.1"
memory: 256Mi
limits:
cpu: "0.5"
memory: 2Gi
tolerations:
- effect: NoSchedule
key: node-role.kubernetes.io/e2e
operator: Exists
wsResources:
requests:
cpu: "0.1"
memory: 256Mi
limits:
cpu: "0.5"
memory: 2Gi
zookeeper:
affinity:
nodeAffinity:
preferredDuringSchedulingIgnoredDuringExecution:
- preference:
matchExpressions:
- key: node-role.kubernetes.io/e2e
operator: Exists
weight: 1
replicaCount: 1
resources:
requests:
cpu: "0.1"
memory: 256Mi
limits:
cpu: "0.5"
memory: 2Gi
tolerations:
- effect: NoSchedule
key: node-role.kubernetes.io/e2e
operator: Exists
tolerations:
- effect: NoSchedule
key: node-role.kubernetes.io/e2e
operator: Exists