feature: 对接deer-flow (#54)

feature: 对接deer-flow
This commit is contained in:
hhhhsc701
2025-11-04 20:30:40 +08:00
committed by GitHub
parent dc30b0d892
commit f3958f08d9
710 changed files with 112812 additions and 52 deletions

View File

@@ -0,0 +1,679 @@
# Milvus Helm Chart
For more information about installing and using Helm, see the [Helm Docs](https://helm.sh/docs/). For a quick introduction to Charts, see the [Chart Guide](https://helm.sh/docs/topics/charts/).
To install Milvus, refer to [Milvus installation](https://milvus.io/docs/v2.0.x/install_cluster-helm.md).
## Introduction
This chart bootstraps Milvus deployment on a Kubernetes cluster using the Helm package manager.
## Prerequisites
- Kubernetes >= 1.20.0
- Helm >= 3.14.0
## Compatibility Notice
- **IMPORTANT** As of helm version 5.0.0, significant architectural changes have been introduced in Milvus v2.6.0:
- Coordinator consolidation: Legacy separate coordinators (dataCoord, queryCoord, indexCoord) have been consolidated into a single mixCoord
- New components: Introduction of Streaming Node for enhanced data processing
- Component removal: indexNode has been removed and consolidated
- Milvus v2.6.0-rc1 is not compatible with v2.6.x. Direct upgrades from release candidates are not supported.
- You must upgrade to v2.5.16 with mixCoordinator enabled before upgrading to v2.6.x.
- **IMPORTANT** For users using pulsar2. Please don't use version 4.2.21~4.2.29 for upgrading, there're some known issues. 4.2.30 or later version is recommended. Remember to add `--set pulsar.enabled=true,pulsarv3.enabled=false` or set them in your values file when upgrading.
- As of version 4.2.21, the Milvus Helm chart introduced pulsar-v3.x chart as dependency. For backward compatibility, please upgrade your helm to v3.14 or later version, and be sure to add the `--reset-then-reuse-values` option whenever you use `helm upgrade`.
- As of version 4.2.0, the Milvus Helm chart no longer supports Milvus v2.3.x. If you need to deploy Milvus v2.3.x using Helm, please use Milvus Helm chart version less than 4.2.0 (e.g 4.1.36).
> **IMPORTANT** The master branch is for the development of Milvus v2.x. On March 9th, 2021, we released Milvus v1.0, the first stable version of Milvus with long-term support. To use Milvus v1.x, switch to [branch 1.1](https://github.com/zilliztech/milvus-helm/tree/1.1).
## Install the Chart
1. Add the stable repository
```bash
$ helm repo add zilliztech https://zilliztech.github.io/milvus-helm/
```
2. Update charts repositories
```
$ helm repo update
```
### Deploy Milvus with standalone mode
Assume the release name is `my-release`:
```bash
# Helm v3.x
$ helm upgrade --install my-release --set cluster.enabled=false --set etcd.replicaCount=1 --set pulsarv3.enabled=false --set minio.mode=standalone zilliztech/milvus
```
By default, milvus standalone uses `woodpecker` as message queue. You can also use `pulsar` or `kafka` as message queue:
```bash
# Helm v3.x
# Milvus Standalone with pulsar as message queue
$ helm upgrade --install my-release --set cluster.enabled=false --set standalone.messageQueue=pulsar --set etcd.replicaCount=1 --set pulsarv3.enabled=true --set minio.mode=standalone zilliztech/milvus
```
```bash
# Helm v3.x
# Milvus Standalone with kafka as message queue
$ helm upgrade --install my-release --set cluster.enabled=false --set standalone.messageQueue=kafka --set etcd.replicaCount=1 --set pulsarv3.enabled=false --set kafka.enabled=true --set minio.mode=standalone zilliztech/milvus
```
If you need to use standalone mode with embedded ETCD and local storage (without starting MinIO and additional ETCD), you can use the following steps:
1. Prepare a values file
```
cat > values.yaml <<EOF
---
cluster:
enabled: false
etcd:
enabled: false
pulsarv3:
enabled: false
minio:
enabled: false
tls:
enabled: false
extraConfigFiles:
user.yaml: |+
etcd:
use:
embed: true
data:
dir: /var/lib/milvus/etcd
common:
storageType: local
EOF
```
2. Helm install with this values file
```
helm upgrade --install -f values.yaml my-release zilliztech/milvus
```
> **Tip**: To list all releases, using `helm list`.
### Deploy Milvus with cluster mode
Assume the release name is `my-release`:
```bash
# Helm v3.x
$ helm upgrade --install my-release zilliztech/milvus
```
By default, milvus cluster uses `pulsar` as message queue. You can also use `kafka` instead of `pulsar` for milvus cluster:
```bash
# Helm v3.x
$ helm upgrade --install my-release zilliztech/milvus --set pulsarv3.enabled=false --set kafka.enabled=true
```
By default, milvus cluster uses `mixCoordinator` which contains all coordinators. This is the recommended deployment approach for Milvus v2.6.x and later versions.
### Upgrading from Milvus v2.5.x to v2.6.x
Upgrading from Milvus 2.5.x to 2.6.x involves significant architectural changes. Please follow these steps carefully:
#### Requirements
**System requirements:**
- Helm version >= 3.14.0
- Kubernetes version >= 1.20.0
- Milvus cluster deployed via Helm Chart
**Compatibility requirements:**
- Milvus v2.6.0-rc1 is not compatible with v2.6.x. Direct upgrades from release candidates are not supported.
- If you are currently running v2.6.0-rc1 and need to preserve your data, please refer to community guides for migration assistance.
- You must upgrade to v2.5.16 with mixCoordinator enabled before upgrading to v2.6.x.
#### Upgrade Process
**Step 1: Upgrade Helm Chart**
First, upgrade your Milvus Helm chart repository:
```bash
helm repo add zilliztech https://zilliztech.github.io/milvus-helm
helm repo update zilliztech
```
**Step 2: Upgrade to v2.5.16 with mixCoordinator**
Check if your cluster currently uses separate coordinators:
```bash
kubectl get pods
```
If you see separate coordinator pods (datacoord, querycoord, indexcoord), upgrade to v2.5.16 and enable mixCoordinator:
```bash
helm upgrade my-release zilliztech/milvus \
--set image.all.tag="v2.5.16" \
--set mixCoordinator.enabled=true \
--set rootCoordinator.enabled=false \
--set indexCoordinator.enabled=false \
--set queryCoordinator.enabled=false \
--set dataCoordinator.enabled=false \
--reset-then-reuse-values \
--version=4.2.58
```
If your cluster already uses mixCoordinator, simply upgrade the image:
```bash
helm upgrade my-release zilliztech/milvus \
--set image.all.tag="v2.5.16" \
--reset-then-reuse-values \
--version=4.2.58
```
**Step 3: Upgrade to v2.6.x**
Once v2.5.16 is running successfully with mixCoordinator, upgrade to v2.6.x:
```bash
helm upgrade my-release zilliztech/milvus \
--set image.all.tag="v2.6.x" \
--set streaming.enabled=true \
--set indexNode.enabled=false \
--reset-then-reuse-values \
--version=5.0.0
```
### Upgrade an existing Milvus cluster (General)
> **IMPORTANT** If you have installed a milvus cluster with version below v2.1.x, you need follow the instructions at here: https://github.com/milvus-io/milvus/blob/master/deployments/migrate-meta/README.md. After meta migration, you use `helm upgrade` to update your cluster again.
E.g. to scale out query node from 1(default) to 2:
```bash
# Helm v3.14.0+
$ helm upgrade --reset-then-reuse-values --install --set queryNode.replicas=2 my-release zilliztech/milvus
```
### Milvus Mix Coordinator Active Standby
> **IMPORTANT** Milvus supports deploying multiple Mix Coordinator instances for high availability. For example, you can run a Milvus cluster with two mixcoord pods:
```bash
helm upgrade --install my-release zilliztech/milvus --set mixCoordinator.activeStandby.enabled=true --set mixCoordinator.replicas=2
```
### Breaking Changes
> **IMPORTANT** Milvus helm chart 4.0.0 has breaking changes for milvus configuration. Previously, you can set segment size like this `--set dataCoordinator.segment.maxSize=1024`. Now we have remove all the shortcut config option. Instead, you can set using `extraConfigFiles` like this:
```bash
extraConfigFiles:
user.yaml: |+
dataCoord:
segment:
maxSize: 1024
```
So if you had deployed a cluster with helm chart version below 4.0.0 and also specified extra config, you need set the configs under `extraConfigFiles` when running `helm upgrade`.
> **IMPORTANT** Milvus has removed mysql as meta store support from v2.3.1. And milvus helm chart has also removed mysql as dependency from chart version 4.1.8.
### Enable log to file
By default, all the logs of milvus components will output stdout. If you wanna log to file, you'd install milvus with `--set log.persistence.enabled=true`. Note that you should have a storageclass with `ReadWriteMany` access modes.
```bash
# Install a milvus cluster with file log output
helm install my-release zilliztech/milvus --set log.persistence.enabled=true --set log.persistence.persistentVolumeClaim.storageClass=<read-write-many-storageclass>
```
It will output log to `/milvus/logs/` directory.
### Enable proxy tls connection
By default the TLS connection to proxy service is false, to enable TLS with users' own certificate and privatge key, it can be specified in `extraConfigFiles` like this:
```bash
extraConfigFiles:
user.yaml: |+
# Enable tlsMode and set the tls cert and key
tls:
serverPemPath: /etc/milvus/certs/tls.crt
serverKeyPath: /etc/milvus/certs/tls.key
common:
security:
tlsMode: 1
```
The path specified above are TLS secret data mounted inside the proxy pod as files. To create a TLS secret, set `proxy.tls.enabled` to `true` then provide base64-encoded values for your certificate and private key files in values.yaml:
```bash
proxy:
enabled: true
tls:
enabled: true
secretName: milvus-tls
#expecting base64 encoded values here: i.e. $(cat tls.crt | base64 -w 0) and $(cat tls.key | base64 -w 0)
key: LS0tLS1C....
crt: LS0tLS1CR...
```
or in cli using --set:
```bash
--set proxy.tls.enabled=true \
--set prox.tls.key=$(cat /path/to/private_key_file | base64 -w 0) \
--set prox.tls.crt=$(cat /path/to/certificate_file | base64 -w 0)
```
In case you want to use a different `secretName` or mount path inside pod, modify `prox.tls.secretName` above, and `serverPemPath` and `serverPemPath` in `extraConfigFles `accordingly, then in the `volume` and `volumeMounts` sections in values.yaml
```bash
volumes:
- secret:
secretName: Your-tls-secret-name
name: milvus-tls
volumeMounts:
- mountPath: /Your/tls/files/path/
name: milvus-tls
```
## Milvus with External Object Storage
As of https://github.com/minio/minio/releases/tag/RELEASE.2022-10-29T06-21-33Z, the MinIO Gateway and the related filesystem mode code have been removed. It is now recommended to utilize the `externalS3` configuration for integrating with various object storage services. Notably, Milvus now provides support for popular object storage platforms such as AWS S3, GCP GCS, Azure Blob, Aliyun OSS and Tencent COS.
The recommended configuration option for `externalS3.cloudProvider` includes the following choices: `aws`, `gcp`, `azure`, `aliyun`, and `tencent`. Here's an example to use AWS S3 for Milvus object storage:
```
minio:
enabled: false
externalS3:
enabled: true
cloudProvider: aws
host: s3.aws.com
port: 443
useSSL: true
bucketName: <bucket-name>
accessKey: <s3-access-key>
secretKey: <s3-secret-key>
```
## Uninstall the Chart
```bash
# Helm v3.x
$ helm uninstall my-release
```
The command removes all the Kubernetes components associated with the chart and deletes the release.
- Completely uninstall Milvus
> **IMPORTANT** Please run this command with care. Maybe you want to keep ETCD data
```bash
MILVUS_LABELS="app.kubernetes.io/instance=my-release"
kubectl delete pvc $(kubectl get pvc -l "${MILVUS_LABELS}" -o jsonpath='{range.items[*]}{.metadata.name} ')
```
## Configuration
### Milvus Service Configuration
The following table lists the configurable parameters of the Milvus Service and their default values.
| Parameter | Description | Default |
|-------------------------------------------|-----------------------------------------------|---------------------------------------------------------|
| `cluster.enabled` | Enable or disable Milvus Cluster mode | `true` |
| `route.enabled` | Enable OpenShift route | `false` |
| `route.hostname` | Hostname for the OpenShift route | `""` |
| `route.path` | Path for the OpenShift route | `""` |
| `route.termination` | TLS termination for the route | `edge` |
| `route.annotations` | Route annotations | `{}` |
| `route.labels` | Route labels | `{}` |
| `image.all.repository` | Image repository | `milvusdb/milvus` |
| `image.all.tag` | Image tag | `v2.6.4` |
| `image.all.pullPolicy` | Image pull policy | `IfNotPresent` |
| `image.all.pullSecrets` | Image pull secrets | `{}` |
| `image.tools.repository` | Config image repository | `milvusdb/milvus-config-tool` |
| `image.tools.tag` | Config image tag | `v0.1.2` |
| `image.tools.pullPolicy` | Config image pull policy | `IfNotPresent` |
| `customConfigMap` | User specified ConfigMap for configuration |
| `extraConfigFiles` | Extra config to override default milvus.yaml | `user.yaml:` |
| `service.type` | Service type | `ClusterIP` |
| `service.port` | Port where service is exposed | `19530` |
| `service.portName` | Useful for [Istio protocol selection](https://istio.io/latest/docs/ops/configuration/traffic-management/protocol-selection/) | `milvus` |
| `service.nodePort` | Service nodePort | `unset` |
| `service.annotations` | Service annotations | `{}` |
| `service.labels` | Service custom labels | `{}` |
| `service.clusterIP` | Internal cluster service IP | `unset` |
| `service.loadBalancerIP` | IP address to assign to load balancer (if supported) | `unset` |
| `service.loadBalancerSourceRanges` | List of IP CIDRs allowed access to lb (if supported) | `[]` |
| `service.externalIPs` | Service external IP addresses | `[]` |
| `ingress.enabled` | If true, Ingress will be created | `false` |
| `ingress.annotations` | Ingress annotations | `{}` |
| `ingress.labels` | Ingress labels | `{}` |
| `ingress.rules` | Ingress rules | `[]` |
| `ingress.tls` | Ingress TLS configuration | `[]` |
| `serviceAccount.create` | Create a custom service account | `false` |
| `serviceAccount.name` | Service Account name | `milvus` |
| `serviceAccount.annotations` | Service Account Annotations | `{}` |
| `serviceAccount.labels` | Service Account labels | `{}` |
| `metrics.enabled` | Export Prometheus monitoring metrics | `true` |
| `metrics.serviceMonitor.enabled` | Create ServiceMonitor for Prometheus operator | `false` |
| `metrics.serviceMonitor.additionalLabels` | Additional labels that can be used so ServiceMonitor will be discovered by Prometheus | `unset` |
| `log.level` | Logging level to be used. Valid levels are `debug`, `info`, `warn`, `error`, `fatal` | `info` |
| `log.file.maxSize` | The size limit of the log file (MB) | `300` |
| `log.file.maxAge` | The maximum number of days that the log is retained. (day) | `10` |
| `log.file.maxBackups` | The maximum number of retained logs. | `20` |
| `log.format` | Format used for the logs. Valid formats are `text` and `json` | `text` |
| `log.persistence.enabled` | Use persistent volume to store Milvus logs data | `false` |
| `log.persistence.mountPath` | Milvus logs data persistence volume mount path | `/milvus/logs` |
| `log.persistence.annotations` | PersistentVolumeClaim annotations | `{}` |
| `log.persistence.persistentVolumeClaim.existingClaim` | Use your own data Persistent Volume existing claim name | `unset` |
| `log.persistence.persistentVolumeClaim.storageClass` | The Milvus logs data Persistent Volume Storage Class | `unset` |
| `log.persistence.persistentVolumeClaim.accessModes` | The Milvus logs data Persistence access modes | `ReadWriteOnce` |
| `log.persistence.persistentVolumeClaim.size` | The size of Milvus logs data Persistent Volume Storage Class | `5Gi` |
| `log.persistence.persistentVolumeClaim.subPath` | SubPath for Milvus logs data mount | `unset` |
| `externalS3.enabled` | Enable or disable external S3 | `false` |
| `externalS3.host` | The host of the external S3 | `unset` |
| `externalS3.port` | The port of the external S3 | `unset` |
| `externalS3.rootPath` | The path prefix of the external S3 | `unset` |
| `externalS3.accessKey` | The Access Key of the external S3 | `unset` |
| `externalS3.secretKey` | The Secret Key of the external S3 | `unset` |
| `externalS3.bucketName` | The Bucket Name of the external S3 | `unset` |
| `externalS3.useSSL` | If true, use SSL to connect to the external S3 | `false` |
| `externalS3.useIAM` | If true, use iam to connect to the external S3 | `false` |
| `externalS3.cloudProvider` | When `useIAM` enabled, only "aws" & "gcp" is supported for now | `aws` |
| `externalS3.iamEndpoint` | The IAM endpoint of the external S3 | `` |
| `externalS3.region` | The region of the external S3 | `` |
| `externalS3.useVirtualHost` | If true, the external S3 whether use virtual host bucket mode | `` |
| `externalEtcd.enabled` | Enable or disable external Etcd | `false` |
| `externalEtcd.endpoints` | The endpoints of the external etcd | `{}` |
| `externalPulsar.enabled` | Enable or disable external Pulsar | `false` |
| `externalPulsar.host` | The host of the external Pulsar | `localhost` |
| `externalPulsar.port` | The port of the external Pulsar | `6650` |
| `externalPulsar.tenant` | The tenant of the external Pulsar | `public` |
| `externalPulsar.namespace` | The namespace of the external Pulsar | `default` |
| `externalPulsar.authPlugin` | The authPlugin of the external Pulsar | `""` |
| `externalPulsar.authParams` | The authParams of the external Pulsar | `""` |
| `externalKafka.enabled` | Enable or disable external Kafka | `false` |
| `externalKafka.brokerList` | The brokerList of the external Kafka separated by comma | `localhost:9092` |
| `externalKafka.securityProtocol` | The securityProtocol used for kafka authentication | `SASL_SSL` |
| `externalKafka.sasl.mechanisms` | SASL mechanism to use for kafka authentication | `PLAIN` |
| `externalKafka.sasl.username` | username for PLAIN or SASL/PLAIN authentication | `` |
| `externalKafka.sasl.password` | password for PLAIN or SASL/PLAIN authentication | `` |
### Milvus Standalone Deployment Configuration
The following table lists the configurable parameters of the Milvus Standalone component and their default values.
| Parameter | Description | Default |
|-------------------------------------------|-----------------------------------------------|---------------------------------------------------------|
| `standalone.resources` | Resource requests/limits for the Milvus Standalone pods | `{}` |
| `standalone.nodeSelector` | Node labels for Milvus Standalone pods assignment | `{}` |
| `standalone.affinity` | Affinity settings for Milvus Standalone pods assignment | `{}` |
| `standalone.tolerations` | Toleration labels for Milvus Standalone pods assignment | `[]` |
| `standalone.heaptrack.enabled` | Whether to enable heaptrack | `false` |
| `standalone.disk.enabled` | Whether to enable disk | `true` |
| `standalone.profiling.enabled` | Whether to enable live profiling | `false` |
| `standalone.extraEnv` | Additional Milvus Standalone container environment variables | `[]` |
| `standalone.messageQueue` | Message queue for Milvus Standalone: rocksmq, natsmq, pulsar, kafka | `rocksmq` |
| `standalone.persistence.enabled` | Use persistent volume to store Milvus standalone data | `true` |
| `standalone.persistence.mountPath` | Milvus standalone data persistence volume mount path | `/var/lib/milvus` |
| `standalone.persistence.annotations` | PersistentVolumeClaim annotations | `{}` |
| `standalone.persistence.persistentVolumeClaim.existingClaim` | Use your own data Persistent Volume existing claim name | `unset` |
| `standalone.persistence.persistentVolumeClaim.storageClass` | The Milvus standalone data Persistent Volume Storage Class | `unset` |
| `standalone.persistence.persistentVolumeClaim.accessModes` | The Milvus standalone data Persistence access modes | `ReadWriteOnce` |
| `standalone.persistence.persistentVolumeClaim.size` | The size of Milvus standalone data Persistent Volume Storage Class | `5Gi` |
| `standalone.persistence.persistentVolumeClaim.subPath` | SubPath for Milvus standalone data mount | `unset` |
### Milvus Proxy Deployment Configuration
The following table lists the configurable parameters of the Milvus Proxy component and their default values.
| Parameter | Description | Default |
|-------------------------------------------|---------------------------------------------------------|---------------|
| `proxy.enabled` | Enable or disable Milvus Proxy Deployment | `true` |
| `proxy.replicas` | Desired number of Milvus Proxy pods | `1` |
| `proxy.resources` | Resource requests/limits for the Milvus Proxy pods | `{}` |
| `proxy.nodeSelector` | Node labels for Milvus Proxy pods assignment | `{}` |
| `proxy.affinity` | Affinity settings for Milvus Proxy pods assignment | `{}` |
| `proxy.tolerations` | Toleration labels for Milvus Proxy pods assignment | `[]` |
| `proxy.heaptrack.enabled` | Whether to enable heaptrack | `false` |
| `proxy.profiling.enabled` | Whether to enable live profiling | `false` |
| `proxy.extraEnv` | Additional Milvus Proxy container environment variables | `[]` |
| `proxy.http.enabled` | Enable rest api for Milvus Proxy | `true` |
| `proxy.maxUserNum` | Modify the Milvus maximum user limit | `100` |
| `proxy.maxRoleNum` | Modify the Milvus maximum role limit | `10` |
| `proxy.http.debugMode.enabled` | Enable debug mode for rest api | `false` |
| `proxy.tls.enabled` | Enable porxy tls connection | `false` |
| `proxy.strategy` | Deployment strategy configuration | RollingUpdate |
| `proxy.annotations` | Additional pod annotations | `{}` |
| `proxy.hpa` | Enable hpa for proxy node | false |
| `proxy.minReplicas` | Specify the minimum number of replicas | 1 |
| `proxy.maxReplicas` | Specify the maximum number of replicas | 5 |
| `proxy.cpuUtilization` | Specify the cpu auto-scaling value | 40 |
### Milvus Query Node Deployment Configuration
The following table lists the configurable parameters of the Milvus Query Node component and their default values.
| Parameter | Description | Default |
|-------------------------------------------|-----------------------------------------------|---------------------------------------------------------|
| `queryNode.enabled` | Enable or disable Milvus Query Node component | `true` |
| `queryNode.replicas` | Desired number of Milvus Query Node pods | `1` |
| `queryNode.resources` | Resource requests/limits for the Milvus Query Node pods | `{}` |
| `queryNode.nodeSelector` | Node labels for Milvus Query Node pods assignment | `{}` |
| `queryNode.affinity` | Affinity settings for Milvus Query Node pods assignment | `{}` |
| `queryNode.tolerations` | Toleration labels for Milvus Query Node pods assignment | `[]` |
| `queryNode.heaptrack.enabled` | Whether to enable heaptrack | `false` |
| `queryNode.disk.enabled` | Whether to enable disk for query | `true` |
| `queryNode.profiling.enabled` | Whether to enable live profiling | `false` |
| `queryNode.extraEnv` | Additional Milvus Query Node container environment variables | `[]` |
| `queryNode.strategy` | Deployment strategy configuration | RollingUpdate |
| `queryNode.annotations` | Additional pod annotations | `{}` |
| `queryNode.hpa` | Enable hpa for query node | false |
| `queryNode.minReplicas` | Specify the minimum number of replicas | 1 |
| `queryNode.maxReplicas` | Specify the maximum number of replicas | 5 |
| `queryNode.cpuUtilization` | Specify the cpu auto-scaling value | 40 |
| `queryNode.memoryUtilization` | Specify the memory auto-scaling value | 60 |
### Milvus Data Node Deployment Configuration
The following table lists the configurable parameters of the Milvus Data Node component and their default values.
| Parameter | Description | Default |
|-------------------------------------------|-----------------------------------------------|---------------------------------------------------------|
| `dataNode.enabled` | Enable or disable Data Node component | `true` |
| `dataNode.replicas` | Desired number of Data Node pods | `1` |
| `dataNode.resources` | Resource requests/limits for the Milvus Data Node pods | `{}` |
| `dataNode.nodeSelector` | Node labels for Milvus Data Node pods assignment | `{}` |
| `dataNode.affinity` | Affinity settings for Milvus Data Node pods assignment | `{}` |
| `dataNode.tolerations` | Toleration labels for Milvus Data Node pods assignment | `[]` |
| `dataNode.heaptrack.enabled` | Whether to enable heaptrack | `false` |
| `dataNode.profiling.enabled` | Whether to enable live profiling | `false` |
| `dataNode.extraEnv` | Additional Milvus Data Node container environment variables | `[]` |
| `dataNode.strategy` | Deployment strategy configuration | RollingUpdate |
| `dataNode.annotations` | Additional pod annotations | `{}` |
| `dataNode.hpa` | Enable hpa for data node | false |
| `dataNode.minReplicas` | Specify the minimum number of replicas | 1 |
| `dataNode.maxReplicas` | Specify the maximum number of replicas | 5 |
| `dataNode.cpuUtilization` | Specify the cpu auto-scaling value | 40 |
### Milvus Mix Coordinator Deployment Configuration
The following table lists the configurable parameters of the Milvus Mix Coordinator component and their default values. The Mix Coordinator consolidates all coordinator functions (Root, Query, Index, and Data Coordinators) into a single component for improved efficiency and simplified deployment.
| Parameter | Description | Default |
|-------------------------------------------|-----------------------------------------------|---------------------------------------------------------|
| `mixCoordinator.enabled` | Enable or disable Mix Coordinator component | `true` |
| `mixCoordinator.replicas` | Desired number of Mix Coordinator pods | `1` |
| `mixCoordinator.resources` | Resource requests/limits for the Milvus Mix Coordinator pods | `{}` |
| `mixCoordinator.nodeSelector` | Node labels for Milvus Mix Coordinator pods assignment | `{}` |
| `mixCoordinator.affinity` | Affinity settings for Milvus Mix Coordinator pods assignment | `{}` |
| `mixCoordinator.tolerations` | Toleration labels for Milvus Mix Coordinator pods assignment | `[]` |
| `mixCoordinator.heaptrack.enabled` | Whether to enable heaptrack | `false` |
| `mixCoordinator.profiling.enabled` | Whether to enable live profiling | `false` |
| `mixCoordinator.activeStandby.enabled` | Whether to enable active-standby | `false` |
| `mixCoordinator.extraEnv` | Additional Milvus Mix Coordinator container environment variables | `[]` |
| `mixCoordinator.service.type` | Service type | `ClusterIP` |
| `mixCoordinator.service.annotations` | Service annotations | `{}` |
| `mixCoordinator.service.labels` | Service custom labels | `{}` |
| `mixCoordinator.service.clusterIP` | Internal cluster service IP | `unset` |
| `mixCoordinator.service.loadBalancerIP` | IP address to assign to load balancer (if supported) | `unset` |
| `mixCoordinator.service.loadBalancerSourceRanges` | List of IP CIDRs allowed access to lb (if supported) | `[]` |
| `mixCoordinator.service.externalIPs` | Service external IP addresses | `[]` |
| `mixCoordinator.strategy` | Deployment strategy configuration | RollingUpdate |
| `mixCoordinator.annotations` | Additional pod annotations | `{}` |
### Milvus Streaming Node Deployment Configuration
The following table lists the configurable parameters of the Milvus Streaming Node component and their default values. The Streaming Node is a new component introduced in Milvus v2.6.x for enhanced data processing and streaming operations.
| Parameter | Description | Default |
|-------------------------------------------|---------------------------------------------------------|---------------|
| `streamingNode.enabled` | Enable or disable Streaming Node component | `true` |
| `streamingNode.replicas` | Desired number of Streaming Node pods | `1` |
| `streamingNode.resources` | Resource requests/limits for the Milvus Streaming Node pods | `{}` |
| `streamingNode.nodeSelector` | Node labels for Milvus Streaming Node pods assignment | `{}` |
| `streamingNode.affinity` | Affinity settings for Milvus Streaming Node pods assignment | `{}` |
| `streamingNode.tolerations` | Toleration labels for Milvus Streaming Node pods assignment | `[]` |
| `streamingNode.securityContext` | Security context for Streaming Node pods | `{}` |
| `streamingNode.containerSecurityContext` | Container security context for Streaming Node | `{}` |
| `streamingNode.heaptrack.enabled` | Whether to enable heaptrack | `false` |
| `streamingNode.profiling.enabled` | Whether to enable live profiling | `false` |
| `streamingNode.extraEnv` | Additional Milvus Streaming Node container environment variables | `[]` |
| `streamingNode.strategy` | Deployment strategy configuration | `{}` |
### TEI Configuration
The following table lists the configurable parameters of the Text Embeddings Inference (TEI) component and their default values.
| Parameter | Description | Default |
|-------------------------------------------|---------------------------------------------------------|---------------------------------------------|
| `tei.enabled` | Enable or disable TEI deployment | `false` |
| `tei.name` | Name of the TEI component | `text-embeddings-inference` |
| `tei.image.repository` | TEI image repository | `ghcr.io/huggingface/text-embeddings-inference` |
| `tei.image.tag` | TEI image tag | `cpu-1.6` |
| `tei.image.pullPolicy` | TEI image pull policy | `IfNotPresent` |
| `tei.service.type` | TEI service type | `ClusterIP` |
| `tei.service.port` | TEI service port | `8080` |
| `tei.service.annotations` | TEI service annotations | `{}` |
| `tei.service.labels` | TEI service custom labels | `{}` |
| `tei.resources.requests.cpu` | CPU resource requests for TEI pods | `4` |
| `tei.resources.requests.memory` | Memory resource requests for TEI pods | `8Gi` |
| `tei.resources.limits.cpu` | CPU resource limits for TEI pods | `8` |
| `tei.resources.limits.memory` | Memory resource limits for TEI pods | `16Gi` |
| `tei.persistence.enabled` | Enable persistence for TEI | `true` |
| `tei.persistence.mountPath` | Mount path for TEI persistence | `/data` |
| `tei.persistence.annotations` | Annotations for TEI PVC | `{}` |
| `tei.persistence.persistentVolumeClaim.existingClaim` | Existing PVC for TEI | `""` |
| `tei.persistence.persistentVolumeClaim.storageClass` | Storage class for TEI PVC | `""` |
| `tei.persistence.persistentVolumeClaim.accessModes` | Access modes for TEI PVC | `ReadWriteOnce` |
| `tei.persistence.persistentVolumeClaim.size` | Size of TEI PVC | `50Gi` |
| `tei.persistence.persistentVolumeClaim.subPath` | SubPath for TEI PVC | `""` |
| `tei.modelId` | Model ID for TEI | `BAAI/bge-large-en-v1.5` |
| `tei.extraArgs` | Additional arguments for TEI | `[]` |
| `tei.nodeSelector` | Node labels for TEI pods assignment | `{}` |
| `tei.affinity` | Affinity settings for TEI pods assignment | `{}` |
| `tei.tolerations` | Toleration labels for TEI pods assignment | `[]` |
| `tei.topologySpreadConstraints` | Topology spread constraints for TEI pods | `[]` |
| `tei.extraEnv` | Additional TEI container environment variables | `[]` |
| `tei.replicas` | Number of TEI replicas | `1` |
### Pulsar Configuration
This version of the chart includes the dependent Pulsar chart in the charts/ directory.
- `pulsar-v3.3.0` is used for Pulsar v3
- `pulsar-v2.7.8` is used for Pulsar v2
Since milvus chart version 4.2.21, pulsar v3 is supported, but pulsar v2 will be still used by default until the release of Milvus v2.5.0.
We recommend creating new instances with pulsar v3 to avoid security vulnerabilities & some bugs in pulsar v2. To use pulsar v3, set `pulsarv3.enabled` to `true` and `pulsar.enabled` to `false`. Set other values for pulsar v3 under `pulsarv3` field.
You can find more information at:
* [https://pulsar.apache.org/charts](https://pulsar.apache.org/charts)
### Etcd Configuration
This version of the chart includes the dependent Etcd chart in the charts/ directory.
You can find more information at:
* [https://artifacthub.io/packages/helm/bitnami/etcd](https://artifacthub.io/packages/helm/bitnami/etcd)
### Minio Configuration
This version of the chart includes the dependent Minio chart in the charts/ directory.
You can find more information at:
* [https://github.com/minio/charts/blob/master/README.md](https://github.com/minio/charts/blob/master/README.md)
### Kafka Configuration
This version of the chart includes the dependent Kafka chart in the charts/ directory.
You can find more information at:
* [https://artifacthub.io/packages/helm/bitnami/kafka](https://artifacthub.io/packages/helm/bitnami/kafka)
### Node Selector, Affinity and Tolerations Configuration Guide
- [Node Selector Configuration Guide](../../docs/node-selector-configuration-guide.md)
- [Affinity Configuration Guide](../../docs/affinity-configuration-guide.md)
- [Tolerations Configuration Guide](../../docs/tolerations-configuration-guide.md)
#### Important Configuration Considerations
When configuring pod scheduling in Kubernetes, be aware of potential conflicts between different scheduling mechanisms:
1. **Node Selectors vs Node Affinity**
- Node selectors provide a simple way to constrain pods to nodes with specific labels
- Node affinity provides more flexible pod scheduling rules
- When used together, **both** node selector and node affinity rules must be satisfied
- Consider using only one mechanism unless you have specific requirements
2. **Scheduling Priority**
- Node Selectors: Hard requirement that must be satisfied
- Required Node Affinity: Hard requirement that must be satisfied
- Preferred Node Affinity: Soft preference that Kubernetes will try to satisfy
- Pod Anti-Affinity: Can prevent pods from being scheduled on the same node
3. **Best Practices**
- Start with simple node selectors for basic constraints
- Use node/pod affinity for more complex scheduling requirements
- Avoid combining multiple scheduling constraints unless necessary
- Test your configuration in a non-production environment first
For detailed examples and configurations, please refer to the documentation guides linked above.
### Milvus Live Profiling
Profiling is an effective way of understanding which parts of your application are consuming the most resources.
Continuous Profiling adds a dimension of time that allows you to understand your systems resource usage (i.e. CPU, Memory, etc.) over time and gives you the ability to locate, debug, and fix issues related to performance.
You can enable profiling with Pyroscope and you can find more information at:
* [https://pyroscope.io/docs/kubernetes-helm-chart/](https://pyroscope.io/docs/kubernetes-helm-chart/)
## Text Embeddings Inference (TEI) Integration Guide for Milvus
Text Embeddings Inference (TEI) is a high-performance text embedding model inference service that converts text into vector representations. Milvus is a vector database that can store and retrieve these vectors. By combining the two, you can build powerful semantic search and retrieval systems.
TEI is an open-source project developed by Hugging Face, available at [https://github.com/huggingface/text-embeddings-inference](https://github.com/huggingface/text-embeddings-inference).
This guide provides two ways to use TEI:
1. Deploy TEI service directly through the Milvus Helm Chart
2. Use external TEI service with Milvus integration
For detailed configuration and usage instructions, please refer to the [README-TEI.md](./README-TEI.md) document.