MinIO is a High Performance Object Storage released under Apache License v2.0. It is API compatible with Amazon S3 cloud storage service. Use MinIO to build high performance infrastructure for machine learning, analytics and application data workloads.
MinIO supports distributed mode. In distributed mode, you can pool multiple drives (even on different machines) into a single object storage server.
For more detailed documentation please visit here
This chart bootstraps MinIO deployment on a Kubernetes cluster using the Helm package manager.
- Kubernetes 1.4+ with Beta APIs enabled for default standalone mode.
- Kubernetes 1.5+ with Beta APIs enabled to run MinIO in distributed mode.
- PV provisioner support in the underlying infrastructure.
$ helm repo add minio https://helm.min.io/
Install this chart using:
$ helm install --namespace minio --generate-name minio/minio
The command deploys MinIO on the Kubernetes cluster in the default configuration. The configuration section lists the parameters that can be configured during installation.
An instance of a chart running in a Kubernetes cluster is called a release. Each release is identified by a unique name within the cluster. Helm automatically assigns a unique release name after installing the chart. You can also set your preferred name by:
$ helm install my-release minio/minio
By default a pre-generated access and secret key will be used. To override the default keys, pass the access and secret keys as arguments to helm install.
$ helm install --set accessKey=myaccesskey,secretKey=mysecretkey --generate-name minio/minio
ConfigMap allows injecting containers with configuration data even while a Helm release is deployed.
To update your MinIO server configuration while it is deployed in a release, you need to
- Check all the configurable values in the MinIO chart using
helm inspect values minio/minio
. - Override the
minio_server_config
settings in a YAML formatted file, and then pass that file like thishelm upgrade -f config.yaml minio/minio
. - Restart the MinIO server(s) for the changes to take effect.
You can also check the history of upgrades to a release using helm history my-release
. Replace my-release
with the actual release name.
Assuming your release is named as my-release
, delete it using the command:
$ helm delete my-release
or
$ helm uninstall my-release
The command removes all the Kubernetes components associated with the chart and deletes the release.
You can use Helm to update MinIO version in a live release. Assuming your release is named as my-release
, get the values using the command:
$ helm get values my-release > old_values.yaml
Then change the field image.tag
in old_values.yaml
file with MinIO image tag you want to use. Now update the chart using
$ helm upgrade -f old_values.yaml my-release minio/minio
Default upgrade strategies are specified in the values.yaml
file. Update these fields if you'd like to use a different strategy.
The following table lists the configurable parameters of the MinIO chart and their default values.
Parameter | Description | Default |
---|---|---|
nameOverride |
Provide a name in place of minio |
"" |
fullnameOverride |
Provide a name to substitute for the full names of resources | "" |
image.repository |
Image repository | minio/minio |
image.tag |
MinIO image tag. Possible values listed here. | RELEASE.2020-09-26T03-44-56Z |
image.pullPolicy |
Image pull policy | IfNotPresent |
imagePullSecrets |
List of container registry secrets | [] |
mcImage.repository |
Client image repository | minio/mc |
mcImage.tag |
mc image tag. Possible values listed here. | RELEASE.2020-08-08T02-33-58Z |
mcImage.pullPolicy |
mc Image pull policy | IfNotPresent |
ingress.enabled |
Enables Ingress | false |
ingress.labels |
Ingress labels | {} |
ingress.annotations |
Ingress annotations | {} |
ingress.hosts |
Ingress accepted hostnames | [] |
ingress.tls |
Ingress TLS configuration | [] |
mode |
MinIO server mode (standalone or distributed ) |
standalone |
extraArgs |
Additional command line arguments to pass to the MinIO server | [] |
replicas |
Number of nodes (applicable only for MinIO distributed mode). | 4 |
zones |
Number of zones (applicable only for MinIO distributed mode). | 1 |
drivesPerNode |
Number of drives per node (applicable only for MinIO distributed mode). | 1 |
existingSecret |
Name of existing secret with access and secret key. | "" |
accessKey |
Default access key (5 to 20 characters) | YOURACCESSKEY |
secretKey |
Default secret key (8 to 40 characters) | YOURSECRETKEY |
certsPath |
Default certs path location | /etc/minio/certs |
configPathmc |
Default config file location for MinIO client - mc | /etc/minio/mc |
mountPath |
Default mount location for persistent drive | /export |
bucketRoot |
Directory from where minio should serve buckets. | Value of .mountPath |
clusterDomain |
domain name of kubernetes cluster where pod is running. | cluster.local |
service.type |
Kubernetes service type | ClusterIP |
service.port |
Kubernetes port where service is exposed | 9000 |
service.externalIPs |
service external IP addresses | nil |
service.annotations |
Service annotations | {} |
serviceAccount.create |
Toggle creation of new service account | true |
serviceAccount.name |
Name of service account to create and/or use | "" |
persistence.enabled |
Use persistent volume to store data | true |
persistence.size |
Size of persistent volume claim | 500Gi |
persistence.existingClaim |
Use an existing PVC to persist data | nil |
persistence.storageClass |
Storage class name of PVC | nil |
persistence.accessMode |
ReadWriteOnce or ReadOnly | ReadWriteOnce |
persistence.subPath |
Mount a sub directory of the persistent volume if set | "" |
resources |
Memory resource requests | Memory: 4Gi |
priorityClassName |
Pod priority settings | "" |
securityContext.enabled |
Enable to run containers as non-root. NOTE: if persistence.enabled=false then securityContext will be automatically disabled |
true |
securityContext.runAsUser |
User id of the user for the container | 1000 |
securityContext.runAsGroup |
Group id of the user for the container | 1000 |
securityContext.fsGroup |
Group id of the persistent volume mount for the container | 1000 |
nodeSelector |
Node labels for pod assignment | {} |
affinity |
Affinity settings for pod assignment | {} |
tolerations |
Toleration labels for pod assignment | [] |
podAnnotations |
Pod annotations | {} |
podLabels |
Pod Labels | {} |
tls.enabled |
Enable TLS for MinIO server | false |
tls.certSecret |
Kubernetes Secret with public.crt and private.key files. |
"" |
livenessProbe.initialDelaySeconds |
Delay before liveness probe is initiated | 120 |
livenessProbe.periodSeconds |
How often to perform the probe | 15 |
livenessProbe.timeoutSeconds |
When the probe times out | 10 |
livenessProbe.successThreshold |
Minimum consecutive successes for the probe to be considered successful after having failed. | 1 |
livenessProbe.failureThreshold |
Minimum consecutive failures for the probe to be considered failed after having succeeded. | 3 |
defaultBucket.enabled |
If set to true, a bucket will be created after MinIO install | false |
defaultBucket.name |
Bucket name | bucket |
defaultBucket.policy |
Bucket policy | none |
defaultBucket.purge |
Purge the bucket if already exists | false |
buckets |
List of buckets to create after MinIO install | [] |
makeBucketJob.annotations |
Additional annotations for the Kubernetes Batch (make-bucket-job) | "" |
makeBucketJob.podAnnotations |
Additional annotations for the pods of the Kubernetes Batch (make-bucket-job) | "" |
makeBucketJob.securityContext.enabled |
Enable to run Kubernetes Batch (make-bucket-job) containers as non-root. | false |
makeBucketJob.securityContext.runAsUser |
User id of the user for the container | 1000 |
makeBucketJob.securityContext.runAsGroup |
Group id of the user for the container | 1000 |
makeBucketJob.securityContext.fsGroup |
Group id of the persistent volume mount for the container | 1000 |
makeBucketJob.resources.requests.memory |
Memory resource requests for 'make bucket' job | 128Mi |
updatePrometheusJob.podAnnotations |
Additional annotations for the pods of the Kubernetes Batch (update-prometheus-secret) | "" |
updatePrometheusJob.securityContext.enabled |
Enable to run Kubernetes Batch (update-prometheus-secret) containers as non-root. | false |
updatePrometheusJob.securityContext.runAsUser |
User id of the user for the container | 1000 |
updatePrometheusJob.securityContext.runAsGroup |
Group id of the user for the container | 1000 |
updatePrometheusJob.securityContext.fsGroup |
Group id of the persistent volume mount for the container | 1000 |
s3gateway.enabled |
Use MinIO as a s3 gateway | false |
s3gateway.replicas |
Number of s3 gateway instances to run in parallel | 4 |
s3gateway.serviceEndpoint |
Endpoint to the S3 compatible service | "" |
s3gateway.accessKey |
Access key of S3 compatible service | "" |
s3gateway.secretKey |
Secret key of S3 compatible service | "" |
azuregateway.enabled |
Use MinIO as an azure gateway | false |
azuregateway.replicas |
Number of azure gateway instances to run in parallel | 4 |
gcsgateway.enabled |
Use MinIO as a Google Cloud Storage gateway | false |
gcsgateway.gcsKeyJson |
credential json file of service account key | "" |
gcsgateway.projectId |
Google cloud project id | "" |
ossgateway.enabled |
Use MinIO as an Alibaba Cloud Object Storage Service gateway | false |
ossgateway.replicas |
Number of oss gateway instances to run in parallel | 4 |
ossgateway.endpointURL |
OSS server endpoint. | "" |
nasgateway.enabled |
Use MinIO as a NAS gateway | false |
nasgateway.replicas |
Number of NAS gateway instances to be run in parallel on a PV | 4 |
b2gateway.enabled |
Use MinIO as a Backblaze B2 gateway | false |
b2gateway.replicas |
Number of b2 gateway instances to run in parallel | 4 |
environment |
Set MinIO server relevant environment variables in values.yaml file. MinIO containers will be passed these variables when they start. |
MINIO_API_READY_DEADLINE: "5s" |
metrics.serviceMonitor.enabled |
Set this to true to create ServiceMonitor for Prometheus operator |
false |
metrics.serviceMonitor.additionalLabels |
Additional labels that can be used so ServiceMonitor will be discovered by Prometheus | {} |
metrics.serviceMonitor.namespace |
Optional namespace in which to create ServiceMonitor | nil |
metrics.serviceMonitor.interval |
Scrape interval. If not set, the Prometheus default scrape interval is used | nil |
metrics.serviceMonitor.scrapeTimeout |
Scrape timeout. If not set, the Prometheus default scrape timeout is used | nil |
etcd.endpoints |
Enpoints of etcd | [] |
etcd.pathPrefix |
Prefix for all etcd keys | "" |
etcd.corednsPathPrefix |
Prefix for all CoreDNS etcd keys | "" |
etcd.clientCert |
Certificate used for SSL/TLS connections to etcd (etcd Security) | "" |
etcd.clientCertKey |
Key for the certificate (etcd Security) | "" |
Some of the parameters above map to the env variables defined in the MinIO DockerHub image.
You can specify each parameter using the --set key=value[,key=value]
argument to helm install
. For example,
$ helm install --name my-release --set persistence.size=1Ti minio/minio
The above command deploys MinIO server with a 100Gi backing persistent volume.
Alternately, you can provide a YAML file that specifies parameter values while installing the chart. For example,
$ helm install --name my-release -f values.yaml minio/minio
Tip: You can use the default values.yaml
This chart provisions a MinIO server in standalone mode, by default. To provision MinIO server in distributed mode, set the mode
field to distributed
,
$ helm install --set mode=distributed minio/minio
This provisions MinIO server in distributed mode with 4 nodes. To change the number of nodes in your distributed MinIO server, set the replicas
field,
$ helm install --set mode=distributed,replicas=8 minio/minio
This provisions MinIO server in distributed mode with 8 nodes. Note that the replicas
value should be a minimum value of 4, there is no limit on number of servers you can run.
You can also expand an existing deployment by adding new zones, following command will create a total of 16 nodes with each zone running 8 nodes.
$ helm install --set mode=distributed,replicas=8,zones=2 minio/minio
StatefulSet limitations applicable to distributed MinIO
- StatefulSets need persistent storage, so the
persistence.enabled
flag is ignored whenmode
is set todistributed
. - When uninstalling a distributed MinIO release, you'll need to manually delete volumes associated with the StatefulSet.
MinIO in NAS gateway mode can be used to create multiple MinIO instances backed by single PV in ReadWriteMany
mode. Currently few Kubernetes volume plugins support ReadWriteMany
mode. To deploy MinIO NAS gateway with Helm chart you'll need to have a Persistent Volume running with one of the supported volume plugins. This document
outlines steps to create a NFS PV in Kubernetes cluster.
To provision MinIO servers in NAS gateway mode, set the nasgateway.enabled
field to true
,
$ helm install --set nasgateway.enabled=true minio/minio
This provisions 4 MinIO NAS gateway instances backed by single storage. To change the number of instances in your MinIO deployment, set the replicas
field,
$ helm install --set nasgateway.enabled=true,nasgateway.replicas=8 minio/minio
This provisions MinIO NAS gateway with 8 instances.
This chart provisions a PersistentVolumeClaim and mounts corresponding persistent volume to default location /export
. You'll need physical storage available in the Kubernetes cluster for this to work. If you'd rather use emptyDir
, disable PersistentVolumeClaim by:
$ helm install --set persistence.enabled=false minio/minio
"An emptyDir volume is first created when a Pod is assigned to a Node, and exists as long as that Pod is running on that node. When a Pod is removed from a node for any reason, the data in the emptyDir is deleted forever."
If a Persistent Volume Claim already exists, specify it during installation.
- Create the PersistentVolume
- Create the PersistentVolumeClaim
- Install the chart
$ helm install --set persistence.existingClaim=PVC_NAME minio/minio
To enable network policy for MinIO,
install a networking plugin that implements the Kubernetes
NetworkPolicy spec,
and set networkPolicy.enabled
to true
.
For Kubernetes v1.5 & v1.6, you must also turn on NetworkPolicy by setting the DefaultDeny namespace annotation. Note: this will enforce policy for all pods in the namespace:
kubectl annotate namespace default "net.beta.kubernetes.io/network-policy={\"ingress\":{\"isolation\":\"DefaultDeny\"}}"
With NetworkPolicy enabled, traffic will be limited to just port 9000.
For more precise policy, set networkPolicy.allowExternal=true
. This will
only allow pods with the generated client label to connect to MinIO.
This label will be displayed in the output of a successful install.
Instead of having this chart create the secret for you, you can supply a preexisting secret, much like an existing PersistentVolumeClaim.
First, create the secret:
$ kubectl create secret generic my-minio-secret --from-literal=accesskey=foobarbaz --from-literal=secretkey=foobarbazqux
Then install the chart, specifying that you want to use an existing secret:
$ helm install --set existingSecret=my-minio-secret minio/minio
The following fields are expected in the secret
accesskey
- the access key IDsecretkey
- the secret keygcs_key.json
- The GCS key if you are using the GCS gateway feature. This is optional.
To enable TLS for MinIO containers, acquire TLS certificates from a CA or create self-signed certificates. While creating / acquiring certificates ensure the corresponding domain names are set as per the standard DNS naming conventions in a Kubernetes StatefulSet (for a distributed MinIO setup). Then create a secret using
$ kubectl create secret generic tls-ssl-minio --from-file=path/to/private.key --from-file=path/to/public.crt
Then install the chart, specifying that you want to use the TLS secret:
$ helm install --set tls.enabled=true,tls.certSecret=tls-ssl-minio minio/minio
To pass environment variables to MinIO containers when deploying via Helm chart, use the below command line format
$ helm install --set environment.MINIO_BROWSER=on,environment.MINIO_DOMAIN=domain-name minio/minio
You can add as many environment variables as required, using the above format. Just add environment.<VARIABLE_NAME>=<value>
under set
flag.
Install the chart, specifying the buckets you want to create after install:
$ helm install --set buckets[0].name=bucket1,buckets[0].policy=none,buckets[0].purge=false minio/minio
Description of the configuration parameters used above -
buckets[].name
- name of the bucket to create, must be a string with length > 0buckets[].policy
- can be one of none|download|upload|publicbuckets[].purge
- purge if bucket exists already