Percona Xtradb Cluster release Only for limited production use
pxc-release is a BOSH release of MySQL Galera that can be used as a backing store for Cloudfoundry. The Galera Cluster Provider is Percona Xtradb Cluster. This release is intended as a drop-in replacement for cf-mysql-release.
The pxc-mysql BOSH job runs mysql using Galera replication, across 1 or several nodes.
The typical clustered topology is 2 proxy nodes and 3 mysql nodes running the pxc-mysql BOSH job. The proxies can be separate vms or co-located with the pxc-mysql nodes.
You can also run this topology with a single mysql node running the pxc-mysql BOSH job and a single proxy job. In this case you would have a galera cluster of size 1, which does not provide high-availability.
The number of mysql nodes should always be odd, with a minimum count of three, to avoid split-brain. When a failed node comes back online, it will automatically rejoin the cluster and sync data from one of the healthy nodes.
Two proxy instances are recommended. The second proxy is intended to be used in a failover capacity. You can also choose to place a load balancer in front of both proxies, or use BOSH DNS to send traffic to both proxies.
In the event the first proxy fails, the second proxy will still be able to route requests to the mysql nodes.
The proxies both will route traffic to the lowest-indexed healthy galera node, according to the galera index (not bosh index).
Traffic to the MySQL cluster is routed through one or more proxy nodes. The current proxy implementation is Switchboard. This proxy acts as an intermediary between the client and the MySQL server, providing failover between MySQL nodes. The number of nodes is configured by the proxy job instance count in the deployment manifest.
NOTE: If the number of proxy nodes is set to zero, apps will be bound to the IP address of the first MySQL node in the cluster. If that IP address should change for any reason (e.g. loss of a VM) or a proxy was subsequently added, one would need to re-bind all apps to the IP address of the new node.
For more details see the proxy documentation.
Use the cf-deployment manifests with the use-pxc.yml ops file.
<a name='deploying-clustered>
To deploy a clustered deployment, use the pxc-deployment.yml manifest and apply the use-clustered opsfile:
bosh -d <deployment> deploy --ops-file operations/use-clustered.yml pxc-deployment.yml
To deploy a standalone deployment, use the pxc-deployment.yml manifest:
bosh -d <deployment> deploy pxc-deployment.ymlRequirements:
cf-mysql-release v36.12.0 or greater
Use the cf-deployment manifests with the migrate-cf-mysql-to-pxc.yml ops file. It is advisable to take a backup first.
⚠️ migrate-cf-mysql-to-pxc.ymlwill scale down a cluster to a single node. This is required for migration. Be sure to re-set to the appropriate number of instances when switching touse-pxc.ymlsubsequently.
The ops file will trigger the same migration procedure described in Using PXC release with other deployments
- If your cf-deployment uses CredHub, be sure to also include the secure-service-credentials-with-pxc-release.yml ops file.
After migrating, use the Deploying CF with pxc-release docs for your next deploy.
-
Make backups according to your usual backup procedure.
-
Get the latest pxc bosh release from bosh.io
-
Add the release to your manifest
-
⚠️ Scale down to 1 node and ensure the persistent disk has enough free space to double the size of the mysql data. -
Make the following changes to your bosh manifest:
-
Add the
pxc-mysqljob frompxc-releaseto the instance group that has themysqljob fromcf-mysql-release -
Configure the
pxc-mysqljob with the same credentials and property values as themysqljob -
To run the migration:
- Set the
cf_mysql_enabled: falseproperty on themysqljob - Set the
pxc_enabled: trueproperty onpxc-mysqljob - Switch the proxies to use the proxy job from
pxc-releaseinstead ofcf-mysql-release - Deploy using BOSH
- Set the
-
To prepare for the migration, but not run it immediately:
- Set the
cf_mysql_enabled: trueproperty on themysqljob - Set the
pxc_enabled: falseproperty onpxc-mysqljob - Deploy using BOSH
- The MySQL will run as normal with only the
cf-mysql-releaserunning - In order to trigger the migration, redeploy with
cf_mysql_enabled: falseandpxc_enabled: true
- Set the
-
⚠️ Do not enable both releases or disable both releases. Only enable one at a time.
-
-
The migration is triggered by deploying with
cf_mysql_enabled: falseandpxc_enabled: true. Thepre-startscript for thepxc-mysqljob inpxc-releasestarts both the Mariadb MySQL from thecf-mysql-releaseand the Percona MySQL frompxc-release. The migration dumps the MariaDB MySQL and loads that data into the Percona MySQL. This is done using pipes, so the dump is not written to disk, in order to reduce the use of disk space. The MariaDB MySQL is then stopped, leaving only the Percona MySQL running.⚠️ MySQL DB will experience downtime during the migration
-
After the migration, you can optionally clean up your deployment:
- The migration will make a copy of the MySQL data on the persistent disk. To reduce disk usage, you can delete the old copy of the data in
/var/vcap/store/mysqlafter you feel comfortable in the success of your migration. Do NOT delete the new copy of the data in/var/vcap/store/pxc-mysql. - Deploy only the
pxc-releaseand not thecf-mysql-releasein future deployments per Deploying new deployments, to free up disk space used by thecf-mysql-release.
- The migration will make a copy of the MySQL data on the persistent disk. To reduce disk usage, you can delete the old copy of the data in
-
Scale back up to the recommended 3 nodes, if desired.
- As of pxc 0.15.x, we implemented bpm support in the pxc-mysql job. bpm puts a hard time limit on monit stop operations and will eventually SIGKILL all processes in the bpm container if mysql takes too long to shut down. When pxc-release is deployed in a Galera topology, this will cause the node to reinitialize via an SST operation. During SST a node will remove its local data directory and replace it with data provided by another member of the cluster.
The Cloud Foundry team uses GitHub and accepts contributions via pull request.
Follow these steps to make a contribution to any of our open source repositories:
-
Ensure that you have completed our CLA Agreement for individuals or corporations.
-
Set your name and email (these should match the information on your submitted CLA)
git config --global user.name "Firstname Lastname" git config --global user.email "your_email@example.com"
- Fork the repository
- Check out
masterof pxc-release - Create a feature branch (
git checkout -b <my_new_branch>) - Make changes on your branch
- Deploy your changes using pxc as the database for cf-deployment to your dev environment and run CF Acceptance Tests (CATS)
- Push to your fork (
git push origin <my_new_branch>) and submit a pull request
We favor pull requests with very small, single commits with a single purpose.
Your pull request is much more likely to be accepted if it is small and focused with a clear message that conveys the intent of your change.