ExternalDNS synchronizes exposed Kubernetes Services and Ingresses with DNS providers.
Inspired by Kubernetes DNS, Kubernetes' cluster-internal DNS server, ExternalDNS makes Kubernetes resources discoverable via public DNS servers. Like KubeDNS, it retrieves a list of resources (Services, Ingresses, etc.) from the Kubernetes API to determine a desired list of DNS records. Unlike KubeDNS, however, it's not a DNS server itself, but merely configures other DNS providers accordingly—e.g. AWS Route 53 or Google Cloud DNS.
In a broader sense, ExternalDNS allows you to control DNS records dynamically via Kubernetes resources in a DNS provider-agnostic way.
The FAQ contains additional information and addresses several questions about key concepts of ExternalDNS.
To see ExternalDNS in action, have a look at this video or read this blogpost.
ExternalDNS' current release is v0.5
. This version allows you to keep selected zones (via --domain-filter
) synchronized with Ingresses and Services of type=LoadBalancer
in various cloud providers:
- Google Cloud DNS
- AWS Route 53
- AWS Service Discovery
- AzureDNS
- CloudFlare
- RcodeZero
- DigitalOcean
- DNSimple
- Infoblox
- Dyn
- OpenStack Designate
- PowerDNS
- CoreDNS
- Exoscale
- Oracle Cloud Infrastructure DNS
- Linode DNS
- RFC2136
- NS1
- TransIP
- VinylDNS
From this release, ExternalDNS can become aware of the records it is managing (enabled via --registry=txt
), therefore ExternalDNS can safely manage non-empty hosted zones. We strongly encourage you to use v0.5
(or greater) with --registry=txt
enabled and --txt-owner-id
set to a unique value that doesn't change for the lifetime of your cluster. You might also want to run ExternalDNS in a dry run mode (--dry-run
flag) to see the changes to be submitted to your DNS Provider API.
Note that all flags can be replaced with environment variables; for instance,
--dry-run
could be replaced with EXTERNAL_DNS_DRY_RUN=1
, or
--registry txt
could be replaced with EXTERNAL_DNS_REGISTRY=txt
.
ExternalDNS supports multiple DNS providers which have been implemented by the ExternalDNS contributors. Maintaining all of those in a central repository is a challenge and we have limited resources to test changes. This means that it is very hard to test all providers for possible regressions and, as written in the Contributing section, we encourage contributors to step in as maintainers for the individual providers and help by testing the integrations.
End-to-end testing of ExternalDNS is currently performed in the separate kubernetes-on-aws repository.
We define the following stability levels for providers:
- Stable: Used for smoke tests before a release, used in production and maintainers are active.
- Beta: Community supported, well tested, but maintainers have no access to resources to execute integration tests on the real platform and/or are not using it in production.
- Alpha: Community provided with no support from the maintainers apart from reviewing PRs.
The following table clarifies the current status of the providers according to the aforementioned stability levels:
Provider | Status |
---|---|
Google Cloud DNS | Stable |
AWS Route 53 | Stable |
AWS Service Discovery | Beta |
AzureDNS | Beta |
CloudFlare | Beta |
RcodeZero | Alpha |
DigitalOcean | Alpha |
DNSimple | Alpha |
Infoblox | Alpha |
Dyn | Alpha |
OpenStack Designate | Alpha |
PowerDNS | Alpha |
CoreDNS | Alpha |
Exoscale | Alpha |
Oracle Cloud Infrastructure DNS | Alpha |
Linode DNS | Alpha |
RFC2136 | Alpha |
NS1 | Alpha |
TransIP | Alpha |
VinylDNS | Alpha |
RancherDNS | Alpha |
The are two ways of running ExternalDNS:
- Deploying to a Cluster
- Running Locally
The following tutorials are provided:
- Alibaba Cloud
- AWS
- Azure
- Cloudflare
- CoreDNS
- DigitalOcean
- DNSimple
- Dyn
- Exoscale
- ExternalName Services
- Google Container Engine
- Headless Services
- Infoblox
- Istio Gateway Source
- Kubernetes Security Context
- Linode
- Nginx Ingress Controller
- NS1
- OpenStack Designate
- Oracle Cloud Infrastructure (OCI) DNS
- PowerDNS
- RcodeZero
- RancherDNS (RDNS)
- RFC2136
- TransIP
- VinylDNS
Make sure you have the following prerequisites:
- A local Go 1.7+ development environment.
- Access to a Google/AWS account with the DNS API enabled.
- Access to a Kubernetes cluster that supports exposing Services, e.g. GKE.
First, get ExternalDNS:
$ git clone https://github.com/kubernetes-incubator/external-dns.git && cd external-dns
This project uses Go modules as introduced in Go 1.11 therefore you need Go >=1.11 installed in order to build. If using Go 1.11 you also need to activate Module support.
Assuming Go has been setup with module support it can be built simply by running:
$ make
This will create external-dns in the build directory directly from master.
Next, run an application and expose it via a Kubernetes Service:
$ kubectl run nginx --image=nginx --replicas=1 --port=80
$ kubectl expose deployment nginx --port=80 --target-port=80 --type=LoadBalancer
Annotate the Service with your desired external DNS name. Make sure to change example.org
to your domain.
$ kubectl annotate service nginx "external-dns.alpha.kubernetes.io/hostname=nginx.example.org."
Optionally, you can customize the TTL value of the resulting DNS record by using the external-dns.alpha.kubernetes.io/ttl
annotation:
$ kubectl annotate service nginx "external-dns.alpha.kubernetes.io/ttl=10"
For more details on configuring TTL, see here.
Locally run a single sync loop of ExternalDNS.
$ external-dns --registry txt --txt-owner-id my-cluster-id --provider google --google-project example-project --source service --once --dry-run
This should output the DNS records it will modify to match the managed zone with the DNS records you desire. It also assumes you are running in the default
namespace. See the FAQ for more information regarding namespaces.
Note: TXT records will have my-cluster-id
value embedded. Those are used to ensure that ExternalDNS is aware of the records it manages.
Once you're satisfied with the result, you can run ExternalDNS like you would run it in your cluster: as a control loop, and not in dry-run mode:
$ external-dns --registry txt --txt-owner-id my-cluster-id --provider google --google-project example-project --source service
Check that ExternalDNS has created the desired DNS record for your Service and that it points to its load balancer's IP. Then try to resolve it:
$ dig +short nginx.example.org.
104.155.60.49
Now you can experiment and watch how ExternalDNS makes sure that your DNS records are configured as desired. Here are a couple of things you can try out:
- Change the desired hostname by modifying the Service's annotation.
- Recreate the Service and see that the DNS record will be updated to point to the new load balancer IP.
- Add another Service to create more DNS records.
- Remove Services to clean up your managed zone.
The tutorials section contains examples, including Ingress resources, and shows you how to set up ExternalDNS in different environments such as other cloud providers and alternative Ingress controllers.
If using a txt registry and attempting to use a CNAME the --txt-prefix
must be set to avoid conflicts. Changing --txt-prefix
will result in lost ownership over previously created records.
ExternalDNS was built with extensibility in mind. Adding and experimenting with new DNS providers and sources of desired DNS records should be as easy as possible. It should also be possible to modify how ExternalDNS behaves—e.g. whether it should add records but never delete them.
Here's a rough outline on what is to come (subject to change):
- Support for Google CloudDNS
- Support for Kubernetes Services
- Support for AWS Route 53
- Support for Kubernetes Ingresses
- Support for AWS Route 53 via ALIAS
- Support for multiple zones
- Ownership System
- Support for AzureDNS
- Support for CloudFlare
- Support for DigitalOcean
- Multiple DNS names per Service
- Support for creating DNS records to multiple targets (for Google and AWS)
- Support for OpenStack Designate
- Support for PowerDNS
- Support for Linode
- Support for RcodeZero
- Support for NS1
- Support for TransIP
- Ability to replace Kops' DNS Controller (This could also directly become
v1.0
)
- Ability to replace Kops' DNS Controller
- Ability to replace Zalando's Mate
- Ability to replace Molecule Software's route53-kubernetes
- Support for CoreDNS
- Support for record weights
- Support for different behavioral policies
- Support for Services with
type=NodePort
- Support for CRDs
- Support for more advanced DNS record configurations
Have a look at the milestones to get an idea of where we currently stand.
We encourage you to get involved with ExternalDNS, as users, contributors or as new maintainers that can take over some parts like different providers and help with code reviews.
Providers which currently need maintainers:
- Azure
- Cloudflare
- Digital Ocean
- Google Cloud Platform
Any provider should have at least one maintainer. It would be nice if you run it in production, but it is not required. You should check changes and make sure your provider is working correctly.
It would be also great to have an automated end-to-end test for different cloud providers, so help from Kubernetes maintainers and their idea on how this can be done would be valuable.
Read the contributing guidelines and have a look at the contributing docs to learn about building the project, the project structure, and the purpose of each package.
If you are interested please reach out to us on the Kubernetes slack in the #external-dns channel.
For an overview on how to write new Sources and Providers check out Sources and Providers.
ExternalDNS is an effort to unify the following similar projects in order to bring the Kubernetes community an easy and predictable way of managing DNS records across cloud providers based on their Kubernetes resources:
- Kops' DNS Controller
- Zalando's Mate
- Molecule Software's route53-kubernetes
This is a Kubernetes Incubator project. The project was established 2017-Feb-9 (initial announcement here). The incubator team for the project is:
- Sponsor: sig-network
- Champion: Tim Hockin (@thockin)
- SIG: sig-network
For more information about sig-network, such as meeting times and agenda, check out the community site.
Participation in the Kubernetes community is governed by the Kubernetes Code of Conduct.