Datadog cli tool to sync resources across organizations.
See Installing section for guides on how to install and setup the tool.
Run the import
command to read the specified resources from the source organization and store them locally into JSON files in the directory resources/source
.
Then, you can run the sync
command which will use the stored files from previous import
command (unless --force-missing-dependencies
flag is passed) to create/modify the resources on the destination organization. The pushed resources are saved in the directory resources/destination
.
Note: The tool uses the resources
directory as the source of truth for determining what resources need to be created and modified. Hence, this directory should not be removed or corrupted.
Example Usage
# Import resources from parent organization and store them locally
$ datadog-sync import \
--source-api-key="..." \
--source-app-key="..." \
--source-api-url="https://api.datadoghq.com"
> 2024-03-14 14:53:54,280 - INFO - Starting import...
> ...
> 2024-03-14 15:00:46,100 - INFO - Finished import
# Check diff output to see what resources will be created/modified
$ datadog-sync diffs \
--destination-api-key="..." \
--destination-app-key="..." \
--destination-api-url="https://api.datadoghq.eu"
> 2024-03-14 15:46:22,014 - INFO - Starting diffs...
> ...
> 2024-03-14 14:51:15,379 - INFO - Finished diffs
# Sync the resources to the child organization from locally stored files and save the output locally
$ datadog-sync sync \
--destination-api-key="..." \
--destination-app-key="..." \
--destination-api-url="https://api.datadoghq.eu"
> 2024-03-14 14:55:56,535 - INFO - Starting sync...
> ...
> 2024-03-14 14:56:00,797 - INFO - Finished sync: 1 successes, 0 errors
The purpose of the datadog-sync-cli package is to provide an easy way to sync Datadog resources across Datadog organizations.
Note: this tool does not, nor is intended, for migrating intake data such as ingested logs, metrics, etc.
The source organization will not be modified, but the destination organization will have resources created and updated by the sync
command.
Note:: Instlling from source requires Python >= v3.9
- Clone the project repo and CD into the directory
git clone https://github.com/DataDog/datadog-sync-cli.git; cd datadog-sync-cli
- Install datadog-sync-cli tool using pip
pip install .
- Invoke the cli tool using
datadog-sync <command> <options>
- Download the executable from the Releases page
- Provide the executable with executable permission
chmod +x datadog-sync-cli-{system-name}-{machine-type}
- Move the executable to your bin directory
sudo mv datadog-sync-cli-{system-name}-{machine-type} /usr/local/bin/datadog-sync
- Invoke the CLI tool using
datadog-sync <command> <options>
- Download the executable with extension
.exe
from the Releases page - Add the directory containing the
exe
file to your path - Invoke the CLI tool in cmd/powershell using the file name and omitting the extension:
datadog-sync-cli-windows-amd64 <command> <options>
- Clone the project repo and CD into the directory
git clone https://github.com/DataDog/datadog-sync-cli.git; cd datadog-sync-cli
- Build the provided Dockerfile
docker build . -t datadog-sync
- Run the Docker image using entrypoint below:
docker run --rm -v <PATH_TO_WORKING_DIR>:/datadog-sync:rw \
-e DD_SOURCE_API_KEY=<DATADOG_API_KEY> \
-e DD_SOURCE_APP_KEY=<DATADOG_APP_KEY> \
-e DD_SOURCE_API_URL=<DATADOG_API_URL> \
-e DD_DESTINATION_API_KEY=<DATADOG_API_KEY> \
-e DD_DESTINATION_APP_KEY=<DATADOG_APP_KEY> \
-e DD_DESTINATION_API_URL=<DATADOG_API_URL> \
datadog-sync:latest <command> <options>
The docker run
command mounts a specified <PATH_TO_WORKING_DIR>
working directory to the container.
Available URL's for the source and destination API URLs are:
https://api.datadoghq.com
https://api.datadoghq.eu
https://api.us5.datadoghq.com
https://api.us3.datadoghq.com
https://api.ddog-gov.com
https://api.ap1.datadoghq.com
For all available regions, see Getting Started with Datadog Sites.
Filtering is done on two levels, at top resources level and per individual resource level using --resources
and --filter
respectively.
By default all resources are imported, synced, etc. If you would like to perform actions on a specific top level resource, or subset of resources, use --resources
option. For example, the command datadog-sync import --resources="dashboard_lists,dashboards"
will import ALL dashboards and dashboard lists in your Datadog organization.
Individual resources can be further filtered using the --filter
flag. For example, the following command datadog-sync import --resources="dashboards,dashboard_lists" --filter='Type=dashboard_lists;Name=name;Value=My custom list'
, will import ALL dashboards and ONLY dashboard lists with the name
attribute equal to My custom list
.
Filter option (--filter
) accepts a string made up of key=value
pairs separated by ;
.
--filter 'Type=<resource>;Name=<attribute_name>;Value=<attribute_value>;Operator=<operator>'
Available keys:
Type
: Resource such as Monitors, Dashboards, and more. [required]Name
: Attribute key to filter on. This can be any attribute represented in dot notation (such asattributes.user_count
). [required]Value
: Regex to filter attribute value by. Note: special regex characters need to be escaped if filtering by raw string. [required]Operator
: Available operators are below. All invalid operator's default toExactMatch
.Not
: Match not equal toValue
.SubString
(Deprecated): Sub string matching. (This operator will be removed in future releases. See SubString and ExactMatch Deprecation section.)ExactMatch
(Deprecated): Exact string match. (This operator will be removed in future releases. See SubString and ExactMatch Deprecation section.)
By default, if multiple filters are passed for the same resource, OR
logic is applied to the filters. This behavior can be adjusted using the --filter-operator
option.
In future releases the SubString
and ExactMatch
Operator will be removed. This is because the Value
key supports regex so both of these scenarios are covered by just writing the appropriate regex. Below is an example:
Let's take the scenario where you would like to filter for monitors that have the filter test
in the name
attribute:
Operator | Command |
---|---|
SubString |
--filter 'Type=monitors;Name=name;Value=filter test;Operator=SubString' |
Using Value |
--filter 'Type=monitors;Name=name;Value=.*filter test.* |
ExactMatch |
--filter 'Type=monitors;Name=name;Value=filter test;Operator=ExactMatch' |
Using Value |
--filter 'Type=monitors;Name=name;Value=^filter test$ |
A Custom config text file can be passed in place of options.
This is an example config file:
# config
destination_api_url="https://api.datadoghq.eu"
destination_api_key="<API_KEY>"
destination_app_key="<APP_KEY>"
source_api_key="<API_KEY>"
source_app_key="<APP_KEY>"
source_api_url="https://api.datadoghq.com"
filter=["Type=Dashboards;Name=title;Value=Test screenboard", "Type=Monitors;Name=tags;Value=sync:true"]
Then, run: datadog-sync import --config config
The tools sync
command provides a cleanup flag (--cleanup
). Passing the cleanup flag will delete resources from the destination organization which have been removed from the source organization. The resources to be deleted are determined based on the difference between the state files of source and destination organization.
For example, ResourceA
and ResourceB
are imported and synced, followed by deleting ResourceA
from the source organization. Running the import
command will update the source organizations state file to only include ResourceB
. The following sync --cleanup=Force
command will now delete ResourceA
from the destination organization.
A resources
directory is generated in the current working directory of the user. This directory contains json
mapping of resources between the source and destination organization. To avoid duplication and loss of mapping, this directory should be retained between tool usage.
When running againts multiple destination organizations, a seperate working directory should be used to ensure seperation of data.
Resource | Description |
---|---|
dashboard_lists | Sync Datadog dashboard lists. |
dashboards | Sync Datadog dashboards. |
downtime_schedules | Sync Datadog downtimes. |
downtimes (deprecated) | Sync Datadog downtimes. |
host_tags | Sync Datadog host tags. |
logs_custom_pipelines (deprecated) | Sync Datadog logs custom pipelines. |
logs_indexes | Sync Datadog logs indexes. |
logs_indexes_order | Sync Datadog logs indexes order. |
logs_metrics | Sync Datadog logs metrics. |
logs_pipelines | Sync Datadog logs OOTB integration and custom pipelines. |
logs_pipelines_order | Sync Datadog logs pipelines order. |
logs_restriction_queries | Sync Datadog logs restriction queries. |
metric_percentiles | Sync Datadog metric percentiles. |
metric_tag_configurations | Sync Datadog metric tags configurations. |
metrics_metadata | Sync Datadog metric metadata. |
monitors | Sync Datadog monitors. |
notebooks | Sync Datadog notebooks. |
powerpacks | Sync Datadog powerpacks. |
restriction_policies | Sync Datadog restriction policies. |
roles | Sync Datadog roles. |
service_level_objectives | Sync Datadog SLOs. |
slo_corrections | Sync Datadog SLO corrections. |
spans_metrics | Sync Datadog spans metrics. |
synthetics_global_variables | Sync Datadog synthetic global variables. |
synthetics_private_locations | Sync Datadog synthetic private locations. |
synthetics_tests | Sync Datadog synthetic tests. |
teams | Sync Datadog teams (excluding users and permissions). |
users | Sync Datadog users. |
Note: logs_custom_pipelines
resource has been deprecated in favor of logs_pipelines
resource which supports both logs OOTB integration and custom pipelines. To migrate to the new resource, rename the existing state files from logs_custom_pipelines.json
to logs_pipelines.json
for both source and destination files.
Many Datadog resources are interdependent. For example, some Datadog resource can reference roles
and dashboards
, which includes widgets that may use Monitors or Synthetics data. The datadog-sync tool syncs these resources in order to ensure dependencies are not broken.
If importing/syncing subset of resources, users should ensure that dependent resources are imported and synced as well.
See Supported resources section below for potential resource dependencies.
Resource | Dependencies |
---|---|
dashboard_lists | dashboards |
dashboards | monitors, roles, powerpacks, service_level_objectives |
downtime_schedules | monitors |
downtimes (deprecated) | monitors |
host_tags | - |
logs_custom_pipelines (deprecated) | - |
logs_indexes | - |
logs_indexes_order | logs_indexes |
logs_metrics | - |
logs_pipelines | - |
logs_pipelines_order | logs_pipelines |
logs_restriction_queries | roles |
metric_percentiles | - |
metric_tag_configurations | - |
metrics_metadata | - |
monitors | roles, service_level_objectives |
notebooks | - |
powerpacks | monitors, service_level_objectives |
restriction_policies | dashboards, service_level_objectives, notebooks, users, roles |
roles | - |
service_level_objectives | monitors, synthetics_tests |
slo_corrections | service_level_objectives |
spans_metrics | - |
synthetics_global_variables | synthetics_tests |
synthetics_private_locations | - |
synthetics_tests | synthetics_private_locations, synthetics_global_variables, roles |
teams | - |
users | roles |