Containerized, auto-scalable private Synthetic locations on Kubernetes

  • Latest Dynatrace
  • How-to guide
  • Published Nov 14, 2025

Dynatrace version 1.264+

Containerized, auto-scalable private Synthetic locations on Kubernetes and its commercial distribution OpenShift are an alternative to deploying Synthetic-enabled ActiveGates on separate hosts or virtual machines and then assigning them to private locations for the execution of synthetic monitors.

Unlike individual Synthetic-enabled ActiveGates that are deployed and assigned to private locations (and then tracked via utilization metrics) containerized locations are deployed as a whole, with a minimum and maximum number of ActiveGates as the necessary input parameters.

Kubernetes and OpenShift aren't just additional supported ActiveGate platforms along with Windows and Linux; with this offering, containerized private Synthetic locations:

  • Are auto-scalable (based on utilization metrics and the maximum/minimum number of ActiveGates specified).
  • Are easy to manage and maintain.
  • Support the synthetic monitoring of cloud-native solutions that require container-based application development.
  • Can be deployed faster while minimizing downtime.
  • Are automatically tracked for resource utilization as a part of auto-scaling operations.

You can manage Kubernetes/OpenShift locations via the Dynatrace web UI and the existing Synthetic - Locations, nodes, and configuration API v2. Additional Early Access endpoints in this API facilitate the deployment of Kubernetes locations; the new endpoints help you generate the commands that need to be executed on the Kubernetes cluster.

You can execute scheduled as well as on-demand executions of all types of synthetic monitors on containerized locations.

Architecture

Containerized private Synthetic locations are deployed as a whole.

  • Each location has multiple Synthetic-enabled ActiveGates configured as pods. You specify a minimum and maximum number of ActiveGates when setting up a location.

  • The StatefulSet is considered the location.

    Locations are scaled automatically by adjusting the number of ActiveGates per location by the following additional parts of the containerized location architecture.

  • The Synthetic metric adapter requests and receives utilization metrics for the containerized ActiveGates from the Dynatrace Cluster.

  • The horizontal pod auto-scaler scales a location by adjusting the number of ActiveGates based on the utilization data it receives from the Synthetic metric adapter.

    There is one horizontal pod auto-scaler per location.

Requirements

Containerized private Synthetic locations are supported with Dynatrace version 1.264+ on Kubernetes 1.22-1.25 with persistent volume and kubectl support.

  • Additional support for Kubernetes 1.26+ is available in the installation workflow.
  • All kinds of Kubernetes implementations are supported, whether cloud or local (for example, Amazon EKS or Minikube).
  • OpenShift versions compatible with the supported Kubernetes versions are supported.

Internet connectivity is required to access the public repositories where Docker images for the Synthetic-enabled ActiveGate and Synthetic metric adapter are available. These image locations are referenced in the respective template files—see Install a containerized location and Update a containerized location below.

Sizing guide

The ActiveGate hardware requirements below are listed by size.

  • CPU and RAM requests refer to the resources reserved by pods upon creation.
  • CPU and RAM limits refer to the maximum resource consumption per pod.
  • If the location is monitored by OneAgent or another deep monitoring solution, memory (RAM) requirements will increase.
  • Browserless pod in FIPS mode has the same requirements as ordinary browserless pod.
Browser-supporting podBrowserless podBrowser-supporting pod in FIPS modeBrowser-supporting pod in FIPS mode with corporate proxyActiveGateSynthetic EngineBrowser workerFIPS proxyFIPS peer
Containers425611211
CPU requests1.4 vCPU0.4 vCPU1.9 vCPU2.05 vCPU0.3 vCPU0.25 vCPU2 × 0.5 vCPU0.5 vCPU0.15 vCPU
CPU limits3.8 vCPU0.8 vCPU5.3 vCPU5.6 vCPU0.3 vCPU0.5 vCPU2 × 1.5 vCPU1.5 vCPU0.3 vCPU
RAM requests3.25 GiB1.25 GiB3.5 GiB3.75 GiB0.25 GiB1 GiB2 × 1 GiB0.25 GiB0.25 GiB
RAM limits7 GiB3 GiB7.5 GiB8 GiB1 GiB2 GiB2 × 2 GiB0.5 GiB0.5 GiB
Ephemeral storage1.5 GiB1.3 GiB1.6 GiB1.7 GiB1.2 GiB0.1 GiB2 × 0.1 GiB0.1 GiB0.1 GiB
Persistent storage3 GiB3 GiB3 GiB3 GiB
RAM disk1 GiB-1 GiB1 GiB

Best practices and caveats

ActiveGates

  • We recommend the S ActiveGate size and a minimum of two ActiveGates per location.
  • When considering node size, keep in mind the possible limitations specific to the Kubernetes service you will be relying on.
  • All ActiveGates within a location are always the same size.
  • Once specified, ActiveGate size for a location can't be changed because persistent storage can't be resized.
  • Kubernetes locations follow the same rules as other locations in that an ActiveGate can't be added to multiple locations simultaneously.
  • You cannot combine containerized and non-containerized ActiveGates in the same location.
  • The image for Synthetic-enabled ActiveGate is in a public registry; this image location is referenced by the template file.

Locations

  • We recommend installing each location in its own namespace.

  • If deploying more than one location per namespace, use different names for the respective ActiveGate resources—see Install a containerized location below.

  • Locations that share a single Kubernetes namespace must be connected to the same Dynatrace environment as the Synthetic metric adapter in order to be auto-scalable. For example, assume that Location A and the metric adapter are configured for Environment X. However, Location A shares a namespace with Location B, which is configured for Environment Y. In such a case, Location A is auto-scalable; Location B is not auto-scalable.

  • If you want to install a location in the same namespace as other Dynatrace resources such as Dynatrace Operator, be aware of the more demanding hardware and system requirements for containerized Synthetic-enabled ActiveGates.

Synthetic metric adapter

  • The best practice is to deploy the Synthetic metric adapter in its own namespace per Kubernetes cluster. The Synthetic metric adapter can share a namespace with a location. However, deploying the metric adapter in its own namespace ensures that it isn't deleted when a location is taken down.
  • The metric adapter can only communicate with a single Dynatrace environment, so location auto-scaling works just for that environment.
  • The image for the Synthetic metric adapter is in a public registry; this image location is referenced by the template file.

Auto-scaling specifics

For auto-scaling purposes, the Synthetic metric adapter needs access to and extends the Kubernetes API by specifying a new API service—v1beta1.external.metrics.k8s.io.

This API service is defined in the Synthetic metric adapter template—see Install and deploy a containerized location below.

API service definition in the metric adapter template
apiVersion: apiregistration.k8s.io/v1
kind: APIService
metadata:
name: v1beta1.external.metrics.k8s.io
spec:
service:
name: dynatrace-metrics-apiserver
namespace: {{adapterNamespace}}
group: external.metrics.k8s.io
version: v1beta1
insecureSkipTLSVerify: true
groupPriorityMinimum: 100
versionPriority: 100

The Synthetic metric adapter also modifies an existing resource in its template—the horizontal-pod-autoscaler ServiceAccount in kube-system namespace.

Existing resource modification in the metric adapter template
apiVersion: rbac.authorization.k8s.io/v1
kind: ClusterRoleBinding
metadata:
name: hpa-controller-dynatrace-metrics
roleRef:
apiGroup: rbac.authorization.k8s.io
kind: ClusterRole
name: dynatrace-metrics-server-resources
subjects:
- kind: ServiceAccount
name: horizontal-pod-autoscaler
namespace: kube-system

Limitations

Note that only one external metric server is allowed in the Kubernetes cluster. Because of this, other components that serve as a metric server (such as the KEDA add-on or the Prometheus adapter) cannot be used along with the Synthetic metric adapter.

Install a containerized location

1. Initial setup for a Kubernetes/OpenShift location

  1. Go to Synthetic Classic Synthetic > Private locations > New private locations > Kubernetes or OpenShift.

  2. Provide a Location name of your choice.

  3. Select a Geographic location, for example, San Francisco, California, United States. (Note that you cannot Save changes until you've specified a name and location.)

  4. In the ActiveGates section:

    1. Specify a Minimum and Maximum number of ActiveGates for your location. These settings are the auto-scaling parameters that the horizontal pod auto-scaler uses.
    2. Select an ActiveGate Node size (XS, S, or M). See also Requirements and Best practices and caveats.
  5. If needed, turn on the FIPS mode. You can choose between just FIPS mode and FIPS mode with corporate proxy. Alternatively, you can turn on the FIPS mode via API. For details, see Install FIPS-enabled containerized location via API.

  6. Kubernetes only If your Kubernetes implementation is based on a later release than 1.21-1.25, turn on Use Kubernetes version 1.26+. See also Requirements and Best practices and caveats.

    If you change this setting after downloading the location template, you need to repeat the deployment procedure.

  7. Optional You can Enable ICMP request type for Network Availability Monitors executions.

  8. Optional You can turn off support for browser monitors. If you do so, the ActiveGate node will be treated as browserless.

  9. Optional Turn on Private location alert. It will generate a problem and send an alert when all ActiveGates at the private location go offline.

  10. Select Next.

2. Deploy the location

  1. Provide an ActiveGate name or use the default. This name is used as the prefix for ActiveGates deployed as part of the location. The first ActiveGate is named <prefix>-0, the second ActiveGate <prefix>-1, and so on. This name is also used as the StatefulSet name.

  2. Provide a Location namespace name or use the default.

  3. Select Download synthetic.yaml. This is the location template file. You can rename the file to match your location for easy identification.

  4. Copy the downloaded location template over to your Kubernetes cluster.

  5. Generate deployment commands.

  6. Execute the following commands in your terminal:

    • The command for creating namespace and secrets for the docker registry and ActiveGate.
    • The command for deploying the ActiveGate.
  7. Select Finish or proceed with Deploy Metric adapter.

1. Deploy the Synthetic metric adapter

This procedure generates a separate template for the Synthetic metric adapter. You then execute generated commands on your Kubernetes cluster to deploy the metric adapter.

  • You need to deploy the Synthetic metric adapter just once per Kubernetes cluster.
  • Installing a Synthetic metric adapter requires a Kubernetes super-user role to create ClusterRoles and ClusterRoleBindings.
  1. Expand Deploy Metric adapter.

  2. Provide a Metric adapter namespace name or use the default.

  3. Select Download synthetic-adapter.yaml. This is the template file for the Synthetic metric adapter.

  4. Copy the downloaded metric adapter template over to your Kubernetes cluster.

  5. Generate deployment commands.

  6. Execute the following commands in your terminal:

    • The command for creating namespace and secrets for the docker registry and ActiveGate.
    • The command for deploying the ActiveGate.
  7. Select Finish.

Install FIPS-enabled containerized location via API

  1. Set location to FIPS mode.

    Set the fipsMode property in the request JSON using REST API.

  1. Perform additional configuration:

    • Browser-supporting.

      • Provide a certificate for re-signing HTTPS requests (explained in FIPS mode):

        kubectl -n $NAMESPACE create secret tls synthetic-fips-proxy-cert --cert=squid.crt --key=squid.key
    • Browser-supporting with corporate proxy.

      • Provide a certificate for re-signing HTTPS requests (explained in FIPS mode):

        kubectl -n $NAMESPACE create secret tls synthetic-fips-proxy-cert --cert=squid.crt --key=squid.key
      • Provide configuration of corporate proxy to Squid (explained in FIPS mode with corporate proxy):

        kubectl -n $NAMESPACE create secret generic synthetic-fips-proxy-peer --from-literal='peer.conf=cache_peer proxy.example.com parent 443 0 default no-digest proxy-only login=proxyuser:proxypass tls tls-min-version=1.2 tls-options=NO_SSLv3'
    • Browserless. No additional configuration is needed.
  2. Continue to download and deploy the YAML template as described in the Installation section.

Update a containerized location or its ActiveGates

Any updates to a location require that you download the location template file again and apply the changes via kubectl.

  1. Go to the Synthetic Classic Synthetic > Private locations

  2. Select your location in Private Synthetic locations.

  3. Switch to the Private location deployment tab.

  4. Re-enter the ActiveGate name and Location namespace that you provided during the location deployment.

  5. Select Download synthetic.yml to download a new location template file.

  6. Rename the file to match your location for easy identification.

  7. Copy the template file over to your Kubernetes cluster.

  8. Execute the following command to apply the changes on your Kubernetes cluster. Be sure to use your location template filename in place of synthetic.yaml. Execute this command from the same location as the template file.

    kubectl apply -f ./synthetic.yaml

Any update redeploys ActiveGates in the reverse order of their deployment. For example, if your location contains the ActiveGates activegate-name-0 and activegate-name-1, activegate-name-1 is stopped and redeployed first.

The redeployed ActiveGate pod uses the same persistent volume deployed for log continuity.

Delete a location or metric adapter on Kubernetes

The commands generated when deploying a location and the Synthetic metric adapter also include code snippets for deleting them on Kubernetes. You may copy and store these commands for future reference.

At any point, you can regenerate the commands for the respective namespaces.

  • If you've renamed a template file, use the new filename in the commands.
  • The cleanup commands shown below don't delete the respective namespaces.

Delete location

  1. Go to the Synthetic Classic Synthetic > Private locations
  2. Select your location in Private Synthetic locations.
  3. Switch to the Private location deployment tab.
  4. Expand Delete command (optional), and then copy and run the provided deletion command.

The procedure only deletes the Synthetic private location but doesn't delete its namespace.

Delete Synthetic metric adapter

  1. Go to the Synthetic Classic Synthetic > Private locations
  2. Select your location in Private Synthetic locations.
  3. Switch to the Metric adapter deployment tab.
  4. Expand Delete command (optional), and then copy and run the provided deletion command.

The procedure only deletes the metric adapter but doesn't delete its namespace.

Multi-AZ PVC access on cloud clusters

Using a multi-availability zone (multi-AZ) cluster with deployments utilizing PVC can result in pods being stuck in a pending state upon recreation. This happens because the storage volumes like EBS are not replicated between zones.

PVC is only shared between nodes located in the same availability zone. When you use a multi-AZ cluster and a node tries to access PVC from a different availability zone, it will become stuck in pending state and display an error message.

Currently, there are two possible solutions for multi-AZ Kubernetes deployments:

  • Use node affinity to limit pods to a specific zone
  • Use shared storage systems like EFS

Use node affinity to limit pods to a specific zone

You can configure node affinities to use only specific zones for a deployment.

To set node affinity

  1. Use the following command to find the zone each node is deployed on:
    kubectl get nodes --show-labels
  2. Look for a label failure-domain.beta.kubernetes.io/zone, for example, failure-domain.beta.kubernetes.io/zone=us-east-1a.
  3. Use the kubectl label command to set a custom label for a node:
    kubectl label nodes node name label=value
    Example
    kubectl label nodes ip-10-179-202-73.ec2.internal zone=us-east-1a
  4. Add the custom node label to the nodeSelector section of the Synthetic deployment template. For example:
    spec:
    nodeSelector:
    zone: us-east-1a
  5. Save your changes.
  6. Apply the template.

Nodes with the same zone label will be deployed in the same availability zone and you'll be able to share PVC between them without causing an error.

Use shared storage systems

Each cloud service provides its own shared storage systems options. To explain how to use shared storage systems, we will use AWS EFS as the example. For information about storage systems used by other cloud storage providers, see:

We assume that you already have EFS that you can use. If you don't, see Getting started with Amazon EFS to learn how to set up EFS.

Be aware that EFS may be more expensive than EBS. Check pricing.

To use storage class with EFS

  1. Complete the Synthetic deployment template similar to the example below:
    kind: StorageClass
    apiVersion: storage.k8s.io/v1
    metadata:
    name: efs-test
    provisioner: efs.csi.aws.com
    parameters:
    fileSystemId: fs-0c155dcd8425aa39d
    provisioningMode: efs-ap
    directoryPerms: "700"
    basePath: "/"
  2. Modify the volumeClaimTemplates section of the template similar to the example below:
    volumeClaimTemplates:
    - metadata:
    name: persistent-storage
    spec:
    storageClassName: efs-test
    accessModes:
    - ReadWriteMany
    resources:
    requests:
    storage: 3Gi
  3. Save your changes.
  4. Apply the template.

Now, if the pod is redeployed on a node in a different zone, the PVC should be automatically bound to the new deployment zone.

NAM monitors on containerized locations

OpenShift

OpenShift uses Security Context Constraint for limiting capabilities used by the pods. By default, deployed pods will use the restricted-v2 SCC, which does not allow any additional capabilities. The recommended solution is to prepare a custom Security Context Constraint.

  1. Create a dedicated Service Account Optional

    • If the custom SCC is used just by the synthetic deployment, it's recommended to create a dedicated Service Account.
      oc -n $NAMESPACE create sa sa-dt-synthetic
      oc -n $NAMESPACE adm policy add-role-to-user edit system:serviceaccount:$NAMESPACE:sa-dt-synthetic
  2. Create a custom Security Context Constraint

    • scc-dt-synthetic.yaml

      apiVersion: security.openshift.io/v1
      kind: SecurityContextConstraints
      metadata:
      name: scc-dt-synthetic
      allowPrivilegedContainer: false
      allowHostDirVolumePlugin: false
      allowHostIPC: false
      allowHostNetwork: false
      allowHostPID: false
      allowHostPorts: false
      runAsUser:
      type: MustRunAsRange
      seccompProfiles:
      - runtime/default
      seLinuxContext:
      type: MustRunAs
      fsGroup:
      type: MustRunAs
      supplementalGroups:
      type: MustRunAs
      volumes:
      - configMap
      - downwardAPI
      - emptyDir
      - persistentVolumeClaim
      - projected
      - secret
      users: []
      groups: []
      priority: null
      readOnlyRootFilesystem: true
      requiredDropCapabilities:
      - ALL
      defaultAddCapabilities: null
      allowedCapabilities:
      - NET_RAW
      allowPrivilegeEscalation: true
    • priority can be set to any number between 1 and 9. If there are two or more SCCs that fulfill the requirements, the one with higher priority is selected.

      oc create -f scc-dt-synthetic.yaml
  3. Add the new SCC to the Service Account used for synthetic deployment

    oc -n $NAMESPACE adm policy add-scc-to-user scc-dt-synthetic system:serviceaccount:$NAMESPACE:default
    • If the sa-dt-synthetic SA was created, substitute it in place of default.
      oc -n $NAMESPACE adm policy add-scc-to-user scc-dt-synthetic system:serviceaccount:$NAMESPACE:sa-dt-synthetic

Azure RedHat OpenShift (ARO)

If the OpenShift cluster is deployed as an Azure Red Hat OpenShift (ARO) resource, by default, the Network Security Group won't allow ICMP traffic outside the cluster.

The AROs Network Security Group is not modifiable, but a custom NSG can be created and imported during the ARO cluster creation. To learn more about it, see Bring your own Network Security Group (NSG) to an Azure Red Hat OpenShift (ARO) cluster.

Running the cluster with default settings will only allow for using ICMP NAM monitors for resources inside the OpenShift cluster. Any requests going outside the cluster will fail.

Proxy configuration

Add the following code at the top of your location template file to insert a ConfigMap resource containing your proxy server information.

In the code sample below:

  • The proxy server is used for connections to the Dynatrace Cluster and tested resources.
  • The namespace (namespace: dynatrace) must be the location namespace.
kind: ConfigMap
apiVersion: v1
data:
custom.properties: |-
[http.client]
proxy-server = 10.102.43.210
proxy-port = 3128
proxy-user = proxyuser
proxy-password = proxypass
metadata:
name: ag-custom-configmap
namespace: dynatrace
---

Add the following code at spec.template.spec.volumes:.

- name: ag-custom-volume
configMap:
name: ag-custom-configmap
items:
- key: custom.properties
path: custom.properties

Add the following code to the ActiveGate container configuration under volumeMounts:.

- name: ag-custom-volume
mountPath: /var/lib/dynatrace/gateway/config_template/custom.properties
subPath: custom.properties

Dynatrace Operator compatibility

If you deploy Dynatrace Operator in the same Kubernetes cluster as a containerized Synthetic location, the generated StatefulSet template must include the pod annotation dynatrace.com/split-mounts: "true". This annotation prevents conflicts that occur when the Operator's injection encounters an ActiveGate image that already contains a /var/lib/dynatrace directory.

Starting with Dynatrace version 1.335, the location template generated by Dynatrace includes this annotation. If you manage the StatefulSet manifest manually or have an earlier version, make sure the annotation is present under spec.template.metadata.annotations. For details, see Dynatrace Operator 1.8.0 release notes.

Browserless private Synthetic locations

In general, we recommend the deployment of complete synthetic private locations to support the execution of all types synthetic monitors (HTTP, browser, NAM).

If you don't need to run browser monitors, consider deploying your location in browserless mode. This mode deploys the location (or ActiveGate belonging to it) without a browser, reducing hardware requirements. However, browser monitors can't run on a browserless location.

Consider browserless locations as an alternative to synthetic private locations with browser monitor support when you’re focused purely on:

  • Network and infrastructure use cases (using NAM monitors)
  • API monitoring (using HTTP monitors)

Compared to the regular template, following changes are introduced:

  1. browser value is set for DT_SYNTHETIC_UNSUPPORTED_MONITORING_MODULES environment variable in synthetic-vuc container under env:

    - name: DT_SYNTHETIC_UNSUPPORTED_MONITORING_MODULES
    value: "browser"
  2. No synthetic-vuc-worker containers are included

  3. No chromium-cache volume is specified or mounted

Kerberos authentication configuration

Add the following code at the top of your location template file to insert a ConfigMap resource containing your Kerberos server information.

In the code sample below:

  • The EXAMPLE.COM realm is used in Kerberos authentication.
  • The example.com domain is used in Kerberos authentication.
  • The kerberos.example.com is the hostname of Key Distribution Center.
  • The namespace (namespace: dynatrace) must be the location namespace.
kind: ConfigMap
apiVersion: v1
data:
krb5.conf: |-
[libdefaults]
dns_lookup_realm = false
ticket_lifetime = 24h
renew_lifetime = 7d
forwardable = true
rdns = false
pkinit_anchors = FILE:/etc/pki/tls/certs/ca-bundle.crt
spake_preauth_groups = edwards25519
dns_canonicalize_hostname = fallback
qualify_shortname = ""
default_realm = EXAMPLE.COM
default_ccache_name = /tmp/krb5cc_%{uid}
[realms]
EXAMPLE.COM = {
kdc = kerberos.example.com
admin_server = kerberos.example.com
}
[domain_realm]
.example.com = EXAMPLE.COM
example.com = EXAMPLE.COM
metadata:
name: krb-map
namespace: dynatrace
---

Add the following code at spec.template.spec.volumes:.

- name: krb5-conf
configMap:
name: krb-map
items:
- key: krb5.conf
path: krb5.conf

Add the following code to every synthetic-vuc-worker container configuration under volumeMounts:.

- name: krb5-conf
mountPath: /etc/krb5.conf
subPath: krb5.conf

Synthetic metric adapter

Disable domain certificate validation

Add the following code to Synthetic metric adapter template under env:

- name: TLS_SECURE
value: "false"

This deactivates certificate validation for the Synthetic metric adapter connection to the Dynatrace Cluster (by default, it is activated).

Proxy configuration

Add the following code to Synthetic metric adapter template under env:

- name: HTTPS_PROXY
value: "http://proxyuser:proxypass@10.102.43.210:3128"
- name: NO_PROXY
value: "172.20.0.0/16" # do not proxy internal calls to Kubernetes cluster

For more details about these environment variables see Go httpproxy package documentation.

The way of obtaining Service CIDR depends on Kubernetes distribution, for example for AWS EKS the following command can be used:

aws eks describe-cluster --name my-cluster --query 'cluster.kubernetesNetworkConfig'

Non-scalable containerized locations

Auto-scalable locations become non-scalable for any of the following reasons.

  • The location reaches its maximum number of pods in the StatefulSet, and location utilization is over the threshold of 80%. No new ActiveGate pods are created until the maximum number of ActiveGates is increased.

  • The Synthetic metric adapter stops working, and the location horizontal pod auto-scalers don't receive the metrics required for auto-scaling.

    You can run the following command to verify the state of a pod auto-scaler. In the example below, dynatrace is the location namespace.

    kubectl describe hpa -n dynatrace

    If ScalingActive is set to False in the output, the auto-scaler isn't receiving metric data.

API: Synthetic - Locations, nodes, and configuration API v2

You can automate the deployment of and manage containerized locations via the existing Synthetic - Locations, nodes, and configuration API v2. Early Access endpoints added to this API to facilitate the deployment of Kubernetes locations. The new endpoints help generate the commands you need to execute on the Kubernetes cluster.

  • The GET location YAML endpoint (/synthetic/locations/{LocationId}/yaml) fetches the location template file based on the location ID of the location you initially set up for containerized deployment.
  • The GET apply commands endpoint (synthetic/locations/commands/apply) fetches the list of commands to deploy a location on Kubernetes/Openshift.
  • The GET delete commands endpoint (synthetic/locations/{LocationId}/commands/delete)fetches the commands to delete a containerized location.
Related tags
Digital Experience