Metrics in OpenPipeline are currently in Preview and only accessible to selected customers. If you would like to share feedback or ideas, please join our dedicated Community user group, or reach out to your Customer Success Manager.
For more information, see
Dynatrace OpenPipeline can reshape incoming data for better understanding, processing, and analysis. OpenPipeline processing is based on rules that you create and offers a flexible way of extracting value from raw records.
Source of ingestion for a data type, collecting data from the provider into Dynatrace Platform, for example, API endpoints or OneAgent.
Assignation of data to a pipeline, based either on matching conditions (dynamic) or direct assignation (static).
Once data is ingested and routed, OpenPipeline processing occurs in pipelines. Each pipeline contains a set of processing instructions (processors), that are executed in an ordered sequence of stages and define how you want Dynatrace to structure, separate, and store your data. After a record is processed, it's sent to storage and is available for further analysis.
By default data types are processed in dedicated built-in pipelines. You can create new custom pipelines to group processing and extraction by technology or team.
Log and business event processing pipeline conditions are included in the built-in OpenPipeline pipelines.
Processing is based on available records, and doesn't take into account record enrichment from external services.
OpenPipeline processing for logs supports DQL only. If you use already the log processing pipeline, ensure your matchers are converted to DQL.
A stage is a phase in a pipeline sequence that focuses on a task, such as masking, filtering, processing, or extraction. Stages contain a predefined list of configurable processors, which define the task of the stage.
The following table is a comprehensive list of stages, ordered in the pipeline sequence of execution, specifying which processors are available and executed for each stage, for the supported data types.
Stage
Description
Processors in the stage
Executed processors
Supported data types
Processing
Prepare data for analysis and storage by parsing values into fields, transforming the schema, and filtering the data records. Fields are edited, and sensitive data is masked.
All matches
Logs, Events, Business events, Metrics 1
Metric extraction
Extract metrics from the records that match the query.
All matches
Logs, Events, Business events, System events
Data extraction
Extract and resend log data into another pipeline.
All matches
Logs, Events, Business events, System events
Permissions
Apply security context to the records that match the query.
First match only
Logs, Events, Business events, Metrics 1
Storage
Assign records to the best-fit bucket.
First match only
Logs, Events, Business events, Spans
Specific metric fields are excluded from matching and processing. To learn more, see OpenPipeline limits.
A processor is a pre-formatted processing instruction that focuses either on modifying (for example, by renaming or adding a new field) or extracting data (for example, by creating an event from a log line or extracting metrics).
While the processor format is predefined, it contains a configurable matcher and processing definition.
The following table lists alphabetically all available processors in a pipeline.
Processor
Description
Add fields
Adds fields with name and value.
Business event
Extracts fields into a new record and sends it to the business event table.
Counter metric
Returns the number of occurrences of a metric, from the records that match the query.
Davis event
Extracts fields into a new record and sends it to an event table.
DQL
Processes a subset of DQL. The output is formatted to string, number, bool, duration, timestamp, and respective arrays of those.
Drop record
Drops a record. The record is not retained.
No storage assignment
Skips storage assignment. The record is not retained.
Remove fields
Removes fields from the record.
Rename fields
Changes the name of fields.
Bucket assignment
Assigns a Grail bucket.
Value metric
Returns the aggregated values of a metric from the records that match the query.
Set dt.security_context
Sets the proper record-level access via dt.security_context
by either copying it from a field, setting it as a static string, or a static array that allows multiple values.
Technology bundle
Matches records for the selected technology and processeses them according to predefined context-sensitive DQL processing statements.