Qorus Integration Engine®  5.1.10_git
Data Pipelines

Data Pipeline Introduction

All Qorus integration objects are described by YAML metadata produced by our IDE, which is available as a free (as in free of charge and also open source) extension to Microsoft Visual Studio Code, a multi-platform editor from Microsoft. The Qorus extension is called Qorus Developer Tools and can be installed directly from Visual Studio Code.

Data pipelines allow for complex and efficient processing of record-based data from a single input to an arbitrary number of outputs in an arbitrary number of processing queues.

Data pipelines by default have a single processing queue (with queue ID 0); to split processing of a single input, the last element of a queue can be a list of queues, each of which taking the data record(s) and processing them in parallel with other queues at the same level.

Besides the final element of a pipeline, which can be a series of queues for onwards parallel processing, each queue is a set of linear elements that takes data record input and outputs zero or more records to the next element in the pipeline. Pipeline elements can:

  • transform data record(s)
  • filter data record(s)
  • generate new record(s)

Note that all of the above operations can be performed at the same time on any given input record.

Pipeline processing elements are made up of objects of the following types:

In fact a mapper is a specialized configuration-driven form of a processor that does data format conversions.

Data Pipeline Input

Data pipelines can process input data in one of three ways:

  • If an input provider was configured for the pipeline, the configuration for the input provider will provide the initial input data for the first elements in the pipeline
  • If no input provider was configured, then the current FSM state data can be submitted as the initial input data to the pipeline
  • The first element in the pipeline can ignore any input data provided by the above two mutually-exclusive options and submit its own data for further processing
Note
If an input data factory is configured, the current state data is used as the $local: data hash when resolving factory options with UserApi::expandTemplatedValue(); if the current state data is a hash, it is used directly and can be referenced in factory options directly in $local: (ex: state key id can be directly referenced as $local:id); otherwise, if the current state data is not a hash, it can be referenced as $local:input.

Data Pipeline Processors

A data pipeline processor is a Qorus class object defined as a data processor (see Class Processor for more information).

Data Pipeline Bulk Processing

Bulk processing is when a data pipeline processor is capable of processing data in "hash of lists" form. This format is the most efficient way to process many records of data in a single operation which results in very fast and efficient processing for large amounts of data.