Pipelines are the center of data movement and payload orchestration in the Intelligence Hub. Use Pipelines to curate, apply logic, and optimize datasets for specific applications using a simple but flexible graphical user interface. Build out stages in a pipeline to model, filter, buffer, or transform data flows to optimize delivery for consuming applications and device nodes.
Orchestrate the movement of data on time intervals or logic-based events. Delay the delivery of data or buffer data based on time or size. Maintain the state of data that has been moved.
Sequentially transform data with pre-built processing stages, custom expressions, or even third-party JavaScript libraries. Process complex event streams of data structures with modeling, validation, and end-to-end observability. Tailor modeled data to meet the needs of multiple consuming applications and services.
Compose data extraction routines from simple input reads to sophisticated parameter-based queries, multi-stage lookups, cross references, payload appending, and more. Publish data to destination systems and dynamically drive schema, topics, keys, and identifiers. Define success/failure criteria of an integration and enforce these criteria with custom error handling.
The transportation of contextualized data continues to be a key pain point for industrial organizations. Solve this pain point with Pipelines. Read data from an input or instance source and then publish that data to one or many target connections with Pipelines. Pipelines can be triggered by events or intervals, while running the execution of each stage in the pipeline. Replay pipeline runs to observe data transformations from stage to stage, observe and react to any errors, and monitor pipeline and stage statistics for delays.
Most target systems have limitations on how they can process inbound data. They may be incapable of parsing or filtering structured data. They may consume multiple records in batches or files. Use Pipelines to dynamically break up objects and arrays, discard unnecessary elements to facilitate easy consumption, and buffer data on time or record count. Publish as JSON payloads or as CSV and Parquet files—or compress those files prior to delivery. This enables target systems to efficiently consume industrial data regardless of how source systems produce and transmit it.
Some data sets must be sourced from multiple systems, where the data from one system is being used in the subsequent read. Based on the values sourced, the pipeline data may go down one or multiple conditional paths with unique structuring, transformation, filtering, and writing to target systems. Pipelines can employ metadata to facilitate dynamic reads as well as curate the presentation of data to the needs of the systems consuming it.
Industrial data is not only produced as primitive tags or points, but also as complex structures. Instead of decomposing these structures and explicitly mapping them into model instances, Pipelines can validate and re-shape incoming data structures to adhere to model definitions. This is ideal for enforcing data quality or modeling many data sources that produce similar data structures with subtle differences.
Stage Type |
Stage Name |
Common |
Breakup Filter Flatten Model Model Validation Size Buffer Timed Buffer Transform |
Control |
On Change Switch |
File Format |
CSV Gzip JSON Parquet Zip |
I/O |
Read Smart Query Write Write New |
Trigger |
Event Flow Polled |
Join the free trial program to get hands-on access to all the features and functionality within HighByte Intelligence Hub and start testing the software in your unique environment.