Dataset Viewer
Auto-converted to Parquet Duplicate
category
stringclasses
11 values
name
stringlengths
3
30
type
stringlengths
4
32
shortDescription
stringlengths
16
269
description
stringlengths
28
927
inputPorts
listlengths
0
3
outputPorts
listlengths
0
4
properties
listlengths
0
46
defaultVisibility
unknown
usage
stringlengths
212
1.18k
viewGridData
bool
2 classes
passThrough
bool
1 class
viewPlainData
bool
1 class
readers
UniversalDataReader
DATA_READER
Reads data from flat files (CSV, fixed-length).
Reads data from delimited or fixed length files and sends data records to first connected output port. If any error occurs, the report is sent to the second port if "Data policy" is set to "Controlled". This component is alias for FlatFileReader.
[ { "type": "singlePort", "name": "0", "required": false, "label": "input" } ]
[ { "type": "singlePort", "name": "0", "required": true, "label": "output" }, { "type": "singlePort", "name": "1", "label": "logs", "metadata": { "id": "UniversalDataReader_Error" } } ]
[ { "name": "fileURL", "label": "File URL", "description": "Specifies the path or URL of the input file to read data from. This is the primary attribute that determines which file will be processed by the component.", "modifiable": true, "nullable": true, "required": true, "primaryAttribut...
null
null
null
null
null
writers
UniversalDataWriter
DATA_WRITER
Writes data to flat files (CSV, fixed-length).
Receives data through connected input port and writes data records to delimited or fixed length files based on input metadata. This component is alias for FlatFileWriter.
[ { "type": "singlePort", "name": "0", "required": true } ]
[ { "type": "singlePort", "name": "0", "required": false, "metadataProvider": { "id": "textBasedWriterMetadataProvider" } } ]
[ { "name": "fileURL", "label": "File URL", "description": "Specifies the path or URL of the file to process. This is the primary attribute that determines which file will be processed by the component.", "modifiable": true, "nullable": true, "required": true, "primaryAttribute": true, ...
null
null
null
null
null
readers
FlatFileReader
FLAT_FILE_READER
Reads data from flat files (CSV, fixed-length). Replaces UniversalDataReader (UDR).
Reads data from delimited or fixed length files and sends data records to first connected output port. If any error occurs, the report is sent to the second port if "Data policy" is set to "Controlled". This component is alias for UniversalDataReader.
[ { "type": "singlePort", "name": "0", "required": false, "label": "input" } ]
[ { "type": "singlePort", "name": "0", "required": true, "label": "output" }, { "type": "singlePort", "name": "1", "label": "logs", "metadata": { "id": "UniversalDataReader_Error" } } ]
[ { "name": "fileURL", "label": "File URL", "description": "Specifies the path or URL of the file to process. This is the primary attribute that determines which file will be processed by the component.", "modifiable": true, "nullable": true, "required": true, "primaryAttribute": true, ...
null
null
null
null
null
writers
FlatFileWriter
FLAT_FILE_WRITER
Writes data to flat files (CSV, fixed-length). Replaces UniversalDataWriter (UDW).
Receives data through connected input port and writes data records to delimited or fixed length files based on input metadata. This component is alias for UniversalDataWriter.
[ { "type": "singlePort", "name": "0", "required": true } ]
[ { "type": "singlePort", "name": "0", "required": false, "metadataProvider": { "id": "textBasedWriterMetadataProvider" } } ]
[ { "name": "fileURL", "label": "File URL", "description": "Specifies the path or URL of the file to process. This is the primary attribute that determines which file will be processed by the component.", "modifiable": true, "nullable": true, "required": true, "primaryAttribute": true, ...
null
null
null
null
null
deprecated
DelimitedDataReader
DELIMITED_DATA_READER
Reads data from delimited files.
Reads data from delimited files and sends data records to all connected output ports.
[ { "type": "singlePort", "name": "0", "required": false, "label": "input" } ]
[ { "type": "multiplePort", "required": true } ]
[ { "name": "fileURL", "label": "File URL", "description": "Specifies the path or URL of the file to process. This is the primary attribute that determines which file will be processed by the component.", "modifiable": true, "nullable": true, "required": true, "primaryAttribute": true, ...
false
null
null
null
null
deprecated
DelimitedDataWriter
DELIMITED_DATA_WRITER
Writes data to delimited files.
Receives data through connected input port and writes data records to delimited files based on input metadata.
[ { "type": "singlePort", "name": "0", "required": true } ]
[ { "type": "singlePort", "name": "0", "required": false } ]
[ { "name": "fileURL", "label": "File URL", "description": "Specifies the path or URL of the file to process. This is the primary attribute that determines which file will be processed by the component.", "modifiable": true, "nullable": true, "required": true, "primaryAttribute": true, ...
false
null
false
null
null
transformers
SimpleCopy
SIMPLE_COPY
Copies data to output ports.
Receives data through connected input port and copies data records to all connected output ports.
[ { "type": "singlePort", "name": "0", "required": true } ]
[ { "type": "multiplePort", "name": "0", "required": true } ]
[]
null
When to use: Broadcast one input stream to multiple output ports unchanged — every record goes to ALL connected outputs simultaneously. Use when the same data needs to feed multiple downstream branches in parallel (e.g. write to DB and to a CSV file at the same time, or feed two different transformation chains with the...
null
true
null
transformers
Concatenate
CONCATENATE
Concatenates data flows.
Receives data through all connected input ports in turn. Gets all records from each input port, sends them to output port and continues with the next input port. Skips ports without incoming data. Terminates when all incoming records are received and sent to connected output port.
[ { "type": "multiplePort", "required": true } ]
[ { "type": "singlePort", "name": "0", "required": true } ]
[]
null
When to use: Merge N streams where output must preserve port order — all records from port 0 first, then all from port 1, etc. Order within each stream is preserved. Constraints: Slower than SIMPLE_GATHER — waits for each port to finish before moving to the next. Input does NOT need to be sorted.
null
true
null
transformers
SimpleGather
SIMPLE_GATHER
Gathers data flows.
Receives data incoming through all connected input ports in the order so that all records are gathered and send out through output port as soon as possible.
[ { "type": "multiplePort", "required": true } ]
[ { "type": "multiplePort", "required": true } ]
[]
null
When to use: Merge N streams where output order doesn't matter. Fastest option — emits records from whichever port has data ready. Constraints: Output order is non-deterministic. Do NOT use if downstream requires sorted or port-ordered output.
null
true
null
transformers
Map
REFORMAT
Maps (transforms) data from one input to one or multiple outputs. Replaces Reformat component.
Receives data through connected input port, transforms them in a user specified way, and sends such new outgoing data records to all connected output ports. Implements org.jetel.component.RecordTransform interface.
[ { "type": "singlePort", "name": "0", "required": true } ]
[ { "type": "multiplePort", "required": true } ]
[ { "name": "transform", "label": "Transform", "description": "Defines the transformation logic used to convert input records to output records.", "modifiable": true, "nullable": true, "required": "!transformClass,!transformURL", "redundant": "transformClass|transformURL", "primaryAttr...
null
When to use: 1->1 transformation where each input record produces one (modified, enriched, or restructured) output record. Primary use is to transform data structure or content between input and output metadata. Optionally, can produce records on different output ports to support routing or different output structures;...
null
null
null
readers
DatabaseReader
DB_INPUT_TABLE
Or DBReader, replaces DBInputTable. Reads data from database using JDBC driver. Supported databases are Amazon Redshift, Microsoft Access, Microsoft SQL Server, MySQL, Oracle, PostgreSQL, SQLite, Sybase, Vertica, DB2 and any other database with JDBC compliant driver.
Connects to database using JDBC driver, executes specified query on DB, extracts the rows returned by the query and sends them as data records to all connected output ports. Output metadata must precisely describe the structure of extracted rows.
[ { "type": "singlePort", "name": "0", "required": false } ]
[ { "type": "multiplePort", "required": true } ]
[ { "name": "dbConnection", "label": "DB connection", "description": "Specifies the database connection to use for database operations.", "modifiable": true, "nullable": false, "required": true, "type": "connection" }, { "name": "sqlQuery", "label": "SQL query", "descriptio...
null
null
null
null
false
deprecated
Sort
SORT
Sorts incoming data records.
Receives data through connected input port, sorts data records according to specified sort key and sends them to all connected output ports. Sort key is name or combination of names of field(s) of incoming records. Sort order can be either Ascending (default) or Descending. This component is deprecated and should not b...
[ { "type": "singlePort", "name": "0", "required": true } ]
[ { "type": "multiplePort", "required": true } ]
[ { "name": "sortOrder", "label": "Sort order", "description": "Specifies whether to sort data in ascending or descending order.", "modifiable": true, "nullable": true, "defaultValue": "Ascending", "primaryAttribute": true, "type": "order" }, { "name": "sortKey", "label": "...
false
null
false
true
false
writers
DatabaseWriter
DB_OUTPUT_TABLE
Or DBWriter, replaces DBOutputTable. Writes data into database using JDBC driver. Supported databases are Amazon Redshift, Microsoft Access, Microsoft SQL Server, MySQL, Oracle, PostgreSQL, SQLite, Sybase, Vertica, DB2 and any other database with JDBC compliant driver.
Receives data through connected input port, connects to database using JDBC driver, performs database operation and loads data records into specified DB table. Correspondence between input metadata and DB table columns must be defined.
[ { "type": "singlePort", "name": "0", "required": true } ]
[ { "type": "singlePort", "name": "0", "label": "rejected" }, { "type": "singlePort", "name": "1", "label": "autogenerated" } ]
[ { "name": "dbConnection", "label": "DB connection", "description": "Specifies the database connection to use for database operations.", "modifiable": true, "nullable": false, "required": true, "type": "connection" }, { "name": "sqlQuery", "label": "SQL query", "descriptio...
null
When to use: Write records from a data stream into a database table via JDBC. Three configuration modes — choose based on how closely field names match column names and how much SQL control is needed: - dbTable (auto-mapping): Field names in metadata exactly match DB column names. CloverDX auto-generates the INSERT. Si...
false
null
false
deprecated
FixLenDataWriter
FIXLEN_DATA_WRITER
Writes data to fixed length files.
Receives data through connected input port and writes data records to fixed length files based on input metadata.
[ { "type": "singlePort", "name": "0", "required": true } ]
[ { "type": "singlePort", "name": "0", "required": false } ]
[ { "name": "fileURL", "label": "File URL", "description": "Specifies the path or URL of the file to process. This is the primary attribute that determines which file will be processed by the component.", "modifiable": true, "nullable": true, "required": true, "primaryAttribute": true, ...
false
null
false
null
null
transformers
Dedup
DEDUP
Removes duplicate records.
Receives sorted data records through connected input port and removes records that are in duplicate with a view to the specified key values. Keeps defined number of records from either the start (First) or the end (Last) of each group with the same key value. If desired, only unique records are kept. Dedup key is name ...
[ { "type": "singlePort", "name": "0", "required": true } ]
[ { "type": "singlePort", "name": "0", "required": true, "label": "unique" }, { "type": "singlePort", "name": "1", "required": false, "label": "duplicate" } ]
[ { "name": "dedupKey", "label": "Dedup key", "description": "Defines the fields used to identify duplicate records for removal.", "modifiable": true, "nullable": true, "primaryAttribute": true, "type": "key", "inputPortName": "0", "keyType": "orderVerification", "rightLabel": ...
null
When to use: Remove duplicate records from a stream based on key field values. Keeps the first, last, or a defined count of records per duplicate group. Optionally routes duplicates to a second output port for inspection or logging rather than silently discarding them. If no dedupKey is defined, the component operates ...
null
true
null
deprecated
FixLenDataReader
FIXLEN_DATA_READER
Reads data from fixed length files.
Reads data from fixed length files and sends data records to all connected output ports.
[ { "type": "singlePort", "name": "0", "required": false, "label": "input" } ]
[ { "type": "multiplePort", "required": true } ]
[ { "name": "fileURL", "label": "File URL", "description": "Specifies the path or URL of the file to process. This is the primary attribute that determines which file will be processed by the component.", "modifiable": true, "nullable": true, "required": true, "primaryAttribute": true, ...
false
null
null
null
null
transformers
Merge
MERGE
Merges data flows.
Receives sorted data through all connected input ports in turn. Gets all incoming records and sends them to connected output port sorted in the same way. Terminates when all incoming records are received and sent to connected output port.
[ { "type": "multiplePort", "required": true } ]
[ { "type": "singlePort", "name": "0", "required": true } ]
[ { "name": "mergeKey", "label": "Merge key", "description": "Defines the fields used to merge multiple sorted data streams.", "modifiable": true, "nullable": true, "required": true, "primaryAttribute": true, "type": "key", "rightLabel": "Merge key:" } ]
null
When to use: Merge N pre-sorted streams into a single sorted output stream. Equivalent to a sorted merge in database terms. Constraints: ALL inputs MUST be pre-sorted on mergeKey. Add sorting upstream if not sorted. Use CONCATENATE if sort order doesn't matter.
null
true
null
joiners
CrossJoin
CROSS_JOIN
Performs cartesian product.
Receives data records through all connected input ports and performs cartesian product on them. Each connected input port represents one set of records. Results are sent to the connected output port.
[ { "type": "singlePort", "name": "0", "required": true, "label": "master" }, { "type": "multiplePort", "required": false, "label": "slave" } ]
[ { "type": "singlePort", "name": "0", "required": true } ]
[ { "name": "transform", "label": "Transform", "description": "Defines the transformation logic used to convert input records to output records.", "modifiable": true, "nullable": true, "redundant": "transformClass|transformURL", "required": false, "defaultValue": "Default transformatio...
null
When to use: Cartesian product — every record from port 0 combined with every record from port 1. Use for combinatorial generation or when explicit all-pairs combination is needed. Constraints: Output size = |port0| × |port1|. Dangerous for large inputs. Only use when cartesian product is intentional.
null
null
null
deprecated
MergeJoin
MERGE_JOIN
Joins sorted data flows.
Receives sorted data through all connected input ports, for each join key value contained in data records incoming through the first input port (drivers) searches corresponding data records incoming through the other port(s) (slaves) with the same key value. Every combination of driver and its corresponding slaves is s...
[ { "type": "singlePort", "name": "0", "required": true, "label": "driver" }, { "type": "multiplePort", "required": true, "label": "slave" } ]
[ { "type": "singlePort", "name": "0", "required": true } ]
[ { "name": "joinKey", "label": "Join key", "description": "Defines the fields used to join data from multiple input sources.", "modifiable": true, "nullable": true, "required": true, "type": "mergeJoinKey" }, { "name": "transform", "label": "Transform", "description": "Def...
false
null
false
null
false
joiners
ExtMergeJoin
EXT_MERGE_JOIN
Joins sorted data flows.
Receives sorted data through all connected input ports, for each join key value contained in data records incoming through the first input port (drivers) searches corresponding data records incoming through the other port(s) (slaves) with the same key value. Every combination of driver and its corresponding slaves is s...
[ { "type": "singlePort", "name": "0", "required": true, "label": "driver" }, { "type": "multiplePort", "required": true, "label": "slave" } ]
[ { "type": "singlePort", "name": "0", "required": true }, { "type": "singlePort", "name": "1" } ]
[ { "name": "joinKey", "label": "Join key", "description": "Defines the fields used to join data from multiple input sources.", "modifiable": true, "nullable": true, "required": true, "type": "mergeJoinKey", "order": "true" }, { "name": "joinType", "label": "Join type", ...
null
When to use: Equality join. Both driver and slave are already sorted on the join key (or you're adding EXT_SORT upstream anyway). More memory-efficient than EXT_HASH_JOIN for large slaves. Constraints: BOTH inputs MUST be pre-sorted on join key — add EXT_SORT or FAST_SORT upstream if not. Use EXT_HASH_JOIN if inputs ar...
null
null
null
writers
Trash
TRASH
Discards all data.
Receives data through connected input port and discards it all. However, data can be written to specified file for debugging.
[ { "type": "multiplePort", "required": true } ]
[]
[ { "name": "debugPrint", "label": "Debug print", "description": "When enabled, prints debug information about processed records.", "modifiable": true, "nullable": true, "defaultValue": "false", "type": "bool" }, { "name": "debugFilename", "label": "Debug file URL", "descri...
null
null
false
null
false
null
null
TRASHIFIER
null
null
[]
[]
[]
null
null
null
null
null
others
DBExecute
DB_EXECUTE
Executes specified commands on database.
Executes specified commands (SQL/DML) on specified database table. If using stored procedure, receives records through connected input port and sends return value and/or output parameters to connected output port.
[ { "type": "singlePort", "name": "0", "label": "input parameters" } ]
[ { "type": "singlePort", "name": "0", "label": "procedure output" }, { "type": "singlePort", "name": "1", "label": "error output" } ]
[ { "name": "dbConnection", "label": "DB connection", "description": "Specifies the database connection to use for database operations.", "modifiable": true, "nullable": false, "required": true, "type": "connection" }, { "name": "sqlQuery", "label": "SQL query", "descriptio...
null
null
null
null
null
deprecated
HashJoin
HASH_JOIN
Joins unsorted data flows.
Receives data through all connected input ports, for every connected slave input port (all input ports except the first) creates hash table from data records incoming through such slave input port (for this reason, number of incoming slave data records must be sufficiently small), for each join key value contained in d...
[ { "type": "singlePort", "name": "0", "required": true, "label": "driver" }, { "type": "multiplePort", "required": true, "label": "slave" } ]
[ { "type": "singlePort", "name": "0", "required": true } ]
[ { "name": "joinKey", "label": "Join key", "description": "Defines the fields used to join data from multiple input sources.", "modifiable": true, "nullable": true, "required": true, "type": "hashJoinKey" }, { "name": "transform", "label": "Transform", "description": "Defi...
false
null
false
null
false
joiners
ExtHashJoin
EXT_HASH_JOIN
Joins unsorted data flows.
Receives data through all connected input ports, for every connected slave input port (all input ports except the first) creates hash table from data records incoming through such slave input port (for this reason, number of incoming slave data records must be sufficiently small), for each join key value contained in d...
[ { "type": "singlePort", "name": "0", "required": true, "label": "driver" }, { "type": "multiplePort", "required": true, "label": "slave" } ]
[ { "type": "singlePort", "name": "0", "required": true }, { "type": "singlePort", "name": "1" } ]
[ { "name": "joinKey", "label": "Join key", "description": "Defines the fields used to join data from multiple input sources.", "modifiable": true, "nullable": true, "required": true, "type": "hashJoinKey" }, { "name": "joinType", "label": "Join type", "description": "Speci...
null
When to use: Equality join. Input does NOT need to be sorted. Slave side loaded entirely into memory as a hash table — slave must fit in RAM. Best for large driver + small-to-medium slave. Constraints: Slave size limited by available memory. Use EXT_MERGE_JOIN instead if both streams are already sorted or if slave is l...
null
null
null
others
CheckForeignKey
CHECK_FOREIGN_KEY
Checks foreign key.
Checks foreign key values against a table of primary key values. Duplicate primary keys are ignored. Invalid foreign key values are replaced by default foreign key value. Resulting foreign data records are sent to all connected output ports. Invalid data records are sent to the optional second output port if connected.
[ { "type": "singlePort", "name": "0", "required": true, "label": "foreign records" }, { "type": "singlePort", "name": "1", "required": true, "label": "primary records" } ]
[ { "type": "singlePort", "name": "0", "required": true, "label": "foreign records" }, { "type": "singlePort", "name": "1", "label": "invalid foreign records" } ]
[ { "name": "foreignKey", "label": "Foreign key", "description": "Defines the foreign key relationship for referential integrity checking.", "modifiable": true, "nullable": true, "required": true, "type": "joinKey", "labels": "Foreign key definition;Foreign key;Primary key;;Foreign key...
"etlGraph"
null
null
null
null
readers
DBFDataReader
DBF_DATA_READER
Reads data from dBase files.
Reads data from dBase files and sends data records to all connected output ports.
[ { "type": "singlePort", "name": "0", "required": false, "label": "input" } ]
[ { "type": "multiplePort", "required": true } ]
[ { "name": "fileURL", "label": "File URL", "description": "Specifies the path or URL of the file to process. This is the primary attribute that determines which file will be processed by the component.", "modifiable": true, "nullable": true, "required": true, "primaryAttribute": true, ...
null
null
null
null
null
writers
DBFDataWriter
DBF_DATA_WRITER
Writes data to dBase files.
Writes data from input port to dBase file(s). Handles Character/Number/Logical/Date dBase data types. Requires fixed-length metadata on input.
[ { "type": "singlePort", "name": "0", "required": true } ]
[]
[ { "name": "fileURL", "label": "File URL", "description": "Specifies the path or URL of the file to process. This is the primary attribute that determines which file will be processed by the component.", "modifiable": true, "nullable": true, "required": true, "primaryAttribute": true, ...
null
null
false
null
false
transformers
Filter
EXT_FILTER
Filters incoming data.
Receives data records through connected input port, removes some of them depending on defined filter expression and sends the rest to the connected first output port. Rejected records are sent to the optional second output port if connected.
[ { "type": "singlePort", "name": "0", "required": true } ]
[ { "type": "singlePort", "name": "0", "required": true, "label": "accepted" }, { "type": "singlePort", "name": "1", "label": "rejected" } ]
[ { "name": "filterExpression", "label": "Filter expression", "description": "Defines the expression used to filter input records.", "modifiable": true, "nullable": true, "required": "!filterClass", "redundant": "filterClass", "primaryAttribute": true, "type": "filter" }, { ...
null
When to use: Split stream into two: records matching a condition (port 0) and records not matching (port 1). Boolean keep/reject. Constraints: Exactly two outputs. Use PARTITION if you need more than two destination ports or content-based routing beyond boolean.
null
true
null
transformers
ExtSort
EXT_SORT
Sorts incoming data.
Receives data records through connected input port, sorts them according to specified sort key and sends them to all connected output ports. Sort key is name or combination of names of field(s) of incoming records. Sort order can be either Ascending (default) or Descending. Any number of records can be sorted. If inter...
[ { "type": "singlePort", "name": "0", "required": true } ]
[ { "type": "multiplePort", "required": true } ]
[ { "name": "sortOrder", "label": "Sort order", "description": "Specifies whether to sort data in ascending or descending order.", "modifiable": true, "nullable": true, "defaultValue": "Ascending", "type": "order" }, { "name": "sortKey", "label": "Sort key", "description": ...
null
When to use: Dataset may be larger than available RAM. Spills to disk if buffer exhausted. Use when data size is unknown or large. When sort stability is required. Constraints: Slower than FAST_SORT due to potential disk I/O.
null
true
null
transformers
SortWithinGroups
SORT_WITHIN_GROUPS
Sorts incoming data in groups.
Receives data records through connected input port, sorts them according to specified sort key within groups of data records. The groups are defined by previous sorting of the records by the group key, with each group consisting of records considered equal when sorting by the group key. Sends sorted data records to all...
[ { "type": "singlePort", "name": "0", "required": true } ]
[ { "type": "multiplePort", "required": true } ]
[ { "name": "groupKey", "label": "Group key", "description": "Defines the fields used to group records for within-group operations.", "modifiable": true, "nullable": true, "required": true, "type": "key", "inputPortName": "0", "rightLabel": "Group key:" }, { "name": "sortKe...
null
When to use: Sort records within pre-existing groups without re-sorting the entire dataset globally. Input is already grouped by groupKey (but not necessarily sorted globally) and records within each group need internal ordering by a different sortKey. Much cheaper than a full EXT_SORT when the stream is already partit...
null
true
null
transformers
Partition
PARTITION
Splits incoming data flow.
Receives data through connected input port, splits incoming data flow into more flows and sends each of them to different connected output port.
[ { "type": "singlePort", "name": "0", "required": true } ]
[ { "type": "multiplePort", "required": true } ]
[ { "name": "partitionSource", "label": "Partition", "description": "Defines the partitioning logic used to split data into multiple output streams.", "modifiable": true, "nullable": true, "primaryAttribute": true, "type": "transformPartition" }, { "name": "partitionURL", "labe...
null
When to use: Route each record to a specific output port based on content — value ranges, hash of a key field, or CTL2 logic returning a port number. Content-aware routing. Constraints: Record goes to exactly one port. Define routing via ranges, partitionKey (hash-based), or partitionSource (CTL2).
null
true
null
transformers
DataIntersection
DATA_INTERSECTION
Intersects two data flows.
Receives sorted data flows through both input ports, finds data records with the same key values, converts these records to outgoing records according to defined transformation and sends these resulting records to the second output port. Data records contained only in the first or second data flow are sent to the first...
[ { "type": "singlePort", "name": "0", "required": true, "label": "set A" }, { "type": "singlePort", "name": "1", "required": true, "label": "set B" } ]
[ { "type": "singlePort", "name": "0", "required": false, "label": "only in A" }, { "type": "singlePort", "name": "1", "required": false, "label": "in A & B" }, { "type": "singlePort", "name": "2", "required": false, "label": "only in B" } ]
[ { "name": "joinKey", "label": "Join key", "description": "Defines the fields used to join data from multiple input sources.", "modifiable": true, "nullable": true, "required": true, "type": "joinKey", "labels": "Join key;Set A key;Set B key;;;Set A key field;Set B key field", "pr...
null
When to use: Set operations (intersect, diff, union) between two sorted streams. Produces three output ports covering all three logical regions simultaneously — more efficient than multiple EXT_FILTER passes. Constraints: BOTH inputs MUST be pre-sorted on joinKey. Use EXT_HASH_JOIN for equality join where you just want...
null
null
null
transformers
Aggregate
AGGREGATE
Aggregates information about data.
Receives data through connected input port, aggregates information about groups of adjacent records with the same aggregate key value, creates outgoing records from this information along with other input fields according to output metadata and sends these resulting records to all connected output ports. Can by used si...
[ { "type": "singlePort", "name": "0", "required": true } ]
[ { "type": "multiplePort", "required": true } ]
[ { "name": "aggregateKey", "label": "Aggregate key", "description": "Defines the key fields used for aggregate key.", "modifiable": true, "nullable": true, "type": "key", "inputPortName": "0", "rightLabel": "Aggregate key:" }, { "name": "mapping", "label": "Aggregation map...
null
When to use: Standard GROUP BY — sum, avg, count, min, max, first, last on groups. No custom logic needed. Fastest option for standard aggregations. Constraints: Only built-in aggregation functions. Input sorted=true (default) requires pre-sorted input; sorted=false handles unsorted but uses more memory. Prefer over RO...
null
null
null
deprecated
SystemExecute
SYS_EXECUTE
Executes system commands.
Executes system commands and sends their outputs to the optional output port or writes it to output file. If command requires some data, it can be get through connected optional input port.
[ { "type": "singlePort", "name": "0", "label": "input for command" } ]
[ { "type": "singlePort", "name": "0", "label": "output of command" } ]
[ { "name": "command", "label": "System command", "description": "Configures system command.", "modifiable": true, "nullable": true, "required": true, "primaryAttribute": true, "type": "multiline" }, { "name": "charset", "label": "Process input/output charset", "descrip...
false
null
null
null
null
deprecated
RunGraph
RUN_GRAPH
Runs graphs in selected instance of Clover.
Runs specified graph(s) in selected instance of Clover. Name(s) of at least one graph can be received through connected input port or name of one graph only can be set as attribute of this component. If graph name(s) is(are) received through connected input port, information about execution of the graph(s) is sent to t...
[ { "type": "singlePort", "name": "0", "required": false, "label": "in regular mode: names of the graphs to be executed, otherwise nothing", "metadata": { "id": "RunGraph_Input" } } ]
[ { "type": "singlePort", "name": "0", "required": false, "label": "regular mode: status messages, pipeline mode: an information record in case of successful termination, otherwise nothing", "metadata": { "id": "RunGraph_Output" } }, { "type": "singlePort", "name": "1", "...
[ { "name": "graphName", "label": "Graph URL", "description": "Configures graph url.", "modifiable": true, "nullable": true, "openActionLabel": "Open Graph", "primaryAttribute": true, "type": "file" }, { "name": "sameInstance", "label": "The same JVM", "description": "W...
false
null
null
null
null
others
HTTPConnector
HTTP_CONNECTOR
HTTPConnector sends HTTP requests to a web server and receives responses.
Receives an HTTP request through the connected input port, reads it from input file or directly from the graph and sends it to the HTTP server. Takes the response from the server and sends it out through the connected output port or writes it to the output file. Optionally also the output file URLs can be sent out thro...
[ { "type": "singlePort", "name": "0", "required": false, "metadata": { "id": "HTTPConnector_Request" } } ]
[ { "type": "singlePort", "name": "0", "required": false, "metadata": { "id": "HTTPConnector_Response" } }, { "type": "singlePort", "name": "1", "required": false, "metadata": { "id": "HTTPConnector_Error" } } ]
[ { "name": "url", "label": "URL", "description": "Specifies the path to an external file containing the SQL query or other configuration.", "modifiable": true, "nullable": true, "type": "httpURL" }, { "name": "requestMethod", "label": "Request method", "description": "Specifie...
null
null
null
null
null
others
RESTConnector
REST_CONNECTOR
Connects to a REST API which provides an OpenAPI specification.
Connects to a REST API which provides an OpenAPI specification. Creates and sends a request based on the data received on input ports. Parses the responses and writes them to the output ports. Provides comfortable user interface to create requests and parse responses.
[ { "type": "multiplePort", "required": false } ]
[ { "type": "multiplePort", "required": false } ]
[ { "name": "openApiUrl", "label": "OpenAPI URL", "description": "Specifies the path or URL for openapi url.", "modifiable": true, "nullable": true, "required": false, "dropUrlTarget": "true", "type": "file", "fields": "input", "dictionary": "input", "httpContext": "request...
null
null
null
null
null
deprecated
KeyGenerator
KEY_GEN
Generates new key field.
Receives data through connected input port, generates one new field based on other specified fields, adds this new field to outgoing data records and sends them to all connected output ports.
[ { "type": "singlePort", "name": "0", "required": true } ]
[ { "type": "singlePort", "name": "0", "required": true } ]
[ { "name": "keyExpression", "label": "Matching key", "description": "Defines the key fields used for matching key.", "modifiable": true, "nullable": true, "required": true, "type": "key", "inputPortName": "0", "keyType": "KeyGen", "rightLabel": "Matching key:" } ]
false
null
false
null
false
deprecated
ApproximativeJoin
APROX_MERGE_JOIN
Joins two similar data flows.
Receives data through two input ports, for each matching key value contained in driver data records (port 0) searches corresponding slave data records (port 1) with the same key value. For every pair of driver and slave records conformity based on join key is computed as Levenstein distance. Pairs with conformity highe...
[ { "type": "singlePort", "name": "0", "required": true, "label": "driver" }, { "type": "singlePort", "name": "1", "required": true, "label": "slave" } ]
[ { "type": "singlePort", "name": "0", "required": true, "label": "joined with conformity > limit" }, { "type": "singlePort", "name": "1", "required": true, "label": "joined with conformity < limit" }, { "type": "singlePort", "name": "2", "label": "driver with no ma...
[ { "name": "joinKey", "label": "Join key", "description": "Defines the fields used to join data from multiple input sources.", "modifiable": true, "nullable": true, "required": true, "type": "aproximativeJoinKey" }, { "name": "matchingKey", "label": "Matching key", "descri...
false
null
false
null
false
joiners
DBJoin
DBJOIN
Joins incoming data flow with records from database table.
Receives data through connected input port (drivers), connects to database, for each join key value contained in driver data records searches corresponding data records with the same key value in database table (slaves). Every pair of driver and slave records is sent to transformation class that joins them to form outg...
[ { "type": "singlePort", "name": "0", "required": true } ]
[ { "type": "singlePort", "name": "0", "required": true, "label": "joined records" }, { "type": "singlePort", "name": "1", "label": "skipped records" } ]
[ { "name": "joinKey", "label": "Join key", "description": "Defines the fields used to join data from multiple input sources.", "modifiable": true, "nullable": true, "required": true, "type": "key", "inputPortName": "0", "keyDuplicates": "true", "rightLabel": "Join key:" }, ...
null
When to use: Join driver stream against a database table via SQL query per record. Slave data lives in DB, not in memory. Use when slave is in a DB and too large for a lookup table, or when slave data must be current (real-time DB lookup per record). Constraints: One DB query per driver record (or per cache miss) — slo...
null
null
null
writers
CloverDataWriter
CLOVER_WRITER
Writes data to Clover files.
Receives data through connected input port and writes data records to output files in internal binary Clover data format.
[ { "type": "singlePort", "name": "0", "required": true } ]
[ { "type": "singlePort", "name": "0", "required": false, "metadataProvider": { "id": "binaryBasedWriterMetadataProvider" } } ]
[ { "name": "fileURL", "label": "File URL", "description": "Specifies the path or URL of the file to process. This is the primary attribute that determines which file will be processed by the component.", "modifiable": true, "nullable": true, "required": true, "primaryAttribute": true, ...
null
null
null
null
false
readers
CloverDataReader
CLOVER_READER
Reads data from Clover files.
Reads data from files written in internal binary Clover data format and sends data records to all connected output ports.
[ { "type": "singlePort", "name": "0", "label": "input" } ]
[ { "type": "multiplePort", "required": true } ]
[ { "name": "fileURL", "label": "File URL", "description": "Specifies the path or URL of the file to process. This is the primary attribute that determines which file will be processed by the component.", "modifiable": true, "nullable": true, "required": true, "primaryAttribute": true, ...
null
null
null
null
false
writers
StructuredDataWriter
STRUCTURE_WRITER
Writes data to files in structured way.
Receives data through connected input port(s) and writes data records to output files according to defined structure consisting of header, body and footer. Header and footer are optional.
[ { "type": "singlePort", "name": "0", "label": "Body port", "required": true }, { "type": "singlePort", "name": "1", "label": "Header port" }, { "type": "singlePort", "name": "2", "label": "Footer port" } ]
[ { "type": "singlePort", "name": "0", "required": false, "metadataProvider": { "id": "textBasedWriterMetadataProvider" } } ]
[ { "name": "fileURL", "label": "File URL", "description": "Specifies the path or URL of the file to process. This is the primary attribute that determines which file will be processed by the component.", "modifiable": true, "nullable": true, "required": true, "primaryAttribute": true, ...
null
null
false
null
null
transformers
Normalizer
NORMALIZER
Normalizes data.
Receives data through connected input port, splits each data record into more records and sends all resulting records to all connected output ports.
[ { "type": "singlePort", "name": "0", "required": true } ]
[ { "type": "singlePort", "name": "0", "required": true } ]
[ { "name": "normalize", "label": "Normalize", "description": "Configures normalize.", "modifiable": true, "nullable": true, "required": "!normalizeClass,!normalizeURL", "primaryAttribute": true, "type": "transformNormalize" }, { "name": "normalizeURL", "label": "Normalize ...
null
When to use: Split ONE input record into a variable number of output records determined at runtime (e.g. expand a list field, unpack a delimited string into individual rows). CTL entry: count() returns how many records to produce, transform(integer idx) produces record at index idx. Constraints: Purpose-built for 1→N. ...
null
null
null
transformers
Denormalizer
DENORMALIZER
Denormalizes data.
Denormalizes input records - ie composes one output record from several input records using user-specified transformation. One group is defined with the same value on the given field(s) or has the fixed size. The transformation is supposed to implement interface RecordDenormalize. Sends all resulting records to all con...
[ { "type": "singlePort", "name": "0", "required": true } ]
[ { "type": "singlePort", "name": "0", "required": true } ]
[ { "name": "key", "label": "Key", "description": "Defines the key fields used for key.", "modifiable": true, "nullable": true, "redundant": "groupSize", "type": "key", "inputPortName": "0", "keyType": "orderVerification", "rightLabel": "Key:" }, { "name": "groupSize", ...
null
When to use: Collapse N consecutive input records into exactly 1 output record per group. Simpler than ROLLUP for the pure N→1 case. Constraints: Input MUST be sorted on key. Exactly one output record per group (unlike ROLLUP which can emit multiple). Use ROLLUP if you need to emit multiple output records per group or ...
null
null
null
transformers
Rollup
ROLLUP
Executes a rollup transform.
Receives data through connected input port, groups incoming records by their key values and lets user transformation arbitrarily create new data records from the group by aggregating, composing or splitting. Sends these new data records through the connected output port(s) as specified by the transformation. Rollup fun...
[ { "type": "singlePort", "name": "0", "required": true } ]
[ { "type": "multiplePort", "required": true } ]
[ { "name": "groupKeyFields", "label": "Group key", "description": "Defines the key fields used for group key.", "modifiable": true, "nullable": true, "type": "key", "inputPortName": "0", "rightLabel": "Group key:" }, { "name": "groupAccumulatorMetadataId", "label": "Group ...
null
When to use: Custom group processing logic that AGGREGATE can't express — e.g. conditional aggregation, producing multiple output records per group, or accumulating complex state. Full CTL2 control via initGroup/updateGroup/finishGroup/transform functions. Constraints: Requires CTL2 code. Input must be sorted on groupK...
null
null
null
readers
JMSReader
JMS_READER
Receives JMS messages.
Receives JMS messages, converts them to data records and sends them to all connected output ports. Implements JmsMsg2DataRecord interface.
[]
[ { "type": "singlePort", "name": "0", "required": true } ]
[ { "name": "connection", "label": "JMS connection", "description": "Configures jms connection.", "modifiable": true, "nullable": false, "required": true, "type": "jmsConnection" }, { "name": "processorCode", "label": "Processor code", "description": "Configures processor c...
null
null
false
null
false
writers
JMSWriter
JMS_WRITER
Sends out JMS messages.
Receives data through connected input port, converts data records to JMS messages and sends them out. Implements DataRecord2JmsMsg interface.
[ { "type": "singlePort", "name": "0", "required": true } ]
[]
[ { "name": "connection", "label": "JMS connection", "description": "Configures jms connection.", "modifiable": true, "nullable": false, "required": true, "type": "jmsConnection" }, { "name": "processorCode", "label": "Processor code", "description": "Configures processor c...
null
null
false
null
false
joiners
LookupJoin
LOOKUP_JOIN
Joins incoming data flow with records from lookup table.
Receives data through connected input port (drivers), for each join key value contained in driver data records searches corresponding data records with the same key value in lookup table (slaves). Every pair of driver and slave records is sent to transformation class that joins them to form outgoing data flow. Resultin...
[ { "type": "singlePort", "name": "0", "required": true } ]
[ { "type": "singlePort", "name": "0", "required": true, "label": "joined records" }, { "type": "singlePort", "name": "1", "label": "skipped records" } ]
[ { "name": "joinKey", "label": "Join key", "description": "Defines the fields used to join data from multiple input sources.", "modifiable": true, "nullable": true, "required": true, "type": "lookupJoinKey" }, { "name": "leftOuterJoin", "label": "Left outer join", "descrip...
null
When to use: Join driver stream against a static lookup table already loaded into memory (defined in <Global>). Eliminates second input port. Best for small reference tables reused across multiple components in the same graph (lookup table loaded once, used many times). Constraints: Lookup table must be defined in grap...
null
null
null
others
LookupTableReaderWriter
LOOKUP_TABLE_READER_WRITER
Reads data from lookup table or write data to lookup table.
Receives data records through connected input port and writes them to specified lookup table (output port is not connected) or reads data records from lookup table and sends them out through all connected output ports (input port is not connected). Lookup table can also be updated before data records are read and sent ...
[ { "type": "singlePort", "name": "0" } ]
[ { "type": "multiplePort" } ]
[ { "name": "lookupTable", "label": "Lookup table", "description": "Configures lookup table.", "modifiable": true, "nullable": true, "required": true, "propagateMetadataOnUpdate": "true", "type": "lookupTableID" }, { "name": "freeLookupTable", "label": "Clear lookup table a...
null
null
null
null
null
readers
DataGenerator
DATA_GENERATOR
Generates data records.
Generates data records according to defined pattern and sends them to all connected output ports.
[]
[ { "type": "multiplePort", "required": true } ]
[ { "name": "generate", "label": "Generator", "description": "Configures generator.", "modifiable": true, "nullable": true, "primaryAttribute": true, "type": "transformGenerator", "input": "key" }, { "name": "generateURL", "label": "Generator URL", "description": "Speci...
null
null
true
null
false
others
SequenceChecker
SEQUENCE_CHECKER
Checks data sequence.
Checks the sort order of incoming data records. If all records are sorted properly for a specified sort key, the component continues, otherwise it fails (thus aborting the graph). Sort key is name or combination of names of field(s) of incoming records. Sort order is either Ascending (default) or Descending, but it is ...
[ { "type": "singlePort", "name": "0", "required": true } ]
[ { "type": "multiplePort", "required": false } ]
[ { "name": "sortOrder", "label": "Sort order", "description": "Specifies whether to sort data in ascending or descending order.", "modifiable": true, "nullable": true, "defaultValue": "Ascending", "type": "order" }, { "name": "sortKey", "label": "Sort key", "description": ...
"etlGraph"
null
null
true
null
readers
XMLXPathReader
XML_XPATH_READER
Reads data from XML files.
Reads data from XML files, converts elements matched by XPath defined in mapping along with their direct children to data records and sends them into different connected output ports as defined by mapping.
[ { "type": "singlePort", "name": "0", "required": false } ]
[ { "type": "multiplePort", "required": true, "metadataProvider": { "id": "treeReaderErrMetadataProvider" } } ]
[ { "name": "fileURL", "label": "File URL", "description": "Specifies the path or URL of the file to process. This is the primary attribute that determines which file will be processed by the component.", "modifiable": true, "nullable": true, "required": true, "primaryAttribute": true, ...
null
null
null
null
null
readers
XMLExtract
XML_EXTRACT
Reads data from XML files.
Reads data from XML files, converts elements matched by mapping along with their children to data records and sends them to different connected output ports as defined by mapping.
[ { "type": "singlePort", "name": "0", "required": false, "label": "input" } ]
[ { "type": "multiplePort", "required": true } ]
[ { "name": "sourceUri", "label": "File URL", "description": "Configures file url.", "modifiable": true, "nullable": true, "required": true, "openActionLabel": "Open input XML", "type": "file", "fields": "input", "dictionary": "input", "httpContext": "request" }, { ...
null
When to use: Large files (streaming SAX parser — low memory). Hierarchical extraction to multiple output ports using parentKey/generatedKey to propagate parent keys to child records. Best for XML/JSON→DB pipelines with parent-child relationships. Constraints: CloverDX Mappings XML syntax (not XPath). Less expressive th...
null
null
null
readers
JSONExtract
JSON_EXTRACT
Reads data from JSON files.
Reads data from JSON files, converts elements matched by mapping along with their children to data records and sends them to different connected output ports as defined by mapping.
[ { "type": "singlePort", "name": "0", "required": false, "label": "input" } ]
[ { "type": "multiplePort", "required": true } ]
[ { "name": "sourceUri", "label": "File URL", "description": "Configures file url.", "modifiable": true, "nullable": true, "required": true, "openActionLabel": "Open input JSON", "type": "file", "fields": "input", "dictionary": "input", "httpContext": "request" }, { ...
null
When to use: Large JSON files (streaming SAX parser — low memory footprint). Hierarchical extraction to multiple output ports using parentKey/generatedKey to propagate parent identifiers to child records. Best for JSON→DB pipelines with nested object/array relationships where the full document should not be loaded into...
null
null
null
readers
XMLReader
XML_READER
Reads data from XML.
Reads data from XML, converts properties matched by XPath defined in mapping along with their direct children to data records and sends them into different connected output ports as defined by mapping.
[ { "type": "singlePort", "name": "0", "required": false } ]
[ { "type": "multiplePort", "required": true, "metadataProvider": { "id": "treeReaderErrMetadataProvider" } } ]
[ { "name": "fileURL", "label": "File URL", "description": "Specifies the path or URL of the file to process. This is the primary attribute that determines which file will be processed by the component.", "modifiable": true, "nullable": true, "required": true, "type": "file", "fields":...
"etlGraph"
When to use: Smaller/medium files where entire document fits in memory (DOM).Complex or irregular structure better expressed with XPath. Constraints: Loads entire document into RAM — avoid for large (>>5M) files.
null
null
null
deprecated
SpeedLimiter
SPEED_LIMITER
Delays incoming records.
Receives data through connected input port, delays each incoming data record by defined value of milliseconds and copies all data records to all connected output ports.
[ { "type": "singlePort", "name": "0", "required": true } ]
[ { "type": "multiplePort", "required": true } ]
[ { "name": "delay", "label": "Delay (ms)", "description": "Configures delay (ms).", "modifiable": true, "nullable": true, "required": "!inputMapping", "type": "interval" }, { "name": "inputMapping", "label": "Input mapping", "description": "Configures input mapping.", ...
false
null
false
true
false
jobControl
Sleep
SLEEP
Delays incoming records.
Receives data through connected input port, delays each incoming data record by defined value of milliseconds and copies all data records to all connected output ports.
[ { "type": "singlePort", "name": "0", "required": false } ]
[ { "type": "multiplePort", "required": false } ]
[ { "name": "delay", "label": "Delay (ms)", "description": "Configures delay (ms).", "modifiable": true, "nullable": true, "required": "!inputMapping", "type": "interval" }, { "name": "inputMapping", "label": "Input mapping", "description": "Configures input mapping.", ...
null
null
null
true
null
deprecated
JavaExecute
JAVA_EXECUTE
Executes Java transformations.
Executes Java commands using runnable transformation.
[]
[]
[ { "name": "runnable", "label": "Runnable", "description": "Configures runnable.", "modifiable": true, "nullable": true, "required": "!runnableClass,!runnableURL", "type": "multilineEditableJava" }, { "name": "runnableURL", "label": "Runnable URL", "description": "Specifie...
false
null
false
null
false
readers
CustomJavaReader
GENERIC_READER
Generic component that allows you to create custom reader.
Runs the specified custom algorithm implemented in Java. Provides an easy way to read, process and write records from and to any number of ports. This is a CustomJavaComponent customized for implementing a Reader.
[ { "type": "multiplePort", "required": false } ]
[ { "type": "multiplePort", "required": false } ]
[ { "name": "genericTransform", "label": "Algorithm", "description": "Configures algorithm.", "modifiable": true, "nullable": true, "required": "!genericTransformClass,!genericTransformURL", "type": "genericComponentMultiline", "genericType": "reader" }, { "name": "genericTrans...
null
null
null
null
false
End of preview. Expand in Data Studio

Not a LLM training dataset but several LLM prompts (and general metadata files) focused on generating and understanding CloverDX transformation graphs artefacts. Can be used as part of LLM system/user prompts.

Downloads last month
35