Skip to content
background-image background-image

Node.js processor

[ | version 3.1]


The connector is permitted for use of the platform background agent.

Connector processing type: Both (Row by row & Bulk), Default type: Row by row!

Debug script enabled.

The NodeJS connector allows you to run JavaScript scripts using the Node.js runtime (version 20.4). NodeJS connector might look very similar to JS mapper but is much more powerfull. Although NodeJS ecosystem is very large, we provide only limited yet powerfull set of imported packages. For security reasons, your JavaScript code can only access limited set of imported types or packages and cannot use dynamic imports or require statement.

Provided packages

Above packages may provide more types, but Integray exposes only limited set of types. All types are covered within Node.js processor examples.

In addition to above packages and types, runtime exposes following APIs

/* Connector logger API */
log { 
  trace: (message) => void,
  debug: (message) => void,
  info: (message) => void,
  warn: (message) => void,
  /* Calling this method will force the connector to mark statement execution as failed */
  error: (message) => void,

/* wait specified amount of time */
sleep: (timeout) => Promise<void>

// for example sleep for 1 second
await sleep(1000);


Node JS statement


JavaScript statement to be executed using the Node.js service.

The output of the connector expects an array of objects that must match the output schema in structure. The defined Statement represents the body of the async function whose output is parsed and returned as the output of the connector itself. The last line of the Statement must therefore be:

return <array_of_objects>;


return [
    { "hello": "Alice" }


As the statement is async you must ensure that all Promises are awaited. Non-awaited code may not complete before your statement code.

Data checkpoint column

The data checkpoint column is a column (field), from which the platform takes the last row value after each executed task run and stores it as a Data checkpoint. The data checkpoint value can be used in the JS statements to control, which data should be processed in the next run. You can refer to the value using the predefined variable DataCheckpoint. Example of use: processing data in cycles, where every cycle processes only a subset of the entire set due to the total size. If you use e.g. record ID as a data checkpoint column, the platform will store after each cycle the last processed ID from the data subset processed by the task run. If your statement is written in a way that will evaluate the value in data checkpoint against the IDs of the records in the data set, you can ensure this way, that only not processed records will be considered in the next task run.

Input & Output Schema


Data schema is optional

The connector does not expect a specific schema. The required data structure can be achieved by correct configuration. Although the selected connector doesn't require a schema generally, the individual integration task step may need to match the output data structure of the preceding task step and use a data schema selected from the repository or create a new input schema.


Data schema is mandatory

The connector requires mandatory input or output data schema, which must be selected by the user from the existing data schema repository or a new one must be created. The connector will fail without structured data.

Release notes


  • Fixed processing sensitive errors


  • First release