Skip to main content
Pentaho Documentation

Unique Rows (HashSet)

Parent article

The Unique Rows (HashSet) step removes duplicate rows and filters only the unique rows as input data for the step.

This step differs from the Unique Rows transformation step by keeping track of the duplicate rows in memory and does not require a sorted input to process duplicate rows.

NoteBecause of memory allocation issues, this step is intended for non-client machines. The required amount of memory and hardware will vary based on the size of the data you are processing. Alternatively, you can use this step with the Spark processing engine to help overcome memory constraint issues. For more information, see Set up the Adaptive Execution Layer (AEL) .

General

The Unique Rows (HashSet) step requires you to define parameters for your output and how you want to sort and track duplicate rows. Unique Row Hashset step

Enter the following information for the Step name field:

  • Step name: Specify the unique name of the transformation on the canvas. You can customize the name or leave it as the default

Settings

The Unique Rows (HashSet) step requires definitions for the following options and parameters:

OptionDescription
Compare using stored row valuesSelect this option to store values for the selected fields in memory for every record. Storing row values requires more memory, but it prevents possible false positives if there are hash collisions.
Redirect duplicate rowSelect this option to process duplicate rows as an error and redirect them to the error stream of the step. If you do not select this option, the duplicate rows are deleted.
Error descriptionSpecify the error handling description that displays when the step detects duplicate rows. This description is only available when Redirect duplicate row is selected.
Fields to compare table

Specify the field names for which you want to find unique values.

-OR-

Select Get to insert all the fields from the input stream.