DeduplicateRecord 2.3.0¶
BUNDLE¶
org.apache.nifi | nifi-standard-nar
DESCRIPTION¶
This processor de-duplicates individual records within a record set. It can operate on a per-file basis using an in-memory hashset or bloom filter. When configured with a distributed map cache, it de-duplicates records across multiple files.
INPUT REQUIREMENT¶
REQUIRED
Supports Sensitive Dynamic Properties¶
false
PROPERTIES¶
Property |
Description |
---|---|
bloom-filter-certainty |
The desired false positive probability when using the BloomFilter type. Using a value of .05 for example, guarantees a five-percent probability that the result is a false positive. The closer to 1 this value is set, the more precise the result at the expense of more storage space utilization. |
cache-identifier |
An optional expression language field that overrides the record’s computed cache key. This field has an additional attribute available: ${record.hash.value}, which contains the cache key derived from dynamic properties (if set) or record fields. |
deduplication-strategy |
The strategy to use for detecting and routing duplicate records. The option for detecting duplicates across a single FlowFile operates in-memory, whereas detection spanning multiple FlowFiles utilises a distributed map cache. |
distributed-map-cache |
This property is required when the deduplication strategy is set to ‘multiple files.’ The map cache will for each record, atomically check whether the cache key exists and if not, set it. |
filter-capacity-hint |
An estimation of the total number of unique records to be processed. The more accurate this number is will lead to fewer false negatives on a BloomFilter. |
filter-type |
The filter used to determine whether a record has been seen before based on the matching RecordPath criteria. If hash set is selected, a Java HashSet object will be used to deduplicate all encountered records. If the bloom filter option is selected, a bloom filter will be used. The bloom filter option is less memory intensive, but has a chance of having false positives. |
include-zero-record-flowfiles |
If a FlowFile sent to either the duplicate or non-duplicate relationships contains no records, a value of |
put-cache-identifier |
For each record, check whether the cache identifier exists in the distributed map cache. If it doesn’t exist and this property is true, put the identifier to the cache. |
record-hashing-algorithm |
The algorithm used to hash the cache key. |
record-reader |
Specifies the Controller Service to use for reading incoming data |
record-writer |
Specifies the Controller Service to use for writing out the records |
RELATIONSHIPS¶
NAME |
DESCRIPTION |
---|---|
original |
The original input FlowFile is sent to this relationship unless a fatal error occurs. |
failure |
If unable to communicate with the cache, the FlowFile will be penalized and routed to this relationship |
duplicate |
Records detected as duplicates are routed to this relationship. |
non-duplicate |
Records not found in the cache are routed to this relationship. |
WRITES ATTRIBUTES¶
NAME |
DESCRIPTION |
---|---|
record.count |
Number of records written to the destination FlowFile. |