@SupportsBatching @SeeAlso({PutS3Object.class, FetchS3Object.class, ListS3.class}) @Tags({"Amazon", "S3", "AWS", "Archive", "Delete"})
@SupportsBatching @SeeAlso({GetSQS.class, PutSQS.class}) @InputRequirement(Requirement.INPUT_REQUIRED)
@SupportsBatching @Tags({"google cloud", "gcs", "google", "storage", "delete"}) @CapabilityDescription("Deletes objects from a Google Cloud Bucket. " +
@SupportsBatching @SeeAlso({GetSQS.class, PutSQS.class}) @InputRequirement(Requirement.INPUT_REQUIRED)
@EventDriven @SupportsBatching @InputRequirement(InputRequirement.Requirement.INPUT_REQUIRED) @Tags({"hadoop", "hbase"})
@SupportsBatching @InputRequirement(Requirement.INPUT_REQUIRED) @Tags({"remote", "copy", "egress", "put", "sftp", "archive", "files"})
@Deprecated @DeprecationNotice(classNames = {"org.apache.nifi.processors.standard.CryptographicHashContent"}, reason = "This processor is deprecated and may be removed in future releases.") @EventDriven @SupportsBatching @InputRequirement(Requirement.INPUT_REQUIRED) @Tags({"hash", "content", "MD5", "SHA-1", "SHA-256"})
@EventDriven @SupportsBatching @Tags({"test", "load", "duplicate"}) @InputRequirement(Requirement.INPUT_REQUIRED)
@EventDriven @SideEffectFree @SupportsBatching @Tags({"encode", "base64"}) @CapabilityDescription("Encodes or decodes content to and from base64")
@SupportsBatching @SeeAlso({ GetSQS.class, DeleteSQS.class }) @InputRequirement(Requirement.INPUT_REQUIRED)
@EventDriven @SupportsBatching @InputRequirement(Requirement.INPUT_REQUIRED) @SideEffectFree
@SupportsBatching @InputRequirement(InputRequirement.Requirement.INPUT_REQUIRED) @Tags({"parquet", "hadoop", "HDFS", "get", "ingest", "fetch", "source", "record"}) @CapabilityDescription("Reads from a given Parquet file and writes records to the content of the flow file using " + "the selected record writer. The original Parquet file will remain unchanged, and the content of the flow file " + "will be replaced with records of the selected type. This processor can be used with ListHDFS or ListFile to obtain " + "a listing of files to fetch.") @WritesAttributes({ @WritesAttribute(attribute="fetch.failure.reason", description="When a FlowFile is routed to 'failure', this attribute is added " + "indicating why the file could not be fetched from the given filesystem."), @WritesAttribute(attribute = "record.count", description = "The number of records in the resulting flow file") }) @SeeAlso({PutParquet.class}) @Restricted(restrictions = { @Restriction( requiredPermission = RequiredPermission.READ_FILESYSTEM, explanation = "Provides operator the ability to retrieve any file that NiFi has access to in HDFS or the local filesystem.") }) public class FetchParquet extends AbstractFetchHDFSRecord { @Override public HDFSRecordReader createHDFSRecordReader(final ProcessContext context, final FlowFile flowFile, final Configuration conf, final Path path) throws IOException { final ParquetReader.Builder<GenericRecord> readerBuilder = AvroParquetReader.<GenericRecord>builder(path).withConf(conf); return new AvroParquetHDFSRecordReader(readerBuilder.build()); } }
@SupportsBatching @InputRequirement(InputRequirement.Requirement.INPUT_FORBIDDEN) @Tags({"listen", "tcp", "tls", "ssl"})
@SupportsBatching @Tags({"ingest", "udp", "listen", "source"}) @InputRequirement(Requirement.INPUT_FORBIDDEN)
@SupportsBatching @Tags({ "Ignite", "get", "read", "cache", "key" }) @InputRequirement(Requirement.INPUT_REQUIRED)
@EventDriven @SideEffectFree @SupportsBatching @InputRequirement(InputRequirement.Requirement.INPUT_REQUIRED) @Tags({"hashing", "fuzzy-hashing", "cyber-security"})
@EventDriven @SideEffectFree @SupportsBatching @Tags({"ISP", "enrich", "ip", "maxmind"}) @InputRequirement(Requirement.INPUT_REQUIRED)
@EventDriven @SideEffectFree @SupportsBatching @InputRequirement(Requirement.INPUT_REQUIRED) @Tags({"logs", "syslog", "attributes", "system", "event", "message"})
@EventDriven @SideEffectFree @SupportsBatching @Tags({"geo", "enrich", "ip", "maxmind"}) @InputRequirement(Requirement.INPUT_REQUIRED)
@SupportsBatching @SeeAlso({GetDynamoDB.class, PutDynamoDB.class}) @InputRequirement(Requirement.INPUT_REQUIRED)