[ 
https://issues.apache.org/jira/browse/NIFI-3724?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15991012#comment-15991012
 ] 

ASF GitHub Bot commented on NIFI-3724:
--------------------------------------

Github user alopresto commented on a diff in the pull request:

    https://github.com/apache/nifi/pull/1712#discussion_r114147981
  
    --- Diff: 
nifi-nar-bundles/nifi-extension-utils/nifi-hadoop-utils/src/main/java/org/apache/nifi/processors/hadoop/AbstractHadoopProcessor.java
 ---
    @@ -67,40 +61,14 @@
      */
     @RequiresInstanceClassLoading(cloneAncestorResources = true)
     public abstract class AbstractHadoopProcessor extends AbstractProcessor {
    -    /**
    -     * Compression Type Enum
    -     */
    -    public enum CompressionType {
    -        NONE,
    -        DEFAULT,
    -        BZIP,
    -        GZIP,
    -        LZ4,
    -        SNAPPY,
    -        AUTOMATIC;
    -
    -        @Override
    -        public String toString() {
    -            switch (this) {
    -                case NONE: return "NONE";
    -                case DEFAULT: return DefaultCodec.class.getName();
    -                case BZIP: return BZip2Codec.class.getName();
    -                case GZIP: return GzipCodec.class.getName();
    -                case LZ4: return Lz4Codec.class.getName();
    -                case SNAPPY: return SnappyCodec.class.getName();
    -                case AUTOMATIC: return "Automatically Detected";
    -            }
    -            return null;
    -        }
    -    }
     
         // properties
         public static final PropertyDescriptor HADOOP_CONFIGURATION_RESOURCES 
= new PropertyDescriptor.Builder()
                 .name("Hadoop Configuration Resources")
                 .description("A file or comma separated list of files which 
contains the Hadoop file system configuration. Without this, Hadoop "
                         + "will search the classpath for a 'core-site.xml' and 
'hdfs-site.xml' file or will revert to a default configuration.")
                 .required(false)
    -            .addValidator(createMultipleFilesExistValidator())
    +            .addValidator(HadoopValidators.MULTIPLE_FILE_EXISTS_VALIDATOR)
    --- End diff --
    
    Minor comment -- until I read the source code for this, my interpretation 
was that this validator ensured that *multiple files existed* -- i.e. one file 
provided would fail. Perhaps we can rename this 
`ONE_OR_MORE_FILES_EXIST_VALIDATOR`? Not a giant issue but potentially 
confusing. 


> Add Put/Fetch Parquet Processors
> --------------------------------
>
>                 Key: NIFI-3724
>                 URL: https://issues.apache.org/jira/browse/NIFI-3724
>             Project: Apache NiFi
>          Issue Type: Improvement
>            Reporter: Bryan Bende
>            Assignee: Bryan Bende
>            Priority: Minor
>             Fix For: 1.2.0
>
>
> Now that we have the record reader/writer services currently in master, it 
> would be nice to have reader and writers for Parquet. Since Parquet's API is 
> based on the Hadoop Path object, and not InputStreams/OutputStreams, we can't 
> really implement direct conversions to and from Parquet in the middle of a 
> flow, but we can we can perform the conversion by taking any record format 
> and writing to a Path as Parquet, or reading Parquet from a Path and writing 
> it out as another record format.
> We should add a PutParquet that uses a record reader and writes records to a 
> Path as Parquet, and a FetchParquet that reads Parquet from a path and writes 
> out records to a flow file using a record writer.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

Reply via email to