hi,all
   my issue is everyday I will receive some json datafile , I want to convert 
them to parquet file and save to hdfs,
   the floder will like this:
   /my_table_base_floder 
          /my_table_base_floder/day_2
          /my_table_base_floder/day_3
          ....
 
where the parquet files of "day_1" was store in  /my_table_base_floder 
then I run : 
sqlContext.createExternalTable("tpc1.customer","hdfs://master1:9000/my_table_base_floder","parquet")
but when I save parquet file to subdir ,for example :  
/my_table_base_floder/day_2  and refresh the metadata.
spark doesn't recognize the data in subdir. How I can do it ?


2016-10-20


lk_spark 

Reply via email to