Your table is missing a "PARTITIONED BY " section. Spark 2.x save the partition information in the TBLPROPERTIES section.
On Sun, Feb 11, 2018 at 10:41 AM, Deepak Sharma <deepakmc...@gmail.com> wrote: > I can see its trying to read the parquet and failing while decompressing > using snappy: > parquet.hadoop.ParquetRecordReader.nextKeyValue(ParquetRecor > dReader.java:201) > > So the table looks good but this needs to be fixed before you can query > the data in hive. > > Thanks > Deepak > > On Sun, Feb 11, 2018 at 1:45 PM, ☼ R Nair (रविशंकर नायर) < > ravishankar.n...@gmail.com> wrote: > >> When I do that , and then do a select, full of errors. I think Hive table >> to read. >> >> select * from mine; >> OK >> SLF4J: Failed to load class "org.slf4j.impl.StaticLoggerBinder". >> SLF4J: Defaulting to no-operation (NOP) logger implementation >> SLF4J: See http://www.slf4j.org/codes.html#StaticLoggerBinder for >> further details. >> java.lang.reflect.InvocationTargetException >> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >> at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAcce >> ssorImpl.java:62) >> at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMe >> thodAccessorImpl.java:43) >> at java.lang.reflect.Method.invoke(Method.java:498) >> at org.xerial.snappy.SnappyLoader.loadNativeLibrary(SnappyLoade >> r.java:317) >> at org.xerial.snappy.SnappyLoader.load(SnappyLoader.java:219) >> at org.xerial.snappy.Snappy.<clinit>(Snappy.java:44) >> at parquet.hadoop.codec.SnappyDecompressor.decompress(SnappyDec >> ompressor.java:62) >> at parquet.hadoop.codec.NonBlockedDecompressorStream.read(NonBl >> ockedDecompressorStream.java:51) >> at java.io.DataInputStream.readFully(DataInputStream.java:195) >> at java.io.DataInputStream.readFully(DataInputStream.java:169) >> at parquet.bytes.BytesInput$StreamBytesInput.toByteArray(BytesI >> nput.java:204) >> at parquet.column.impl.ColumnReaderImpl.readPageV1(ColumnReader >> Impl.java:557) >> at parquet.column.impl.ColumnReaderImpl.access$300(ColumnReader >> Impl.java:57) >> at parquet.column.impl.ColumnReaderImpl$3.visit(ColumnReaderImp >> l.java:516) >> at parquet.column.impl.ColumnReaderImpl$3.visit(ColumnReaderImp >> l.java:513) >> at parquet.column.page.DataPageV1.accept(DataPageV1.java:96) >> at parquet.column.impl.ColumnReaderImpl.readPage(ColumnReaderIm >> pl.java:513) >> at parquet.column.impl.ColumnReaderImpl.checkRead(ColumnReaderI >> mpl.java:505) >> at parquet.column.impl.ColumnReaderImpl.consume(ColumnReaderImp >> l.java:607) >> at parquet.column.impl.ColumnReaderImpl.<init>(ColumnReaderImpl.java:351) >> at parquet.column.impl.ColumnReadStoreImpl.newMemColumnReader(C >> olumnReadStoreImpl.java:66) >> at parquet.column.impl.ColumnReadStoreImpl.getColumnReader(Colu >> mnReadStoreImpl.java:61) >> at parquet.io.RecordReaderImplementation.<init>(RecordReaderImp >> lementation.java:270) >> at parquet.io.MessageColumnIO$1.visit(MessageColumnIO.java:134) >> at parquet.io.MessageColumnIO$1.visit(MessageColumnIO.java:99) >> at parquet.filter2.compat.FilterCompat$NoOpFilter.accept( >> FilterCompat.java:154) >> at parquet.io.MessageColumnIO.getRecordReader(MessageColumnIO.java:99) >> at parquet.hadoop.InternalParquetRecordReader.checkRead(Interna >> lParquetRecordReader.java:137) >> at parquet.hadoop.InternalParquetRecordReader.nextKeyValue(Inte >> rnalParquetRecordReader.java:208) >> at parquet.hadoop.ParquetRecordReader.nextKeyValue(ParquetRecor >> dReader.java:201) >> at org.apache.hadoop.hive.ql.io.parquet.read.ParquetRecordReade >> rWrapper.<init>(ParquetRecordReaderWrapper.java:122) >> at org.apache.hadoop.hive.ql.io.parquet.read.ParquetRecordReade >> rWrapper.<init>(ParquetRecordReaderWrapper.java:85) >> at org.apache.hadoop.hive.ql.io.parquet.MapredParquetInputForma >> t.getRecordReader(MapredParquetInputFormat.java:72) >> at org.apache.hadoop.hive.ql.exec.FetchOperator$FetchInputForma >> tSplit.getRecordReader(FetchOperator.java:673) >> at org.apache.hadoop.hive.ql.exec.FetchOperator.getRecordReader >> (FetchOperator.java:323) >> at org.apache.hadoop.hive.ql.exec.FetchOperator.getNextRow(Fetc >> hOperator.java:445) >> at org.apache.hadoop.hive.ql.exec.FetchOperator.pushRow(FetchOp >> erator.java:414) >> at org.apache.hadoop.hive.ql.exec.FetchTask.fetch(FetchTask.java:140) >> at org.apache.hadoop.hive.ql.Driver.getResults(Driver.java:1670) >> at org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriv >> er.java:233) >> at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:165) >> at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:376) >> at org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:736) >> at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:681) >> at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:621) >> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >> at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAcce >> ssorImpl.java:62) >> at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMe >> thodAccessorImpl.java:43) >> at java.lang.reflect.Method.invoke(Method.java:498) >> at org.apache.hadoop.util.RunJar.run(RunJar.java:221) >> at org.apache.hadoop.util.RunJar.main(RunJar.java:136) >> Caused by: java.lang.UnsatisfiedLinkError: no snappyjava in >> java.library.path >> at java.lang.ClassLoader.loadLibrary(ClassLoader.java:1867) >> at java.lang.Runtime.loadLibrary0(Runtime.java:870) >> at java.lang.System.loadLibrary(System.java:1122) >> at org.xerial.snappy.SnappyNativeLoader.loadLibrary(SnappyNativ >> eLoader.java:52) >> ... 52 more >> Exception in thread "main" org.xerial.snappy.SnappyError: >> [FAILED_TO_LOAD_NATIVE_LIBRARY] null >> at org.xerial.snappy.SnappyLoader.load(SnappyLoader.java:229) >> at org.xerial.snappy.Snappy.<clinit>(Snappy.java:44) >> at parquet.hadoop.codec.SnappyDecompressor.decompress(SnappyDec >> ompressor.java:62) >> at parquet.hadoop.codec.NonBlockedDecompressorStream.read(NonBl >> ockedDecompressorStream.java:51) >> at java.io.DataInputStream.readFully(DataInputStream.java:195) >> at java.io.DataInputStream.readFully(DataInputStream.java:169) >> at parquet.bytes.BytesInput$StreamBytesInput.toByteArray(BytesI >> nput.java:204) >> at parquet.column.impl.ColumnReaderImpl.readPageV1(ColumnReader >> Impl.java:557) >> at parquet.column.impl.ColumnReaderImpl.access$300(ColumnReader >> Impl.java:57) >> at parquet.column.impl.ColumnReaderImpl$3.visit(ColumnReaderImp >> l.java:516) >> at parquet.column.impl.ColumnReaderImpl$3.visit(ColumnReaderImp >> l.java:513) >> at parquet.column.page.DataPageV1.accept(DataPageV1.java:96) >> at parquet.column.impl.ColumnReaderImpl.readPage(ColumnReaderIm >> pl.java:513) >> at parquet.column.impl.ColumnReaderImpl.checkRead(ColumnReaderI >> mpl.java:505) >> at parquet.column.impl.ColumnReaderImpl.consume(ColumnReaderImp >> l.java:607) >> at parquet.column.impl.ColumnReaderImpl.<init>(ColumnReaderImpl.java:351) >> at parquet.column.impl.ColumnReadStoreImpl.newMemColumnReader(C >> olumnReadStoreImpl.java:66) >> at parquet.column.impl.ColumnReadStoreImpl.getColumnReader(Colu >> mnReadStoreImpl.java:61) >> at parquet.io.RecordReaderImplementation.<init>(RecordReaderImp >> lementation.java:270) >> at parquet.io.MessageColumnIO$1.visit(MessageColumnIO.java:134) >> at parquet.io.MessageColumnIO$1.visit(MessageColumnIO.java:99) >> at parquet.filter2.compat.FilterCompat$NoOpFilter.accept( >> FilterCompat.java:154) >> at parquet.io.MessageColumnIO.getRecordReader(MessageColumnIO.java:99) >> at parquet.hadoop.InternalParquetRecordReader.checkRead(Interna >> lParquetRecordReader.java:137) >> at parquet.hadoop.InternalParquetRecordReader.nextKeyValue(Inte >> rnalParquetRecordReader.java:208) >> at parquet.hadoop.ParquetRecordReader.nextKeyValue(ParquetRecor >> dReader.java:201) >> at org.apache.hadoop.hive.ql.io.parquet.read.ParquetRecordReade >> rWrapper.<init>(ParquetRecordReaderWrapper.java:122) >> at org.apache.hadoop.hive.ql.io.parquet.read.ParquetRecordReade >> rWrapper.<init>(ParquetRecordReaderWrapper.java:85) >> at org.apache.hadoop.hive.ql.io.parquet.MapredParquetInputForma >> t.getRecordReader(MapredParquetInputFormat.java:72) >> at org.apache.hadoop.hive.ql.exec.FetchOperator$FetchInputForma >> tSplit.getRecordReader(FetchOperator.java:673) >> at org.apache.hadoop.hive.ql.exec.FetchOperator.getRecordReader >> (FetchOperator.java:323) >> at org.apache.hadoop.hive.ql.exec.FetchOperator.getNextRow(Fetc >> hOperator.java:445) >> at org.apache.hadoop.hive.ql.exec.FetchOperator.pushRow(FetchOp >> erator.java:414) >> at org.apache.hadoop.hive.ql.exec.FetchTask.fetch(FetchTask.java:140) >> at org.apache.hadoop.hive.ql.Driver.getResults(Driver.java:1670) >> at org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriv >> er.java:233) >> at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:165) >> at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:376) >> at org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:736) >> at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:681) >> at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:621) >> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >> at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAcce >> ssorImpl.java:62) >> at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMe >> thodAccessorImpl.java:43) >> at java.lang.reflect.Method.invoke(Method.java:498) >> at org.apache.hadoop.util.RunJar.run(RunJar.java:221) >> at org.apache.hadoop.util.RunJar.main(RunJar.java:136) >> Feb 11, 2018 3:14:06 AM WARNING: parquet.hadoop.ParquetRecordReader: Can >> not initialize counter due to context is not a instance of >> TaskInputOutputContext, but is org.apache.hadoop.mapreduce.ta >> sk.TaskAttemptContextImpl >> Feb 11, 2018 3:14:06 AM INFO: parquet.hadoop.InternalParquetRecordReader: >> RecordReader initialized will read a total of 36635 records. >> Feb 11, 2018 3:14:06 AM INFO: parquet.hadoop.InternalParquetRecordReader: >> at row 0. reading next block >> Feb 11, 2018 3:14:06 AM INFO: parquet.hadoop.InternalParquetRecordReader: >> block read in memory in 27 ms. row count = 36635 >> >> >> On Sun, Feb 11, 2018 at 3:10 AM, Deepak Sharma <deepakmc...@gmail.com> >> wrote: >> >>> There was a typo: >>> Instead of : >>> alter table mine set locations "hdfs://localhost:8020/user/hi >>> ve/warehouse/mine"; >>> >>> Use : >>> alter table mine set location "hdfs://localhost:8020/user/hi >>> ve/warehouse/mine"; >>> >>> On Sun, Feb 11, 2018 at 1:38 PM, Deepak Sharma <deepakmc...@gmail.com> >>> wrote: >>> >>>> Try this in hive: >>>> alter table mine set locations "hdfs://localhost:8020/user/hi >>>> ve/warehouse/mine"; >>>> >>>> Thanks >>>> Deepak >>>> >>>> On Sun, Feb 11, 2018 at 1:24 PM, ☼ R Nair (रविशंकर नायर) < >>>> ravishankar.n...@gmail.com> wrote: >>>> >>>>> Hi, >>>>> Here you go: >>>>> >>>>> hive> show create table mine; >>>>> OK >>>>> CREATE TABLE `mine`( >>>>> `policyid` int, >>>>> `statecode` string, >>>>> `socialid` string, >>>>> `county` string, >>>>> `eq_site_limit` decimal(10,2), >>>>> `hu_site_limit` decimal(10,2), >>>>> `fl_site_limit` decimal(10,2), >>>>> `fr_site_limit` decimal(10,2), >>>>> `tiv_2014` decimal(10,2), >>>>> `tiv_2015` decimal(10,2), >>>>> `eq_site_deductible` int, >>>>> `hu_site_deductible` int, >>>>> `fl_site_deductible` int, >>>>> `fr_site_deductible` int, >>>>> `latitude` decimal(6,6), >>>>> `longitude` decimal(6,6), >>>>> `line` string, >>>>> `construction` string, >>>>> `point_granularity` int) >>>>> ROW FORMAT SERDE >>>>> 'org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe' >>>>> WITH SERDEPROPERTIES ( >>>>> 'path'='hdfs://localhost:8020/user/hive/warehouse/mine') >>>>> STORED AS INPUTFORMAT >>>>> 'org.apache.hadoop.hive.ql.io.parquet.MapredParquetInputFormat' >>>>> OUTPUTFORMAT >>>>> 'org.apache.hadoop.hive.ql.io.parquet.MapredParquetOutputFormat' >>>>> LOCATION >>>>> 'file:/Users/ravishankarnair/spark-warehouse/mine' >>>>> TBLPROPERTIES ( >>>>> 'spark.sql.sources.provider'='parquet', >>>>> 'spark.sql.sources.schema.numParts'='1', >>>>> 'spark.sql.sources.schema.part.0'='{\"type\":\"struct\",\"fi >>>>> elds\":[{\"name\":\"policyid\",\"type\":\"integer\",\"nullab >>>>> le\":true,\"metadata\":{\"name\":\"policyid\",\"scale\":0}}, >>>>> {\"name\":\"statecode\",\"type\":\"string\",\"nullable\":tru >>>>> e,\"metadata\":{\"name\":\"statecode\",\"scale\":0}},{\"name >>>>> \":\"Socialid\",\"type\":\"string\",\"nullable\":true,\"meta >>>>> data\":{\"name\":\"Socialid\",\"scale\":0}},{\"name\":\"coun >>>>> ty\",\"type\":\"string\",\"nullable\":true,\"metadata\":{ >>>>> \"name\":\"county\",\"scale\":0}},{\"name\":\"eq_site_limit\ >>>>> ",\"type\":\"decimal(10,2)\",\"nullable\":true,\"metadata\": >>>>> {\"name\":\"eq_site_limit\",\"scale\":2}},{\"name\":\"hu_ >>>>> site_limit\",\"type\":\"decimal(10,2)\",\"nullable\": >>>>> true,\"metadata\":{\"name\":\"hu_site_limit\",\"scale\":2}}, >>>>> {\"name\":\"fl_site_limit\",\"type\":\"decimal(10,2)\",\" >>>>> nullable\":true,\"metadata\":{\"name\":\"fl_site_limit\",\" >>>>> scale\":2}},{\"name\":\"fr_site_limit\",\"type\":\" >>>>> decimal(10,2)\",\"nullable\":true,\"metadata\":{\"name\":\" >>>>> fr_site_limit\",\"scale\":2}},{\"name\":\"tiv_2014\",\"type\ >>>>> ":\"decimal(10,2)\",\"nullable\":true,\"metadata\":{ >>>>> \"name\":\"tiv_2014\",\"scale\":2}},{\"name\":\"tiv_2015\",\ >>>>> "type\":\"decimal(10,2)\",\"nullable\":true,\"metadata\":{ >>>>> \"name\":\"tiv_2015\",\"scale\":2}},{\"name\":\"eq_site_ >>>>> deductible\",\"type\":\"integer\",\"nullable\":true,\" >>>>> metadata\":{\"name\":\"eq_site_deductible\",\"scale\":0} >>>>> },{\"name\":\"hu_site_deductible\",\"type\":\"intege >>>>> r\",\"nullable\":true,\"metadata\":{\"name\":\"hu_site_ >>>>> deductible\",\"scale\":0}},{\"name\":\"fl_site_deductible\", >>>>> \"type\":\"integer\",\"nullable\":true,\"metadata\":{\"name\ >>>>> ":\"fl_site_deductible\",\"scale\":0}},{\"name\":\"fr_ >>>>> site_deductible\",\"type\":\"integer\",\"nullable\":true,\" >>>>> metadata\":{\"name\":\"fr_site_deductible\",\"scale\":0}},{\ >>>>> "name\":\"latitude\",\"type\":\"decimal(6,6)\",\"nullable\": >>>>> true,\"metadata\":{\"name\":\"latitude\",\"scale\":6}},{\" >>>>> name\":\"longitude\",\"type\":\"decimal(6,6)\",\"nullable\": >>>>> true,\"metadata\":{\"name\":\"longitude\",\"scale\":6}},{\" >>>>> name\":\"line\",\"type\":\"string\",\"nullable\":true,\" >>>>> metadata\":{\"name\":\"line\",\"scale\":0}},{\"name\":\" >>>>> construction\",\"type\":\"string\",\"nullable\":true,\" >>>>> metadata\":{\"name\":\"construction\",\"scale\":0}},{ >>>>> \"name\":\"point_granularity\",\"type\":\"integer\",\"nullab >>>>> le\":true,\"metadata\":{\"name\":\"point_granularity\",\ >>>>> "scale\":0}}]}', >>>>> 'transient_lastDdlTime'='1518335598') >>>>> Time taken: 0.13 seconds, Fetched: 35 row(s) >>>>> >>>>> On Sun, Feb 11, 2018 at 2:36 AM, Shmuel Blitz < >>>>> shmuel.bl...@similarweb.com> wrote: >>>>> >>>>>> Please run the following command, and paste the result: >>>>>> SHOW CREATE TABLE <<TABLE-NAME>> >>>>>> >>>>>> On Sun, Feb 11, 2018 at 7:56 AM, ☼ R Nair (रविशंकर नायर) < >>>>>> ravishankar.n...@gmail.com> wrote: >>>>>> >>>>>>> No, No luck. >>>>>>> >>>>>>> Thanks >>>>>>> >>>>>>> On Sun, Feb 11, 2018 at 12:48 AM, Deepak Sharma < >>>>>>> deepakmc...@gmail.com> wrote: >>>>>>> >>>>>>>> In hive cli: >>>>>>>> msck repair table 《table_name》; >>>>>>>> >>>>>>>> Thanks >>>>>>>> Deepak >>>>>>>> >>>>>>>> On Feb 11, 2018 11:14, "☼ R Nair (रविशंकर नायर)" < >>>>>>>> ravishankar.n...@gmail.com> wrote: >>>>>>>> >>>>>>>>> NO, can you pease explain the command ? Let me try now. >>>>>>>>> >>>>>>>>> Best, >>>>>>>>> >>>>>>>>> On Sun, Feb 11, 2018 at 12:40 AM, Deepak Sharma < >>>>>>>>> deepakmc...@gmail.com> wrote: >>>>>>>>> >>>>>>>>>> I am not sure about the exact issue bjt i see you are partioning >>>>>>>>>> while writing from spark. >>>>>>>>>> Did you tried msck repair on the table before reading it in hive ? >>>>>>>>>> >>>>>>>>>> Thanks >>>>>>>>>> Deepak >>>>>>>>>> >>>>>>>>>> On Feb 11, 2018 11:06, "☼ R Nair (रविशंकर नायर)" < >>>>>>>>>> ravishankar.n...@gmail.com> wrote: >>>>>>>>>> >>>>>>>>>>> All, >>>>>>>>>>> >>>>>>>>>>> Thanks for the inputs. Again I am not successful. I think, we >>>>>>>>>>> need to resolve this, as this is a very common requirement. >>>>>>>>>>> >>>>>>>>>>> Please go through my complete code: >>>>>>>>>>> >>>>>>>>>>> STEP 1: Started Spark shell as spark-shell --master yarn >>>>>>>>>>> >>>>>>>>>>> STEP 2: Flowing code is being given as inout to shark shell >>>>>>>>>>> >>>>>>>>>>> import org.apache.spark.sql.Row >>>>>>>>>>> import org.apache.spark.sql.SparkSession >>>>>>>>>>> val warehouseLocation ="/user/hive/warehouse" >>>>>>>>>>> >>>>>>>>>>> val spark = SparkSession.builder().appName("Spark Hive >>>>>>>>>>> Example").config("spark.sql.warehouse.dir", >>>>>>>>>>> warehouseLocation).enableHiveSupport().getOrCreate() >>>>>>>>>>> >>>>>>>>>>> import org.apache.spark.sql._ >>>>>>>>>>> var passion_df = spark.read. >>>>>>>>>>> format("jdbc"). >>>>>>>>>>> option("url", "jdbc:mysql://localhost:3307/policies"). >>>>>>>>>>> option("driver" ,"com.mysql.jdbc.Driver"). >>>>>>>>>>> option("user", "root"). >>>>>>>>>>> option("password", "root"). >>>>>>>>>>> option("dbtable", "insurancedetails"). >>>>>>>>>>> option("partitionColumn", "policyid"). >>>>>>>>>>> option("lowerBound", "1"). >>>>>>>>>>> option("upperBound", "100000"). >>>>>>>>>>> option("numPartitions", "4"). >>>>>>>>>>> load() >>>>>>>>>>> //Made sure that passion_df is created, as passion_df.show(5) >>>>>>>>>>> shows me correct data. >>>>>>>>>>> passion_df.write.saveAsTable("default.mine") //Default parquet >>>>>>>>>>> >>>>>>>>>>> STEP 3: Went to HIVE. Started HIVE prompt. >>>>>>>>>>> >>>>>>>>>>> hive> show tables; >>>>>>>>>>> OK >>>>>>>>>>> callcentervoicelogs >>>>>>>>>>> mine >>>>>>>>>>> Time taken: 0.035 seconds, Fetched: 2 row(s) >>>>>>>>>>> //As you can see HIVE is showing the table "mine" in default >>>>>>>>>>> schema. >>>>>>>>>>> >>>>>>>>>>> STEP 4: HERE IS THE PROBLEM. >>>>>>>>>>> >>>>>>>>>>> hive> select * from mine; >>>>>>>>>>> OK >>>>>>>>>>> Time taken: 0.354 seconds >>>>>>>>>>> hive> >>>>>>>>>>> //Where is the data ??? >>>>>>>>>>> >>>>>>>>>>> STEP 5: >>>>>>>>>>> >>>>>>>>>>> See the below command on HIVE >>>>>>>>>>> >>>>>>>>>>> describe formatted mine; >>>>>>>>>>> OK >>>>>>>>>>> # col_name data_type comment >>>>>>>>>>> >>>>>>>>>>> policyid int >>>>>>>>>>> statecode string >>>>>>>>>>> socialid string >>>>>>>>>>> county string >>>>>>>>>>> eq_site_limit decimal(10,2) >>>>>>>>>>> hu_site_limit decimal(10,2) >>>>>>>>>>> fl_site_limit decimal(10,2) >>>>>>>>>>> fr_site_limit decimal(10,2) >>>>>>>>>>> tiv_2014 decimal(10,2) >>>>>>>>>>> tiv_2015 decimal(10,2) >>>>>>>>>>> eq_site_deductible int >>>>>>>>>>> hu_site_deductible int >>>>>>>>>>> fl_site_deductible int >>>>>>>>>>> fr_site_deductible int >>>>>>>>>>> latitude decimal(6,6) >>>>>>>>>>> longitude decimal(6,6) >>>>>>>>>>> line string >>>>>>>>>>> construction string >>>>>>>>>>> point_granularity int >>>>>>>>>>> >>>>>>>>>>> # Detailed Table Information >>>>>>>>>>> Database: default >>>>>>>>>>> Owner: ravishankarnair >>>>>>>>>>> CreateTime: Sun Feb 11 00:26:40 EST 2018 >>>>>>>>>>> LastAccessTime: UNKNOWN >>>>>>>>>>> Protect Mode: None >>>>>>>>>>> Retention: 0 >>>>>>>>>>> Location: file:/Users/ravishankarnair/sp >>>>>>>>>>> ark-warehouse/mine >>>>>>>>>>> Table Type: MANAGED_TABLE >>>>>>>>>>> Table Parameters: >>>>>>>>>>> spark.sql.sources.provider parquet >>>>>>>>>>> spark.sql.sources.schema.numParts 1 >>>>>>>>>>> spark.sql.sources.schema.part.0 {\"type\":\"struct\",\"fields\ >>>>>>>>>>> ":[{\"name\":\"policyid\",\"type\":\"integer\",\"nullable\": >>>>>>>>>>> true,\"metadata\":{\"name\":\"policyid\",\"scale\":0}},{\"na >>>>>>>>>>> me\":\"statecode\",\"type\":\"string\",\"nullable\":true,\"m >>>>>>>>>>> etadata\":{\"name\":\"statecode\",\"scale\":0}},{\"name\":\" >>>>>>>>>>> Socialid\",\"type\":\"string\",\"nullable\":true,\"metadata\ >>>>>>>>>>> ":{\"name\":\"Socialid\",\"scale\":0}},{\"name\":\"county\", >>>>>>>>>>> \"type\":\"string\",\"nullable\":true,\"metadata\":{\"name\" >>>>>>>>>>> :\"county\",\"scale\":0}},{\"name\":\"eq_site_limit\",\"type >>>>>>>>>>> \":\"decimal(10,2)\",\"nullable\":true,\"metadata\":{\"name\ >>>>>>>>>>> ":\"eq_site_limit\",\"scale\":2}},{\"name\":\"hu_site_limit\ >>>>>>>>>>> ",\"type\":\"decimal(10,2)\",\"nullable\":true,\"metadata\": >>>>>>>>>>> {\"name\":\"hu_site_limit\",\"scale\":2}},{\"name\":\"fl_sit >>>>>>>>>>> e_limit\",\"type\":\"decimal(10,2)\",\"nullable\":true,\"met >>>>>>>>>>> adata\":{\"name\":\"fl_site_limit\",\"scale\":2}},{\"name\": >>>>>>>>>>> \"fr_site_limit\",\"type\":\"decimal(10,2)\",\"nullable\":tr >>>>>>>>>>> ue,\"metadata\":{\"name\":\"fr_site_limit\",\"scale\":2}},{\ >>>>>>>>>>> "name\":\"tiv_2014\",\"type\":\"decimal(10,2)\",\"nullable\" >>>>>>>>>>> :true,\"metadata\":{\"name\":\"tiv_2014\",\"scale\":2}},{\"n >>>>>>>>>>> ame\":\"tiv_2015\",\"type\":\"decimal(10,2)\",\"nullable\":t >>>>>>>>>>> rue,\"metadata\":{\"name\":\"tiv_2015\",\"scale\":2}},{\"nam >>>>>>>>>>> e\":\"eq_site_deductible\",\"type\":\"integer\",\"nullable\" >>>>>>>>>>> :true,\"metadata\":{\"name\":\"eq_site_deductible\",\"scale\ >>>>>>>>>>> ":0}},{\"name\":\"hu_site_deductible\",\"type\":\"integer\", >>>>>>>>>>> \"nullable\":true,\"metadata\":{\"name\":\"hu_site_deductibl >>>>>>>>>>> e\",\"scale\":0}},{\"name\":\"fl_site_deductible\",\"type\": >>>>>>>>>>> \"integer\",\"nullable\":true,\"metadata\":{\"name\":\"fl_ >>>>>>>>>>> site_deductible\",\"scale\":0}},{\"name\":\"fr_site_ >>>>>>>>>>> deductible\",\"type\":\"integer\",\"nullable\":true,\"metada >>>>>>>>>>> ta\":{\"name\":\"fr_site_deductible\",\"scale\":0}},{\" >>>>>>>>>>> name\":\"latitude\",\"type\":\"decimal(6,6)\",\"nullable\":t >>>>>>>>>>> rue,\"metadata\":{\"name\":\"latitude\",\"scale\":6}},{\"nam >>>>>>>>>>> e\":\"longitude\",\"type\":\"decimal(6,6)\",\"nullable\":tr >>>>>>>>>>> ue,\"metadata\":{\"name\":\"longitude\",\"scale\":6}},{\"nam >>>>>>>>>>> e\":\"line\",\"type\":\"string\",\"nullable\":true,\"metadat >>>>>>>>>>> a\":{\"name\":\"line\",\"scale\":0}},{\"name\":\"constr >>>>>>>>>>> uction\",\"type\":\"string\",\"nullable\":true,\"metadata\": >>>>>>>>>>> {\"name\":\"construction\",\"scale\":0}},{\"name\":\"point_ >>>>>>>>>>> granularity\",\"type\":\"integer\",\"nullable\":true,\" >>>>>>>>>>> metadata\":{\"name\":\"point_granularity\",\"scale\":0}}]} >>>>>>>>>>> transient_lastDdlTime 1518326800 >>>>>>>>>>> >>>>>>>>>>> # Storage Information >>>>>>>>>>> SerDe Library: org.apache.hadoop.hive.ql.io.p >>>>>>>>>>> arquet.serde.ParquetHiveSerDe >>>>>>>>>>> InputFormat: org.apache.hadoop.hive.ql.io.p >>>>>>>>>>> arquet.MapredParquetInputFormat >>>>>>>>>>> OutputFormat: org.apache.hadoop.hive.ql.io.p >>>>>>>>>>> arquet.MapredParquetOutputFormat >>>>>>>>>>> Compressed: No >>>>>>>>>>> Num Buckets: -1 >>>>>>>>>>> Bucket Columns: [] >>>>>>>>>>> Sort Columns: [] >>>>>>>>>>> Storage Desc Params: >>>>>>>>>>> path hdfs://localhost:8020/user/hiv >>>>>>>>>>> e/warehouse/mine >>>>>>>>>>> serialization.format 1 >>>>>>>>>>> Time taken: 0.077 seconds, Fetched: 48 row(s) >>>>>>>>>>> >>>>>>>>>>> Now, I see your advise and support. Whats the issue? Am I doing >>>>>>>>>>> wrong, it it a bug ? I am using Spark 2.2.1, HIVE 1.2.1, HADOOP >>>>>>>>>>> 2.7.3. All >>>>>>>>>>> class path, configuration are set properly. >>>>>>>>>>> >>>>>>>>>>> Best, >>>>>>>>>>> >>>>>>>>>>> Ravion >>>>>>>>>>> >>>>>>>>>>> On Fri, Feb 9, 2018 at 1:29 PM, Nicholas Hakobian < >>>>>>>>>>> nicholas.hakob...@rallyhealth.com> wrote: >>>>>>>>>>> >>>>>>>>>>>> Its possible that the format of your table is not compatible >>>>>>>>>>>> with your version of hive, so Spark saved it in a way such that >>>>>>>>>>>> only Spark >>>>>>>>>>>> can read it. When this happens it prints out a very visible >>>>>>>>>>>> warning letting >>>>>>>>>>>> you know this has happened. >>>>>>>>>>>> >>>>>>>>>>>> We've seen it most frequently when trying to save a parquet >>>>>>>>>>>> file with a column in date format into a Hive table. In older >>>>>>>>>>>> versions of >>>>>>>>>>>> hive, its parquet reader/writer did not support Date formats >>>>>>>>>>>> (among a >>>>>>>>>>>> couple others). >>>>>>>>>>>> >>>>>>>>>>>> Nicholas Szandor Hakobian, Ph.D. >>>>>>>>>>>> Staff Data Scientist >>>>>>>>>>>> Rally Health >>>>>>>>>>>> nicholas.hakob...@rallyhealth.com >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> On Fri, Feb 9, 2018 at 9:59 AM, Prakash Joshi < >>>>>>>>>>>> prakashcjos...@gmail.com> wrote: >>>>>>>>>>>> >>>>>>>>>>>>> Ravi, >>>>>>>>>>>>> >>>>>>>>>>>>> Can you send the result of >>>>>>>>>>>>> Show create table your_table_name >>>>>>>>>>>>> >>>>>>>>>>>>> Thanks >>>>>>>>>>>>> Prakash >>>>>>>>>>>>> >>>>>>>>>>>>> On Feb 9, 2018 8:20 PM, "☼ R Nair (रविशंकर नायर)" < >>>>>>>>>>>>> ravishankar.n...@gmail.com> wrote: >>>>>>>>>>>>> >>>>>>>>>>>>>> All, >>>>>>>>>>>>>> >>>>>>>>>>>>>> It has been three days continuously I am on this issue. Not >>>>>>>>>>>>>> getting any clue. >>>>>>>>>>>>>> >>>>>>>>>>>>>> Environment: Spark 2.2.x, all configurations are correct. >>>>>>>>>>>>>> hive-site.xml is in spark's conf. >>>>>>>>>>>>>> >>>>>>>>>>>>>> 1) Step 1: I created a data frame DF1 reading a csv file. >>>>>>>>>>>>>> >>>>>>>>>>>>>> 2) Did manipulations on DF1. Resulting frame is passion_df. >>>>>>>>>>>>>> >>>>>>>>>>>>>> 3) passion_df.write.format("orc") >>>>>>>>>>>>>> .saveAsTable("sampledb.passion") >>>>>>>>>>>>>> >>>>>>>>>>>>>> 4) The metastore shows the hive table., when I do "show >>>>>>>>>>>>>> tables" in HIVE, I can see table name >>>>>>>>>>>>>> >>>>>>>>>>>>>> 5) I can't select in HIVE, though I can select from SPARK as >>>>>>>>>>>>>> spark.sql("select * from sampledb.passion") >>>>>>>>>>>>>> >>>>>>>>>>>>>> Whats going on here? Please help. Why I am not seeing data >>>>>>>>>>>>>> from HIVE prompt? >>>>>>>>>>>>>> The "describe formatted " command on the table in HIVE shows >>>>>>>>>>>>>> he data is is in default warehouse location ( >>>>>>>>>>>>>> /user/hive/warehouse) since I >>>>>>>>>>>>>> set it. >>>>>>>>>>>>>> >>>>>>>>>>>>>> I am not getting any definite answer anywhere. Many >>>>>>>>>>>>>> suggestions and answers given in Stackoverflow et al.Nothing >>>>>>>>>>>>>> really works. >>>>>>>>>>>>>> >>>>>>>>>>>>>> So asking experts here for some light on this, thanks >>>>>>>>>>>>>> >>>>>>>>>>>>>> Best, >>>>>>>>>>>>>> Ravion >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> -- >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> -- >>>>>>>>> >>>>>>>>> >>>>>>> >>>>>>> >>>>>>> -- >>>>>>> >>>>>>> >>>>>> >>>>>> >>>>>> -- >>>>>> Shmuel Blitz >>>>>> Big Data Developer >>>>>> Email: shmuel.bl...@similarweb.com >>>>>> www.similarweb.com >>>>>> <https://www.facebook.com/SimilarWeb/> >>>>>> <https://www.linkedin.com/company/429838/> >>>>>> <https://twitter.com/similarweb> >>>>>> >>>>> >>>>> >>>>> >>>>> -- >>>>> >>>>> >>>> >>>> >>>> -- >>>> Thanks >>>> Deepak >>>> www.bigdatabig.com >>>> www.keosha.net >>>> >>> >>> >>> >>> -- >>> Thanks >>> Deepak >>> www.bigdatabig.com >>> www.keosha.net >>> >> >> >> >> -- >> >> > > > -- > Thanks > Deepak > www.bigdatabig.com > www.keosha.net > -- Shmuel Blitz Big Data Developer Email: shmuel.bl...@similarweb.com www.similarweb.com <https://www.facebook.com/SimilarWeb/> <https://www.linkedin.com/company/429838/> <https://twitter.com/similarweb>