Try this in hive:
alter table mine set locations "hdfs://localhost:8020/
user/hive/warehouse/mine";

Thanks
Deepak

On Sun, Feb 11, 2018 at 1:24 PM, ☼ R Nair (रविशंकर नायर) <
ravishankar.n...@gmail.com> wrote:

> Hi,
> Here you go:
>
> hive> show create table mine;
> OK
> CREATE TABLE `mine`(
>   `policyid` int,
>   `statecode` string,
>   `socialid` string,
>   `county` string,
>   `eq_site_limit` decimal(10,2),
>   `hu_site_limit` decimal(10,2),
>   `fl_site_limit` decimal(10,2),
>   `fr_site_limit` decimal(10,2),
>   `tiv_2014` decimal(10,2),
>   `tiv_2015` decimal(10,2),
>   `eq_site_deductible` int,
>   `hu_site_deductible` int,
>   `fl_site_deductible` int,
>   `fr_site_deductible` int,
>   `latitude` decimal(6,6),
>   `longitude` decimal(6,6),
>   `line` string,
>   `construction` string,
>   `point_granularity` int)
> ROW FORMAT SERDE
>   'org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe'
> WITH SERDEPROPERTIES (
>   'path'='hdfs://localhost:8020/user/hive/warehouse/mine')
> STORED AS INPUTFORMAT
>   'org.apache.hadoop.hive.ql.io.parquet.MapredParquetInputFormat'
> OUTPUTFORMAT
>   'org.apache.hadoop.hive.ql.io.parquet.MapredParquetOutputFormat'
> LOCATION
>   'file:/Users/ravishankarnair/spark-warehouse/mine'
> TBLPROPERTIES (
>   'spark.sql.sources.provider'='parquet',
>   'spark.sql.sources.schema.numParts'='1',
>   'spark.sql.sources.schema.part.0'='{\"type\":\"struct\",
> \"fields\":[{\"name\":\"policyid\",\"type\":\"integer\
> ",\"nullable\":true,\"metadata\":{\"name\":\"policyid\",\"scale\":0}},{\"
> name\":\"statecode\",\"type\":\"string\",\"nullable\":true,\
> "metadata\":{\"name\":\"statecode\",\"scale\":0}},{\"
> name\":\"Socialid\",\"type\":\"string\",\"nullable\":true,\"
> metadata\":{\"name\":\"Socialid\",\"scale\":0}},{\"
> name\":\"county\",\"type\":\"string\",\"nullable\":true,\"
> metadata\":{\"name\":\"county\",\"scale\":0}},{\"name\":\"
> eq_site_limit\",\"type\":\"decimal(10,2)\",\"nullable\":
> true,\"metadata\":{\"name\":\"eq_site_limit\",\"scale\":2}},
> {\"name\":\"hu_site_limit\",\"type\":\"decimal(10,2)\",\"
> nullable\":true,\"metadata\":{\"name\":\"hu_site_limit\",\"
> scale\":2}},{\"name\":\"fl_site_limit\",\"type\":\"
> decimal(10,2)\",\"nullable\":true,\"metadata\":{\"name\":\"
> fl_site_limit\",\"scale\":2}},{\"name\":\"fr_site_limit\",\"
> type\":\"decimal(10,2)\",\"nullable\":true,\"metadata\":{
> \"name\":\"fr_site_limit\",\"scale\":2}},{\"name\":\"tiv_
> 2014\",\"type\":\"decimal(10,2)\",\"nullable\":true,\"
> metadata\":{\"name\":\"tiv_2014\",\"scale\":2}},{\"name\"
> :\"tiv_2015\",\"type\":\"decimal(10,2)\",\"nullable\":
> true,\"metadata\":{\"name\":\"tiv_2015\",\"scale\":2}},{\"
> name\":\"eq_site_deductible\",\"type\":\"integer\",\"
> nullable\":true,\"metadata\":{\"name\":\"eq_site_deductible\
> ",\"scale\":0}},{\"name\":\"hu_site_deductible\",\"type\":
> \"integer\",\"nullable\":true,\"metadata\":{\"name\":\"hu_
> site_deductible\",\"scale\":0}},{\"name\":\"fl_site_
> deductible\",\"type\":\"integer\",\"nullable\":true,\"
> metadata\":{\"name\":\"fl_site_deductible\",\"scale\":0}
> },{\"name\":\"fr_site_deductible\",\"type\":\"
> integer\",\"nullable\":true,\"metadata\":{\"name\":\"fr_
> site_deductible\",\"scale\":0}},{\"name\":\"latitude\",\"
> type\":\"decimal(6,6)\",\"nullable\":true,\"metadata\":{
> \"name\":\"latitude\",\"scale\":6}},{\"name\":\"longitude\",
> \"type\":\"decimal(6,6)\",\"nullable\":true,\"metadata\":{
> \"name\":\"longitude\",\"scale\":6}},{\"name\":\"line\"
> ,\"type\":\"string\",\"nullable\":true,\"metadata\":{
> \"name\":\"line\",\"scale\":0}},{\"name\":\"construction\",\
> "type\":\"string\",\"nullable\":true,\"metadata\":{\"name\":
> \"construction\",\"scale\":0}},{\"name\":\"point_granularity\",\"type\":\"
> integer\",\"nullable\":true,\"metadata\":{\"name\":\"point_
> granularity\",\"scale\":0}}]}',
>   'transient_lastDdlTime'='1518335598')
> Time taken: 0.13 seconds, Fetched: 35 row(s)
>
> On Sun, Feb 11, 2018 at 2:36 AM, Shmuel Blitz <shmuel.bl...@similarweb.com
> > wrote:
>
>> Please run the following command, and paste the result:
>> SHOW CREATE TABLE <<TABLE-NAME>>
>>
>> On Sun, Feb 11, 2018 at 7:56 AM, ☼ R Nair (रविशंकर नायर) <
>> ravishankar.n...@gmail.com> wrote:
>>
>>> No, No luck.
>>>
>>> Thanks
>>>
>>> On Sun, Feb 11, 2018 at 12:48 AM, Deepak Sharma <deepakmc...@gmail.com>
>>> wrote:
>>>
>>>> In hive cli:
>>>> msck repair table 《table_name》;
>>>>
>>>> Thanks
>>>> Deepak
>>>>
>>>> On Feb 11, 2018 11:14, "☼ R Nair (रविशंकर नायर)" <
>>>> ravishankar.n...@gmail.com> wrote:
>>>>
>>>>> NO, can you pease explain the command ? Let me try now.
>>>>>
>>>>> Best,
>>>>>
>>>>> On Sun, Feb 11, 2018 at 12:40 AM, Deepak Sharma <deepakmc...@gmail.com
>>>>> > wrote:
>>>>>
>>>>>> I am not sure about the exact issue bjt i see you are partioning
>>>>>> while writing from spark.
>>>>>> Did you tried msck repair on the table before reading it in hive ?
>>>>>>
>>>>>> Thanks
>>>>>> Deepak
>>>>>>
>>>>>> On Feb 11, 2018 11:06, "☼ R Nair (रविशंकर नायर)" <
>>>>>> ravishankar.n...@gmail.com> wrote:
>>>>>>
>>>>>>> All,
>>>>>>>
>>>>>>> Thanks for the inputs. Again I am not successful. I think, we need
>>>>>>> to resolve this, as this is a very common requirement.
>>>>>>>
>>>>>>> Please go through my complete code:
>>>>>>>
>>>>>>> STEP 1:  Started Spark shell as spark-shell --master yarn
>>>>>>>
>>>>>>> STEP 2: Flowing code is being given as inout to shark shell
>>>>>>>
>>>>>>> import org.apache.spark.sql.Row
>>>>>>> import org.apache.spark.sql.SparkSession
>>>>>>> val warehouseLocation ="/user/hive/warehouse"
>>>>>>>
>>>>>>> val spark = SparkSession.builder().appName("Spark Hive
>>>>>>> Example").config("spark.sql.warehouse.dir",
>>>>>>> warehouseLocation).enableHiveSupport().getOrCreate()
>>>>>>>
>>>>>>> import org.apache.spark.sql._
>>>>>>> var passion_df = spark.read.
>>>>>>> format("jdbc").
>>>>>>> option("url", "jdbc:mysql://localhost:3307/policies").
>>>>>>> option("driver" ,"com.mysql.jdbc.Driver").
>>>>>>> option("user", "root").
>>>>>>> option("password", "root").
>>>>>>> option("dbtable", "insurancedetails").
>>>>>>> option("partitionColumn", "policyid").
>>>>>>> option("lowerBound", "1").
>>>>>>> option("upperBound", "100000").
>>>>>>> option("numPartitions", "4").
>>>>>>> load()
>>>>>>> //Made sure that passion_df is created, as passion_df.show(5) shows
>>>>>>> me correct data.
>>>>>>> passion_df.write.saveAsTable("default.mine") //Default parquet
>>>>>>>
>>>>>>> STEP 3: Went to HIVE. Started HIVE prompt.
>>>>>>>
>>>>>>> hive> show tables;
>>>>>>> OK
>>>>>>> callcentervoicelogs
>>>>>>> mine
>>>>>>> Time taken: 0.035 seconds, Fetched: 2 row(s)
>>>>>>> //As you can see HIVE is showing the table "mine" in default schema.
>>>>>>>
>>>>>>> STEP 4: HERE IS THE PROBLEM.
>>>>>>>
>>>>>>> hive> select * from mine;
>>>>>>> OK
>>>>>>> Time taken: 0.354 seconds
>>>>>>> hive>
>>>>>>> //Where is the data ???
>>>>>>>
>>>>>>> STEP 5:
>>>>>>>
>>>>>>> See the below command on HIVE
>>>>>>>
>>>>>>> describe formatted mine;
>>>>>>> OK
>>>>>>> # col_name             data_type           comment
>>>>>>>
>>>>>>> policyid             int
>>>>>>> statecode           string
>>>>>>> socialid             string
>>>>>>> county               string
>>>>>>> eq_site_limit       decimal(10,2)
>>>>>>> hu_site_limit       decimal(10,2)
>>>>>>> fl_site_limit       decimal(10,2)
>>>>>>> fr_site_limit       decimal(10,2)
>>>>>>> tiv_2014             decimal(10,2)
>>>>>>> tiv_2015             decimal(10,2)
>>>>>>> eq_site_deductible   int
>>>>>>> hu_site_deductible   int
>>>>>>> fl_site_deductible   int
>>>>>>> fr_site_deductible   int
>>>>>>> latitude             decimal(6,6)
>>>>>>> longitude           decimal(6,6)
>>>>>>> line                 string
>>>>>>> construction         string
>>>>>>> point_granularity   int
>>>>>>>
>>>>>>> # Detailed Table Information
>>>>>>> Database:           default
>>>>>>> Owner:               ravishankarnair
>>>>>>> CreateTime:         Sun Feb 11 00:26:40 EST 2018
>>>>>>> LastAccessTime:     UNKNOWN
>>>>>>> Protect Mode:       None
>>>>>>> Retention:           0
>>>>>>> Location:           file:/Users/ravishankarnair/spark-warehouse/mine
>>>>>>>
>>>>>>> Table Type:         MANAGED_TABLE
>>>>>>> Table Parameters:
>>>>>>> spark.sql.sources.provider parquet
>>>>>>> spark.sql.sources.schema.numParts 1
>>>>>>> spark.sql.sources.schema.part.0 {\"type\":\"struct\",\"fields\
>>>>>>> ":[{\"name\":\"policyid\",\"type\":\"integer\",\"nullable\":
>>>>>>> true,\"metadata\":{\"name\":\"policyid\",\"scale\":0}},{\"na
>>>>>>> me\":\"statecode\",\"type\":\"string\",\"nullable\":true,\"m
>>>>>>> etadata\":{\"name\":\"statecode\",\"scale\":0}},{\"name\":\"
>>>>>>> Socialid\",\"type\":\"string\",\"nullable\":true,\"metadata\
>>>>>>> ":{\"name\":\"Socialid\",\"scale\":0}},{\"name\":\"county\",
>>>>>>> \"type\":\"string\",\"nullable\":true,\"metadata\":{\"name\"
>>>>>>> :\"county\",\"scale\":0}},{\"name\":\"eq_site_limit\",\"type
>>>>>>> \":\"decimal(10,2)\",\"nullable\":true,\"metadata\":{\"name\
>>>>>>> ":\"eq_site_limit\",\"scale\":2}},{\"name\":\"hu_site_limit\
>>>>>>> ",\"type\":\"decimal(10,2)\",\"nullable\":true,\"metadata\":
>>>>>>> {\"name\":\"hu_site_limit\",\"scale\":2}},{\"name\":\"fl_sit
>>>>>>> e_limit\",\"type\":\"decimal(10,2)\",\"nullable\":true,\"met
>>>>>>> adata\":{\"name\":\"fl_site_limit\",\"scale\":2}},{\"name\
>>>>>>> ":\"fr_site_limit\",\"type\":\"decimal(10,2)\",\"nullable\":
>>>>>>> true,\"metadata\":{\"name\":\"fr_site_limit\",\"scale\":2}},
>>>>>>> {\"name\":\"tiv_2014\",\"type\":\"decimal(10,2)\",\"
>>>>>>> nullable\":true,\"metadata\":{\"name\":\"tiv_2014\",\"scale\
>>>>>>> ":2}},{\"name\":\"tiv_2015\",\"type\":\"decimal(10,2)\",\"
>>>>>>> nullable\":true,\"metadata\":{\"name\":\"tiv_2015\",\"scale\
>>>>>>> ":2}},{\"name\":\"eq_site_deductible\",\"type\":\"
>>>>>>> integer\",\"nullable\":true,\"metadata\":{\"name\":\"eq_
>>>>>>> site_deductible\",\"scale\":0}},{\"name\":\"hu_site_deductib
>>>>>>> le\",\"type\":\"integer\",\"nullable\":true,\"metadata\":{
>>>>>>> \"name\":\"hu_site_deductible\",\"scale\":0}},{\"name\":\"
>>>>>>> fl_site_deductible\",\"type\":\"integer\",\"nullable\":true,
>>>>>>> \"metadata\":{\"name\":\"fl_site_deductible\",\"scale\":0}
>>>>>>> },{\"name\":\"fr_site_deductible\",\"type\":\"intege
>>>>>>> r\",\"nullable\":true,\"metadata\":{\"name\":\"fr_site_
>>>>>>> deductible\",\"scale\":0}},{\"name\":\"latitude\",\"type\":\
>>>>>>> "decimal(6,6)\",\"nullable\":true,\"metadata\":{\"name\":\"l
>>>>>>> atitude\",\"scale\":6}},{\"name\":\"longitude\",\"type\":\"
>>>>>>> decimal(6,6)\",\"nullable\":true,\"metadata\":{\"name\":\"lo
>>>>>>> ngitude\",\"scale\":6}},{\"name\":\"line\",\"type\":\"string
>>>>>>> \",\"nullable\":true,\"metadata\":{\"name\":\"line\",\"
>>>>>>> scale\":0}},{\"name\":\"construction\",\"type\":\"string\",\
>>>>>>> "nullable\":true,\"metadata\":{\"name\":\"construction\",\"
>>>>>>> scale\":0}},{\"name\":\"point_granularity\",\"type\":\"
>>>>>>> integer\",\"nullable\":true,\"metadata\":{\"name\":\"point_
>>>>>>> granularity\",\"scale\":0}}]}
>>>>>>> transient_lastDdlTime 1518326800
>>>>>>>
>>>>>>> # Storage Information
>>>>>>> SerDe Library:       org.apache.hadoop.hive.ql.io.p
>>>>>>> arquet.serde.ParquetHiveSerDe
>>>>>>> InputFormat:         org.apache.hadoop.hive.ql.io.p
>>>>>>> arquet.MapredParquetInputFormat
>>>>>>> OutputFormat:       org.apache.hadoop.hive.ql.io.p
>>>>>>> arquet.MapredParquetOutputFormat
>>>>>>> Compressed:         No
>>>>>>> Num Buckets:         -1
>>>>>>> Bucket Columns:     []
>>>>>>> Sort Columns:       []
>>>>>>> Storage Desc Params:
>>>>>>> path                 hdfs://localhost:8020/user/hive/warehouse/mine
>>>>>>> serialization.format 1
>>>>>>> Time taken: 0.077 seconds, Fetched: 48 row(s)
>>>>>>>
>>>>>>> Now, I see your advise and support. Whats the issue? Am I doing
>>>>>>> wrong, it it a bug ? I am using Spark 2.2.1, HIVE 1.2.1, HADOOP 2.7.3. 
>>>>>>> All
>>>>>>> class path, configuration are set properly.
>>>>>>>
>>>>>>> Best,
>>>>>>>
>>>>>>> Ravion
>>>>>>>
>>>>>>> On Fri, Feb 9, 2018 at 1:29 PM, Nicholas Hakobian <
>>>>>>> nicholas.hakob...@rallyhealth.com> wrote:
>>>>>>>
>>>>>>>> Its possible that the format of your table is not compatible with
>>>>>>>> your version of hive, so Spark saved it in a way such that only Spark 
>>>>>>>> can
>>>>>>>> read it. When this happens it prints out a very visible warning 
>>>>>>>> letting you
>>>>>>>> know this has happened.
>>>>>>>>
>>>>>>>> We've seen it most frequently when trying to save a parquet file
>>>>>>>> with a column in date format into a Hive table. In older versions of 
>>>>>>>> hive,
>>>>>>>> its parquet reader/writer did not support Date formats (among a couple
>>>>>>>> others).
>>>>>>>>
>>>>>>>> Nicholas Szandor Hakobian, Ph.D.
>>>>>>>> Staff Data Scientist
>>>>>>>> Rally Health
>>>>>>>> nicholas.hakob...@rallyhealth.com
>>>>>>>>
>>>>>>>>
>>>>>>>> On Fri, Feb 9, 2018 at 9:59 AM, Prakash Joshi <
>>>>>>>> prakashcjos...@gmail.com> wrote:
>>>>>>>>
>>>>>>>>> Ravi,
>>>>>>>>>
>>>>>>>>> Can you send the result of
>>>>>>>>> Show create table your_table_name
>>>>>>>>>
>>>>>>>>> Thanks
>>>>>>>>> Prakash
>>>>>>>>>
>>>>>>>>> On Feb 9, 2018 8:20 PM, "☼ R Nair (रविशंकर नायर)" <
>>>>>>>>> ravishankar.n...@gmail.com> wrote:
>>>>>>>>>
>>>>>>>>>> All,
>>>>>>>>>>
>>>>>>>>>> It has been three days continuously I am on this issue. Not
>>>>>>>>>> getting any clue.
>>>>>>>>>>
>>>>>>>>>> Environment: Spark 2.2.x, all configurations are correct.
>>>>>>>>>> hive-site.xml is in spark's conf.
>>>>>>>>>>
>>>>>>>>>> 1) Step 1: I created a data frame DF1 reading a csv file.
>>>>>>>>>>
>>>>>>>>>> 2) Did  manipulations on DF1. Resulting frame is passion_df.
>>>>>>>>>>
>>>>>>>>>> 3) passion_df.write.format("orc").saveAsTable("sampledb.passion")
>>>>>>>>>>
>>>>>>>>>> 4) The metastore shows the hive table., when I do "show tables"
>>>>>>>>>> in HIVE, I can see table name
>>>>>>>>>>
>>>>>>>>>> 5) I can't select in HIVE, though I can select from SPARK as
>>>>>>>>>> spark.sql("select * from sampledb.passion")
>>>>>>>>>>
>>>>>>>>>> Whats going on here? Please help. Why I am not seeing data from
>>>>>>>>>> HIVE prompt?
>>>>>>>>>> The "describe formatted " command on the table in HIVE shows he
>>>>>>>>>> data is is in default warehouse location ( /user/hive/warehouse) 
>>>>>>>>>> since I
>>>>>>>>>> set it.
>>>>>>>>>>
>>>>>>>>>> I am not getting any definite answer anywhere. Many suggestions
>>>>>>>>>> and answers given in Stackoverflow et al.Nothing really works.
>>>>>>>>>>
>>>>>>>>>> So asking experts here for some light on this, thanks
>>>>>>>>>>
>>>>>>>>>> Best,
>>>>>>>>>> Ravion
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> --
>>>>>>>
>>>>>>>
>>>>>
>>>>>
>>>>> --
>>>>>
>>>>>
>>>
>>>
>>> --
>>>
>>>
>>
>>
>> --
>> Shmuel Blitz
>> Big Data Developer
>> Email: shmuel.bl...@similarweb.com
>> www.similarweb.com
>> <https://www.facebook.com/SimilarWeb/>
>> <https://www.linkedin.com/company/429838/>
>> <https://twitter.com/similarweb>
>>
>
>
>
> --
>
>


-- 
Thanks
Deepak
www.bigdatabig.com
www.keosha.net

Reply via email to