Hi Ayan,

I don’t know I did something wrong but still couldn’t set 
hive.metastore.warehouse.dir property.

I set 3 hive-site.xml files in spark location, zeppelin, hive as well but still 
didn’t work.

zeppeline/conf/hive-site.xml 
spark/conf/hive-site.xml
hive/conf/hive-site.xml



My hive-site.xml

<configuration>
        <property>
          <name>hive.metastore.metadb.dir</name>
          <value>alluxio://master1:19998/metadb</value>
          <description>
          Required by metastore server or if the uris argument below is not 
supplied
          </description>
        </property>
        <property>
          <name>hive.metastore.warehouse.dir</name>
          <value>alluxio://master1:19998/warehouse</value>
          <description>
          Required by metastore server or if the uris argument below is not 
supplied
          </description>
        </property>
</configuration>

Is there anything I can do?

Regards,
Chanh



> On Jul 13, 2016, at 12:43 PM, ayan guha <guha.a...@gmail.com> wrote:
> 
> Hi
> 
> Create table always happens through Hive. In Hive, when you create a 
> database, the default metadata storage location is driven by 
> hive.metastore.metadb.dir and data storage is driven by 
> hive.metastore.warehouse.dir property (set in hive site xml). So, you do not 
> need to set this property in Zeppelin. 
> 
> What you can do:
>            a. Modify  hive-site.xml to include those properties, if they are 
> not already set.  use the same hive site.xml to run STS. Then connect through 
> JDBC, create table and you should find metadata & data in your desired 
> location. 
> b. I think you can set these properties (same way you'd do in hive cli)
> c. You can create tables/databases with a LOCATION clause,  in case you need 
> to use non-standard path. 
> 
> Best
> Ayan
> 
> On Wed, Jul 13, 2016 at 3:20 PM, Chanh Le <giaosu...@gmail.com 
> <mailto:giaosu...@gmail.com>> wrote:
> Hi Ayan,
> Thank you for replying. 
> But I wanna create a table in Zeppelin and store the metadata in Alluxio like 
> I tried to do set hive.metastore.warehouse.dir=alluxio://master1:19998/metadb 
> <>   <>So I can share data with STS.
> 
> The way you’ve mentioned through JDBC I already did and it works but I can’t 
> create table in Spark way easily.
> 
> Regards,
> Chanh
> 
> 
>> On Jul 13, 2016, at 12:06 PM, ayan guha <guha.a...@gmail.com 
>> <mailto:guha.a...@gmail.com>> wrote:
>> 
>> HI
>> 
>> I quickly tried with available hive interpreter 
>> 
>> <image.png>
>> 
>> Please try similarly. 
>> 
>> I will try with jdbc interpreter but I need to add it to zeppelin :)
>> 
>> Best
>> Ayan
>> 
>> On Wed, Jul 13, 2016 at 1:53 PM, Chanh Le <giaosu...@gmail.com 
>> <mailto:giaosu...@gmail.com>> wrote:
>> Hi Ayan,
>> How to set hive metastore in Zeppelin. I tried but not success.
>> The way I do I add into Spark Interpreter
>> 
>> <Screen Shot 2016-07-13 at 10.50.53 AM.png>
>> 
>> And also try in a notebook by 
>> %sql
>> set hive.metastore.metadb.dir=alluxio://master1:19998/metadb <>
>> 
>> %sql 
>> set hive.metastore.warehouse.dir=alluxio://master1:19998/metadb <>
>> 
>> %spark
>> sqlContext.setConf("hive.metastore.warehouse.dir", 
>> "alluxio://master1:19998/metadb <>")
>> sqlContext.setConf("hive.metastore.metadb.dir", 
>> "alluxio://master1:19998/metadb <>")
>> sqlContext.read.parquet("alluxio://master1:19998/etl_info/WEBSITE 
>> <>").saveAsTable("tests_5”)
>> 
>> But It’s <Screen Shot 2016-07-13 at 10.53.10 AM.png>
>> 
>>> On Jul 11, 2016, at 1:26 PM, ayan guha <guha.a...@gmail.com 
>>> <mailto:guha.a...@gmail.com>> wrote:
>>> 
>>> Hi
>>> 
>>> When you say "Zeppelin and STS", I am assuming you mean "Spark Interpreter" 
>>> and "JDBC interpreter" respectively. 
>>> 
>>> Through Zeppelin, you can either run your own spark application (by using 
>>> Zeppelin's own spark context) using spark interpreter OR you can access 
>>> STS, which  is a spark application ie separate Spark Context using JDBC 
>>> interpreter. There should not be any need for these 2 contexts to coexist. 
>>> 
>>> If you want to share data, save it to hive from either context, and you 
>>> should be able to see the data from other context. 
>>> 
>>> Best
>>> Ayan
>>> 
>>> 
>>> 
>>> On Mon, Jul 11, 2016 at 3:00 PM, Chanh Le <giaosu...@gmail.com 
>>> <mailto:giaosu...@gmail.com>> wrote:
>>> Hi Ayan,
>>> I tested It works fine but one more confuse is If my (technical) users want 
>>> to write some code in zeppelin to apply thing into Hive table? 
>>> Zeppelin and STS can’t share Spark Context that mean we need separated 
>>> process? Is there anyway to use the same Spark Context of STS?
>>> 
>>> Regards,
>>> Chanh
>>> 
>>> 
>>>> On Jul 11, 2016, at 10:05 AM, Takeshi Yamamuro <linguin....@gmail.com 
>>>> <mailto:linguin....@gmail.com>> wrote:
>>>> 
>>>> Hi,
>>>> 
>>>> ISTM multiple sparkcontexts are not recommended in spark.
>>>> See: https://issues.apache.org/jira/browse/SPARK-2243 
>>>> <https://issues.apache.org/jira/browse/SPARK-2243>
>>>> 
>>>> // maropu
>>>> 
>>>> 
>>>> On Mon, Jul 11, 2016 at 12:01 PM, ayan guha <guha.a...@gmail.com 
>>>> <mailto:guha.a...@gmail.com>> wrote:
>>>> Hi
>>>> 
>>>> Can you try using JDBC interpreter with STS? We are using Zeppelin+STS on 
>>>> YARN for few months now without much issue. 
>>>> 
>>>> On Mon, Jul 11, 2016 at 12:48 PM, Chanh Le <giaosu...@gmail.com 
>>>> <mailto:giaosu...@gmail.com>> wrote:
>>>> Hi everybody,
>>>> We are using Spark to query big data and currently we’re using Zeppelin to 
>>>> provide a UI for technical users.
>>>> Now we also need to provide a UI for business users so we use Oracle BI 
>>>> tools and set up a Spark Thrift Server (STS) for it.
>>>> 
>>>> When I run both Zeppelin and STS throw error:
>>>> 
>>>> INFO [2016-07-11 09:40:21,905] ({pool-2-thread-4} 
>>>> SchedulerFactory.java[jobStarted]:131) - Job 
>>>> remoteInterpretJob_1468204821905 started by scheduler 
>>>> org.apache.zeppelin.spark.SparkInterpreter835015739
>>>>  INFO [2016-07-11 09:40:21,911] ({pool-2-thread-4} 
>>>> Logging.scala[logInfo]:58) - Changing view acls to: giaosudau
>>>>  INFO [2016-07-11 09:40:21,912] ({pool-2-thread-4} 
>>>> Logging.scala[logInfo]:58) - Changing modify acls to: giaosudau
>>>>  INFO [2016-07-11 09:40:21,912] ({pool-2-thread-4} 
>>>> Logging.scala[logInfo]:58) - SecurityManager: authentication disabled; ui 
>>>> acls disabled; users with view permissions: Set(giaosudau); users with 
>>>> modify permissions: Set(giaosudau)
>>>>  INFO [2016-07-11 09:40:21,918] ({pool-2-thread-4} 
>>>> Logging.scala[logInfo]:58) - Starting HTTP Server
>>>>  INFO [2016-07-11 09:40:21,919] ({pool-2-thread-4} 
>>>> Server.java[doStart]:272) - jetty-8.y.z-SNAPSHOT
>>>>  INFO [2016-07-11 09:40:21,920] ({pool-2-thread-4} 
>>>> AbstractConnector.java[doStart]:338) - Started 
>>>> SocketConnector@0.0.0.0:54818 <http://SocketConnector@0.0.0.0:54818/>
>>>>  INFO [2016-07-11 09:40:21,922] ({pool-2-thread-4} 
>>>> Logging.scala[logInfo]:58) - Successfully started service 'HTTP class 
>>>> server' on port 54818.
>>>>  INFO [2016-07-11 09:40:22,408] ({pool-2-thread-4} 
>>>> SparkInterpreter.java[createSparkContext]:233) - ------ Create new 
>>>> SparkContext local[*] -------
>>>>  WARN [2016-07-11 09:40:22,411] ({pool-2-thread-4} 
>>>> Logging.scala[logWarning]:70) - Another SparkContext is being constructed 
>>>> (or threw an exception in its constructor).  This may indicate an error, 
>>>> since only one SparkContext may be running in this JVM (see SPARK-2243). 
>>>> The other SparkContext was created at:
>>>> 
>>>> Is that mean I need to setup allow multiple context? Because It’s only 
>>>> test in local with local mode If I deploy on mesos cluster what would 
>>>> happened?
>>>> 
>>>> Need you guys suggests some solutions for that. Thanks.
>>>> 
>>>> Chanh
>>>> ---------------------------------------------------------------------
>>>> To unsubscribe e-mail: user-unsubscr...@spark.apache.org 
>>>> <mailto:user-unsubscr...@spark.apache.org>
>>>> 
>>>> 
>>>> 
>>>> 
>>>> -- 
>>>> Best Regards,
>>>> Ayan Guha
>>>> 
>>>> 
>>>> 
>>>> -- 
>>>> ---
>>>> Takeshi Yamamuro
>>> 
>>> 
>>> 
>>> 
>>> -- 
>>> Best Regards,
>>> Ayan Guha
>> 
>> 
>> 
>> 
>> -- 
>> Best Regards,
>> Ayan Guha
> 
> 
> 
> 
> -- 
> Best Regards,
> Ayan Guha

Reply via email to