Hi Ayan, I succeed with Tableau but I still can’t import metadata from Hive to Oracle BI.
Is that Oracle BI still can’t connect to STS. Regards, Chanh > On Jul 15, 2016, at 11:44 AM, ayan guha <guha.a...@gmail.com> wrote: > > Its possible that transfar protocols are not matching, thats what Simba is > complaining about. Try to change the protocol to SASL? > > On Fri, Jul 15, 2016 at 1:20 PM, Chanh Le <giaosu...@gmail.com > <mailto:giaosu...@gmail.com>> wrote: > Hi Ayan, > Thanks. I got it. > Did you have any problem when connecting Oracle BI with STS? > > I have some error > <IMG_15072016_095351.png> > > If I use Tableau > <Screen Shot 2016-07-15 at 10.19.40 AM.png> > > >> On Jul 15, 2016, at 10:03 AM, ayan guha <guha.a...@gmail.com >> <mailto:guha.a...@gmail.com>> wrote: >> >> This looks like a Spark code. I am not sure if this is what you intended to >> use with STS? I think it should be a insert overwrite command (SQL) >> >> On Fri, Jul 15, 2016 at 12:22 PM, Chanh Le <giaosu...@gmail.com >> <mailto:giaosu...@gmail.com>> wrote: >> Hi Ayan, >> >> Spark Code: >> df.write.mode(SaveMode.Overwrite). >> parquet(dataPath) >> I overwrite current data frame every hour to update data. >> dataPath is /etl_info/WEBSITE >> So that mean the part-xxx file will change every hour as well. >> This also happen on Spark if I register as tempTable and I need to drop and >> recreate every time data was changed. >> >> Regards, >> Chanh >> >> >> >>> On Jul 14, 2016, at 10:39 PM, ayan guha <guha.a...@gmail.com >>> <mailto:guha.a...@gmail.com>> wrote: >>> >>> Can you kindly share essential parts your spark code? >>> >>> On 14 Jul 2016 21:12, "Chanh Le" <giaosu...@gmail.com >>> <mailto:giaosu...@gmail.com>> wrote: >>> Hi Ayan, >>> Thank you for your suggestion. I switch to use STS for primary and Zeppelin >>> just call through JDBC. >>> But my data is a bit update data so Every hour I need to rewrite of some >>> data and therefore it has a issue with parquet file. >>> Scenario is: I have parquet file NETWORK and each one hour I need overwrite >>> that to update a new one and that gonna happened. >>> Caused by: alluxio.exception.FileDoesNotExistException: Path >>> /etl_info/WEBSITE/part-r-00001-a26015c3-6de5-4a10-bda8-d985d7241953.snappy.parquet >>> does not exist. >>> So I need to drop old table and create a new one. >>> >>> Is there anyway to work around this? >>> >>> Regards, >>> Chanh >>> >>> >>>> On Jul 14, 2016, at 11:36 AM, ayan guha <guha.a...@gmail.com >>>> <mailto:guha.a...@gmail.com>> wrote: >>>> >>>> Hi >>>> >>>> Thanks for the information. However, I still strongly believe you should >>>> be able to set URI in STS hive site xml and then create table through >>>> JDBC. >>>> >>>> On Thu, Jul 14, 2016 at 1:49 PM, Chanh Le <giaosu...@gmail.com >>>> <mailto:giaosu...@gmail.com>> wrote: >>>> Hi Ayan, >>>> I found that Zeppelin 0.6.0 can’t set hive.metastore.warehouse.dir >>>> property in somehow. >>>> Because I change the hive-site.xml inside $ZEP_DIR/conf/hive-site.xml for >>>> other config it works for instance I changed hive.metastore.metadb.dir >>>> it’s ok but only for hive.metastore.warehouse.dir didn’t work. >>>> >>>> This is wired. >>>> >>>> >>>> >>>> >>>>> On Jul 13, 2016, at 4:53 PM, ayan guha <guha.a...@gmail.com >>>>> <mailto:guha.a...@gmail.com>> wrote: >>>>> >>>>> I would suggest you to restart zeppelin and STS. >>>>> >>>>> On Wed, Jul 13, 2016 at 6:35 PM, Chanh Le <giaosu...@gmail.com >>>>> <mailto:giaosu...@gmail.com>> wrote: >>>>> Hi Ayan, >>>>> >>>>> I don’t know I did something wrong but still couldn’t set >>>>> hive.metastore.warehouse.dir property. >>>>> >>>>> I set 3 hive-site.xml files in spark location, zeppelin, hive as well but >>>>> still didn’t work. >>>>> >>>>> zeppeline/conf/hive-site.xml >>>>> spark/conf/hive-site.xml >>>>> hive/conf/hive-site.xml >>>>> >>>>> <Screen Shot 2016-07-13 at 3.32.49 PM.png> >>>>> >>>>> My hive-site.xml >>>>> >>>>> <configuration> >>>>> <property> >>>>> <name>hive.metastore.metadb.dir</name> >>>>> <value>alluxio://master1:19998/metadb</value> <> >>>>> <description> >>>>> Required by metastore server or if the uris argument below is not >>>>> supplied >>>>> </description> >>>>> </property> >>>>> <property> >>>>> <name>hive.metastore.warehouse.dir</name> >>>>> <value>alluxio://master1:19998/warehouse</value> <> >>>>> <description> >>>>> Required by metastore server or if the uris argument below is not >>>>> supplied >>>>> </description> >>>>> </property> >>>>> </configuration> >>>>> >>>>> Is there anything I can do? >>>>> >>>>> Regards, >>>>> Chanh >>>>> >>>>> >>>>> >>>>>> On Jul 13, 2016, at 12:43 PM, ayan guha <guha.a...@gmail.com >>>>>> <mailto:guha.a...@gmail.com>> wrote: >>>>>> >>>>>> Hi >>>>>> >>>>>> Create table always happens through Hive. In Hive, when you create a >>>>>> database, the default metadata storage location is driven by >>>>>> hive.metastore.metadb.dir and data storage is driven by >>>>>> hive.metastore.warehouse.dir property (set in hive site xml). So, you do >>>>>> not need to set this property in Zeppelin. >>>>>> >>>>>> What you can do: >>>>>> a. Modify hive-site.xml to include those properties, if they >>>>>> are not already set. use the same hive site.xml to run STS. Then >>>>>> connect through JDBC, create table and you should find metadata & data >>>>>> in your desired location. >>>>>> b. I think you can set these properties (same way you'd do in hive cli) >>>>>> c. You can create tables/databases with a LOCATION clause, in case you >>>>>> need to use non-standard path. >>>>>> >>>>>> Best >>>>>> Ayan >>>>>> >>>>>> On Wed, Jul 13, 2016 at 3:20 PM, Chanh Le <giaosu...@gmail.com >>>>>> <mailto:giaosu...@gmail.com>> wrote: >>>>>> Hi Ayan, >>>>>> Thank you for replying. >>>>>> But I wanna create a table in Zeppelin and store the metadata in Alluxio >>>>>> like I tried to do set >>>>>> hive.metastore.warehouse.dir=alluxio://master1:19998/metadb <> <>So I >>>>>> can share data with STS. >>>>>> >>>>>> The way you’ve mentioned through JDBC I already did and it works but I >>>>>> can’t create table in Spark way easily. >>>>>> >>>>>> Regards, >>>>>> Chanh >>>>>> >>>>>> >>>>>>> On Jul 13, 2016, at 12:06 PM, ayan guha <guha.a...@gmail.com >>>>>>> <mailto:guha.a...@gmail.com>> wrote: >>>>>>> >>>>>>> HI >>>>>>> >>>>>>> I quickly tried with available hive interpreter >>>>>>> >>>>>>> <image.png> >>>>>>> >>>>>>> Please try similarly. >>>>>>> >>>>>>> I will try with jdbc interpreter but I need to add it to zeppelin :) >>>>>>> >>>>>>> Best >>>>>>> Ayan >>>>>>> >>>>>>> On Wed, Jul 13, 2016 at 1:53 PM, Chanh Le <giaosu...@gmail.com >>>>>>> <mailto:giaosu...@gmail.com>> wrote: >>>>>>> Hi Ayan, >>>>>>> How to set hive metastore in Zeppelin. I tried but not success. >>>>>>> The way I do I add into Spark Interpreter >>>>>>> >>>>>>> <Screen Shot 2016-07-13 at 10.50.53 AM.png> >>>>>>> >>>>>>> And also try in a notebook by >>>>>>> %sql >>>>>>> set hive.metastore.metadb.dir=alluxio://master1:19998/metadb <> >>>>>>> >>>>>>> %sql >>>>>>> set hive.metastore.warehouse.dir=alluxio://master1:19998/metadb <> >>>>>>> >>>>>>> %spark >>>>>>> sqlContext.setConf("hive.metastore.warehouse.dir", >>>>>>> "alluxio://master1:19998/metadb <>") >>>>>>> sqlContext.setConf("hive.metastore.metadb.dir", >>>>>>> "alluxio://master1:19998/metadb <>") >>>>>>> sqlContext.read.parquet("alluxio://master1:19998/etl_info/WEBSITE >>>>>>> <>").saveAsTable("tests_5”) >>>>>>> >>>>>>> But It’s <Screen Shot 2016-07-13 at 10.53.10 AM.png> >>>>>>> >>>>>>>> On Jul 11, 2016, at 1:26 PM, ayan guha <guha.a...@gmail.com >>>>>>>> <mailto:guha.a...@gmail.com>> wrote: >>>>>>>> >>>>>>>> Hi >>>>>>>> >>>>>>>> When you say "Zeppelin and STS", I am assuming you mean "Spark >>>>>>>> Interpreter" and "JDBC interpreter" respectively. >>>>>>>> >>>>>>>> Through Zeppelin, you can either run your own spark application (by >>>>>>>> using Zeppelin's own spark context) using spark interpreter OR you can >>>>>>>> access STS, which is a spark application ie separate Spark Context >>>>>>>> using JDBC interpreter. There should not be any need for these 2 >>>>>>>> contexts to coexist. >>>>>>>> >>>>>>>> If you want to share data, save it to hive from either context, and >>>>>>>> you should be able to see the data from other context. >>>>>>>> >>>>>>>> Best >>>>>>>> Ayan >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> On Mon, Jul 11, 2016 at 3:00 PM, Chanh Le <giaosu...@gmail.com >>>>>>>> <mailto:giaosu...@gmail.com>> wrote: >>>>>>>> Hi Ayan, >>>>>>>> I tested It works fine but one more confuse is If my (technical) users >>>>>>>> want to write some code in zeppelin to apply thing into Hive table? >>>>>>>> Zeppelin and STS can’t share Spark Context that mean we need separated >>>>>>>> process? Is there anyway to use the same Spark Context of STS? >>>>>>>> >>>>>>>> Regards, >>>>>>>> Chanh >>>>>>>> >>>>>>>> >>>>>>>>> On Jul 11, 2016, at 10:05 AM, Takeshi Yamamuro <linguin....@gmail.com >>>>>>>>> <mailto:linguin....@gmail.com>> wrote: >>>>>>>>> >>>>>>>>> Hi, >>>>>>>>> >>>>>>>>> ISTM multiple sparkcontexts are not recommended in spark. >>>>>>>>> See: https://issues.apache.org/jira/browse/SPARK-2243 >>>>>>>>> <https://issues.apache.org/jira/browse/SPARK-2243> >>>>>>>>> >>>>>>>>> // maropu >>>>>>>>> >>>>>>>>> >>>>>>>>> On Mon, Jul 11, 2016 at 12:01 PM, ayan guha <guha.a...@gmail.com >>>>>>>>> <mailto:guha.a...@gmail.com>> wrote: >>>>>>>>> Hi >>>>>>>>> >>>>>>>>> Can you try using JDBC interpreter with STS? We are using >>>>>>>>> Zeppelin+STS on YARN for few months now without much issue. >>>>>>>>> >>>>>>>>> On Mon, Jul 11, 2016 at 12:48 PM, Chanh Le <giaosu...@gmail.com >>>>>>>>> <mailto:giaosu...@gmail.com>> wrote: >>>>>>>>> Hi everybody, >>>>>>>>> We are using Spark to query big data and currently we’re using >>>>>>>>> Zeppelin to provide a UI for technical users. >>>>>>>>> Now we also need to provide a UI for business users so we use Oracle >>>>>>>>> BI tools and set up a Spark Thrift Server (STS) for it. >>>>>>>>> >>>>>>>>> When I run both Zeppelin and STS throw error: >>>>>>>>> >>>>>>>>> INFO [2016-07-11 09:40:21,905] ({pool-2-thread-4} >>>>>>>>> SchedulerFactory.java[jobStarted]:131) - Job >>>>>>>>> remoteInterpretJob_1468204821905 started by scheduler >>>>>>>>> org.apache.zeppelin.spark.SparkInterpreter835015739 >>>>>>>>> INFO [2016-07-11 09:40:21,911] ({pool-2-thread-4} >>>>>>>>> Logging.scala[logInfo]:58) - Changing view acls to: giaosudau >>>>>>>>> INFO [2016-07-11 09:40:21,912] ({pool-2-thread-4} >>>>>>>>> Logging.scala[logInfo]:58) - Changing modify acls to: giaosudau >>>>>>>>> INFO [2016-07-11 09:40:21,912] ({pool-2-thread-4} >>>>>>>>> Logging.scala[logInfo]:58) - SecurityManager: authentication >>>>>>>>> disabled; ui acls disabled; users with view permissions: >>>>>>>>> Set(giaosudau); users with modify permissions: Set(giaosudau) >>>>>>>>> INFO [2016-07-11 09:40:21,918] ({pool-2-thread-4} >>>>>>>>> Logging.scala[logInfo]:58) - Starting HTTP Server >>>>>>>>> INFO [2016-07-11 09:40:21,919] ({pool-2-thread-4} >>>>>>>>> Server.java[doStart]:272) - jetty-8.y.z-SNAPSHOT >>>>>>>>> INFO [2016-07-11 09:40:21,920] ({pool-2-thread-4} >>>>>>>>> AbstractConnector.java[doStart]:338) - Started >>>>>>>>> SocketConnector@0.0.0.0:54818 <http://SocketConnector@0.0.0.0:54818/> >>>>>>>>> INFO [2016-07-11 09:40:21,922] ({pool-2-thread-4} >>>>>>>>> Logging.scala[logInfo]:58) - Successfully started service 'HTTP class >>>>>>>>> server' on port 54818. >>>>>>>>> INFO [2016-07-11 09:40:22,408] ({pool-2-thread-4} >>>>>>>>> SparkInterpreter.java[createSparkContext]:233) - ------ Create new >>>>>>>>> SparkContext local[*] ------- >>>>>>>>> WARN [2016-07-11 09:40:22,411] ({pool-2-thread-4} >>>>>>>>> Logging.scala[logWarning]:70) - Another SparkContext is being >>>>>>>>> constructed (or threw an exception in its constructor). This may >>>>>>>>> indicate an error, since only one SparkContext may be running in this >>>>>>>>> JVM (see SPARK-2243). The other SparkContext was created at: >>>>>>>>> >>>>>>>>> Is that mean I need to setup allow multiple context? Because It’s >>>>>>>>> only test in local with local mode If I deploy on mesos cluster what >>>>>>>>> would happened? >>>>>>>>> >>>>>>>>> Need you guys suggests some solutions for that. Thanks. >>>>>>>>> >>>>>>>>> Chanh >>>>>>>>> --------------------------------------------------------------------- >>>>>>>>> To unsubscribe e-mail: user-unsubscr...@spark.apache.org >>>>>>>>> <mailto:user-unsubscr...@spark.apache.org> >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> -- >>>>>>>>> Best Regards, >>>>>>>>> Ayan Guha >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> -- >>>>>>>>> --- >>>>>>>>> Takeshi Yamamuro >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> -- >>>>>>>> Best Regards, >>>>>>>> Ayan Guha >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> -- >>>>>>> Best Regards, >>>>>>> Ayan Guha >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> -- >>>>>> Best Regards, >>>>>> Ayan Guha >>>>> >>>>> >>>>> >>>>> >>>>> -- >>>>> Best Regards, >>>>> Ayan Guha >>>> >>>> >>>> >>>> >>>> -- >>>> Best Regards, >>>> Ayan Guha >>> >> >> >> >> >> -- >> Best Regards, >> Ayan Guha > > > > > -- > Best Regards, > Ayan Guha