When you run it in local mode ^^ Thanks Best Regards
On Thu, Mar 26, 2015 at 2:06 PM, ÐΞ€ρ@Ҝ (๏̯͡๏) <deepuj...@gmail.com> wrote: > I don;t think thats correct. load data local should pick input from local > directory. > > On Thu, Mar 26, 2015 at 1:59 PM, Akhil Das <ak...@sigmoidanalytics.com> > wrote: > >> Not sure, but you can create that path in all workers and put that file >> in it. >> >> Thanks >> Best Regards >> >> On Thu, Mar 26, 2015 at 1:56 PM, ÐΞ€ρ@Ҝ (๏̯͡๏) <deepuj...@gmail.com> >> wrote: >> >>> The Hive command >>> >>> LOAD DATA LOCAL INPATH >>> '/home/dvasthimal/spark1.3/spark-1.3.0-bin-hadoop2.4/examples/src/main/resources/kv1.txt' >>> INTO TABLE src_spark >>> >>> 1. LOCAL INPATH. if i push to HDFS then how will it work ? >>> >>> 2. I cant use sc.addFile, cause i want to run Hive (Spark SQL) queries. >>> >>> On Thu, Mar 26, 2015 at 1:41 PM, Akhil Das <ak...@sigmoidanalytics.com> >>> wrote: >>> >>>> Now its clear that the workers are not having the file kv1.txt in their >>>> local filesystem. You can try putting that in hdfs and use the URI to that >>>> file or try adding the file with sc.addFile >>>> >>>> Thanks >>>> Best Regards >>>> >>>> On Thu, Mar 26, 2015 at 1:38 PM, ÐΞ€ρ@Ҝ (๏̯͡๏) <deepuj...@gmail.com> >>>> wrote: >>>> >>>>> Does not work >>>>> >>>>> 15/03/26 01:07:05 INFO HiveMetaStore.audit: ugi=dvasthimal >>>>> ip=unknown-ip-addr cmd=get_table : db=default tbl=src_spark >>>>> 15/03/26 01:07:06 ERROR ql.Driver: FAILED: SemanticException Line 1:23 >>>>> Invalid path >>>>> ''/home/dvasthimal/spark1.3/spark-1.3.0-bin-hadoop2.4/examples/src/main/resources/kv1.txt'': >>>>> No files matching path >>>>> file:/home/dvasthimal/spark1.3/spark-1.3.0-bin-hadoop2.4/examples/src/main/resources/kv1.txt >>>>> org.apache.hadoop.hive.ql.parse.SemanticException: Line 1:23 Invalid >>>>> path >>>>> ''/home/dvasthimal/spark1.3/spark-1.3.0-bin-hadoop2.4/examples/src/main/resources/kv1.txt'': >>>>> No files matching path >>>>> file:/home/dvasthimal/spark1.3/spark-1.3.0-bin-hadoop2.4/examples/src/main/resources/kv1.txt >>>>> at >>>>> org.apache.hadoop.hive.ql.parse.LoadSemanticAnalyzer.applyConstraints(LoadSemanticAnalyzer.java:142) >>>>> at >>>>> org.apache.hadoop.hive.ql.parse.LoadSemanticAnalyzer.analyzeInternal(LoadSemanticAnalyzer.java:233) >>>>> at >>>>> org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:327) >>>>> at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:422) >>>>> >>>>> >>>>> >>>>> Does the input file needs to be passed to executor via -- jars ? >>>>> >>>>> On Thu, Mar 26, 2015 at 12:15 PM, Akhil Das < >>>>> ak...@sigmoidanalytics.com> wrote: >>>>> >>>>>> Try to give the complete path to the file kv1.txt. >>>>>> On 26 Mar 2015 11:48, "ÐΞ€ρ@Ҝ (๏̯͡๏)" <deepuj...@gmail.com> wrote: >>>>>> >>>>>>> I am now seeing this error. >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> 15/03/25 19:44:03 ERROR yarn.ApplicationMaster: User class threw >>>>>>> exception: FAILED: SemanticException Line 1:23 Invalid path >>>>>>> ''examples/src/main/resources/kv1.txt'': No files matching path >>>>>>> file:/hadoop/10/scratch/local/usercache/dvasthimal/appcache/application_1426715280024_89893/container_1426715280024_89893_01_000002/examples/src/main/resources/kv1.txt >>>>>>> >>>>>>> org.apache.spark.sql.execution.QueryExecutionException: FAILED: >>>>>>> SemanticException Line 1:23 Invalid path >>>>>>> ''examples/src/main/resources/kv1.txt'': No files matching path >>>>>>> file:/hadoop/10/scratch/local/usercache/dvasthimal/appcache/application_1426715280024_89893/container_1426715280024_89893_01_000002/examples/src/main/resources/kv1.txt >>>>>>> >>>>>>> at >>>>>>> org.apache.spark.sql.hive.HiveContext.runHive(HiveContext.scala:312) >>>>>>> >>>>>>> at >>>>>>> org.apache.spark.sql.hive.HiveContext.runSqlHive(HiveContext.scala:280) >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> -sh-4.1$ pwd >>>>>>> >>>>>>> /home/dvasthimal/spark1.3/spark-1.3.0-bin-hadoop2.4 >>>>>>> >>>>>>> -sh-4.1$ ls examples/src/main/resources/kv1.txt >>>>>>> >>>>>>> examples/src/main/resources/kv1.txt >>>>>>> >>>>>>> -sh-4.1$ >>>>>>> >>>>>>> >>>>>>> >>>>>>> On Thu, Mar 26, 2015 at 8:08 AM, Zhan Zhang <zzh...@hortonworks.com> >>>>>>> wrote: >>>>>>> >>>>>>>> You can do it in $SPARK_HOME/conf/spark-defaults.con >>>>>>>> >>>>>>>> spark.driver.extraJavaOptions -XX:MaxPermSize=512m >>>>>>>> >>>>>>>> Thanks. >>>>>>>> >>>>>>>> Zhan Zhang >>>>>>>> >>>>>>>> >>>>>>>> On Mar 25, 2015, at 7:25 PM, ÐΞ€ρ@Ҝ (๏̯͡๏) <deepuj...@gmail.com> >>>>>>>> wrote: >>>>>>>> >>>>>>>> Where and how do i pass this or other JVM argument ? >>>>>>>> -XX:MaxPermSize=512m >>>>>>>> >>>>>>>> On Wed, Mar 25, 2015 at 11:36 PM, Zhan Zhang < >>>>>>>> zzh...@hortonworks.com> wrote: >>>>>>>> >>>>>>>>> I solve this by increase the PermGen memory size in driver. >>>>>>>>> >>>>>>>>> -XX:MaxPermSize=512m >>>>>>>>> >>>>>>>>> Thanks. >>>>>>>>> >>>>>>>>> Zhan Zhang >>>>>>>>> >>>>>>>>> On Mar 25, 2015, at 10:54 AM, ÐΞ€ρ@Ҝ (๏̯͡๏) <deepuj...@gmail.com> >>>>>>>>> wrote: >>>>>>>>> >>>>>>>>> I am facing same issue, posted a new thread. Please respond. >>>>>>>>> >>>>>>>>> On Wed, Jan 14, 2015 at 4:38 AM, Zhan Zhang < >>>>>>>>> zzh...@hortonworks.com> wrote: >>>>>>>>> >>>>>>>>>> Hi Folks, >>>>>>>>>> >>>>>>>>>> I am trying to run hive context in yarn-cluster mode, but met >>>>>>>>>> some error. Does anybody know what cause the issue. >>>>>>>>>> >>>>>>>>>> I use following cmd to build the distribution: >>>>>>>>>> >>>>>>>>>> ./make-distribution.sh -Phive -Phive-thriftserver -Pyarn >>>>>>>>>> -Phadoop-2.4 >>>>>>>>>> >>>>>>>>>> 15/01/13 17:59:42 INFO cluster.YarnClusterScheduler: >>>>>>>>>> YarnClusterScheduler.postStartHook done >>>>>>>>>> 15/01/13 17:59:42 INFO storage.BlockManagerMasterActor: >>>>>>>>>> Registering block manager cn122-10.l42scl.hortonworks.com:56157 >>>>>>>>>> with 1589.8 MB RAM, BlockManagerId(2, >>>>>>>>>> cn122-10.l42scl.hortonworks.com, 56157) >>>>>>>>>> 15/01/13 17:59:43 INFO parse.ParseDriver: Parsing command: CREATE >>>>>>>>>> TABLE IF NOT EXISTS src (key INT, value STRING) >>>>>>>>>> 15/01/13 17:59:43 INFO parse.ParseDriver: Parse Completed >>>>>>>>>> 15/01/13 17:59:44 INFO metastore.HiveMetaStore: 0: Opening raw >>>>>>>>>> store with implemenation >>>>>>>>>> class:org.apache.hadoop.hive.metastore.ObjectStore >>>>>>>>>> 15/01/13 17:59:44 INFO metastore.ObjectStore: ObjectStore, >>>>>>>>>> initialize called >>>>>>>>>> 15/01/13 17:59:44 INFO DataNucleus.Persistence: Property >>>>>>>>>> datanucleus.cache.level2 unknown - will be ignored >>>>>>>>>> 15/01/13 17:59:44 INFO DataNucleus.Persistence: Property >>>>>>>>>> hive.metastore.integral.jdo.pushdown unknown - will be ignored >>>>>>>>>> 15/01/13 17:59:44 WARN DataNucleus.Connection: BoneCP specified >>>>>>>>>> but not present in CLASSPATH (or one of dependencies) >>>>>>>>>> 15/01/13 17:59:44 WARN DataNucleus.Connection: BoneCP specified >>>>>>>>>> but not present in CLASSPATH (or one of dependencies) >>>>>>>>>> 15/01/13 17:59:52 INFO metastore.ObjectStore: Setting MetaStore >>>>>>>>>> object pin classes with >>>>>>>>>> hive.metastore.cache.pinobjtypes="Table,StorageDescriptor,SerDeInfo,Partition,Database,Type,FieldSchema,Order" >>>>>>>>>> 15/01/13 17:59:52 INFO metastore.MetaStoreDirectSql: MySQL check >>>>>>>>>> failed, assuming we are not on mysql: Lexical error at line 1, >>>>>>>>>> column 5. >>>>>>>>>> Encountered: "@" (64), after : "". >>>>>>>>>> 15/01/13 17:59:53 INFO DataNucleus.Datastore: The class >>>>>>>>>> "org.apache.hadoop.hive.metastore.model.MFieldSchema" is tagged as >>>>>>>>>> "embedded-only" so does not have its own datastore table. >>>>>>>>>> 15/01/13 17:59:53 INFO DataNucleus.Datastore: The class >>>>>>>>>> "org.apache.hadoop.hive.metastore.model.MOrder" is tagged as >>>>>>>>>> "embedded-only" so does not have its own datastore table. >>>>>>>>>> 15/01/13 17:59:59 INFO DataNucleus.Datastore: The class >>>>>>>>>> "org.apache.hadoop.hive.metastore.model.MFieldSchema" is tagged as >>>>>>>>>> "embedded-only" so does not have its own datastore table. >>>>>>>>>> 15/01/13 17:59:59 INFO DataNucleus.Datastore: The class >>>>>>>>>> "org.apache.hadoop.hive.metastore.model.MOrder" is tagged as >>>>>>>>>> "embedded-only" so does not have its own datastore table. >>>>>>>>>> 15/01/13 18:00:00 INFO metastore.ObjectStore: Initialized >>>>>>>>>> ObjectStore >>>>>>>>>> 15/01/13 18:00:00 WARN metastore.ObjectStore: Version information >>>>>>>>>> not found in metastore. hive.metastore.schema.verification is not >>>>>>>>>> enabled >>>>>>>>>> so recording the schema version 0.13.1aa >>>>>>>>>> 15/01/13 18:00:01 INFO metastore.HiveMetaStore: Added admin role >>>>>>>>>> in metastore >>>>>>>>>> 15/01/13 18:00:01 INFO metastore.HiveMetaStore: Added public role >>>>>>>>>> in metastore >>>>>>>>>> 15/01/13 18:00:01 INFO metastore.HiveMetaStore: No user is added >>>>>>>>>> in admin role, since config is empty >>>>>>>>>> 15/01/13 18:00:01 INFO session.SessionState: No Tez session >>>>>>>>>> required at this point. hive.execution.engine=mr. >>>>>>>>>> 15/01/13 18:00:02 INFO log.PerfLogger: <PERFLOG method=Driver.run >>>>>>>>>> from=org.apache.hadoop.hive.ql.Driver> >>>>>>>>>> 15/01/13 18:00:02 INFO log.PerfLogger: <PERFLOG >>>>>>>>>> method=TimeToSubmit from=org.apache.hadoop.hive.ql.Driver> >>>>>>>>>> 15/01/13 18:00:02 INFO ql.Driver: Concurrency mode is disabled, >>>>>>>>>> not creating a lock manager >>>>>>>>>> 15/01/13 18:00:02 INFO log.PerfLogger: <PERFLOG method=compile >>>>>>>>>> from=org.apache.hadoop.hive.ql.Driver> >>>>>>>>>> 15/01/13 18:00:03 INFO log.PerfLogger: <PERFLOG method=parse >>>>>>>>>> from=org.apache.hadoop.hive.ql.Driver> >>>>>>>>>> 15/01/13 18:00:03 INFO parse.ParseDriver: Parsing command: CREATE >>>>>>>>>> TABLE IF NOT EXISTS src (key INT, value STRING) >>>>>>>>>> 15/01/13 18:00:03 INFO parse.ParseDriver: Parse Completed >>>>>>>>>> 15/01/13 18:00:03 INFO log.PerfLogger: </PERFLOG method=parse >>>>>>>>>> start=1421190003030 end=1421190003031 duration=1 >>>>>>>>>> from=org.apache.hadoop.hive.ql.Driver> >>>>>>>>>> 15/01/13 18:00:03 INFO log.PerfLogger: <PERFLOG >>>>>>>>>> method=semanticAnalyze from=org.apache.hadoop.hive.ql.Driver> >>>>>>>>>> 15/01/13 18:00:03 INFO parse.SemanticAnalyzer: Starting Semantic >>>>>>>>>> Analysis >>>>>>>>>> 15/01/13 18:00:03 INFO parse.SemanticAnalyzer: Creating table src >>>>>>>>>> position=27 >>>>>>>>>> 15/01/13 18:00:03 INFO metastore.HiveMetaStore: 0: get_table : >>>>>>>>>> db=default tbl=src >>>>>>>>>> 15/01/13 18:00:03 INFO HiveMetaStore.audit: ugi=zzhang >>>>>>>>>> ip=unknown-ip-addr cmd=get_table : db=default tbl=src >>>>>>>>>> 15/01/13 18:00:03 INFO metastore.HiveMetaStore: 0: get_database: >>>>>>>>>> default >>>>>>>>>> 15/01/13 18:00:03 INFO HiveMetaStore.audit: ugi=zzhang >>>>>>>>>> ip=unknown-ip-addr cmd=get_database: default >>>>>>>>>> 15/01/13 18:00:03 INFO ql.Driver: Semantic Analysis Completed >>>>>>>>>> 15/01/13 18:00:03 INFO log.PerfLogger: </PERFLOG >>>>>>>>>> method=semanticAnalyze start=1421190003031 end=1421190003406 >>>>>>>>>> duration=375 >>>>>>>>>> from=org.apache.hadoop.hive.ql.Driver> >>>>>>>>>> 15/01/13 18:00:03 INFO ql.Driver: Returning Hive schema: >>>>>>>>>> Schema(fieldSchemas:null, properties:null) >>>>>>>>>> 15/01/13 18:00:03 INFO log.PerfLogger: </PERFLOG method=compile >>>>>>>>>> start=1421190002998 end=1421190003416 duration=418 >>>>>>>>>> from=org.apache.hadoop.hive.ql.Driver> >>>>>>>>>> 15/01/13 18:00:03 INFO log.PerfLogger: <PERFLOG >>>>>>>>>> method=Driver.execute from=org.apache.hadoop.hive.ql.Driver> >>>>>>>>>> 15/01/13 18:00:03 INFO ql.Driver: Starting command: CREATE TABLE >>>>>>>>>> IF NOT EXISTS src (key INT, value STRING) >>>>>>>>>> 15/01/13 18:00:03 INFO log.PerfLogger: </PERFLOG >>>>>>>>>> method=TimeToSubmit start=1421190002995 end=1421190003421 >>>>>>>>>> duration=426 >>>>>>>>>> from=org.apache.hadoop.hive.ql.Driver> >>>>>>>>>> 15/01/13 18:00:03 INFO log.PerfLogger: <PERFLOG method=runTasks >>>>>>>>>> from=org.apache.hadoop.hive.ql.Driver> >>>>>>>>>> 15/01/13 18:00:03 INFO log.PerfLogger: <PERFLOG >>>>>>>>>> method=task.DDL.Stage-0 from=org.apache.hadoop.hive.ql.Driver> >>>>>>>>>> 15/01/13 18:00:03 INFO exec.DDLTask: Default to LazySimpleSerDe >>>>>>>>>> for table src >>>>>>>>>> 15/01/13 18:00:05 INFO log.PerfLogger: </PERFLOG >>>>>>>>>> method=Driver.execute start=1421190003416 end=1421190005498 >>>>>>>>>> duration=2082 >>>>>>>>>> from=org.apache.hadoop.hive.ql.Driver> >>>>>>>>>> Exception in thread "Driver" >>>>>>>>>> Exception: java.lang.OutOfMemoryError thrown from the >>>>>>>>>> UncaughtExceptionHandler in thread "Driver" >>>>>>>>>> -- >>>>>>>>>> CONFIDENTIALITY NOTICE >>>>>>>>>> NOTICE: This message is intended for the use of the individual or >>>>>>>>>> entity to >>>>>>>>>> which it is addressed and may contain information that is >>>>>>>>>> confidential, >>>>>>>>>> privileged and exempt from disclosure under applicable law. If >>>>>>>>>> the reader >>>>>>>>>> of this message is not the intended recipient, you are hereby >>>>>>>>>> notified that >>>>>>>>>> any printing, copying, dissemination, distribution, disclosure or >>>>>>>>>> forwarding of this communication is strictly prohibited. If you >>>>>>>>>> have >>>>>>>>>> received this communication in error, please contact the sender >>>>>>>>>> immediately >>>>>>>>>> and delete it from your system. Thank You. >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> --------------------------------------------------------------------- >>>>>>>>>> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org >>>>>>>>>> For additional commands, e-mail: user-h...@spark.apache.org >>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> -- >>>>>>>>> Deepak >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> -- >>>>>>>> Deepak >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>> >>>>>>> >>>>>>> -- >>>>>>> Deepak >>>>>>> >>>>>>> >>>>> >>>>> >>>>> -- >>>>> Deepak >>>>> >>>>> >>>> >>> >>> >>> -- >>> Deepak >>> >>> >> > > > -- > Deepak > >