Could you try putting that file in hdfs and try like: LOAD DATA INPATH 'hdfs://sigmoid/test/kv1.txt' INTO TABLE src_spark
Thanks Best Regards On Thu, Mar 26, 2015 at 2:07 PM, Akhil Das <ak...@sigmoidanalytics.com> wrote: > When you run it in local mode ^^ > > Thanks > Best Regards > > On Thu, Mar 26, 2015 at 2:06 PM, ÐΞ€ρ@Ҝ (๏̯͡๏) <deepuj...@gmail.com> > wrote: > >> I don;t think thats correct. load data local should pick input from local >> directory. >> >> On Thu, Mar 26, 2015 at 1:59 PM, Akhil Das <ak...@sigmoidanalytics.com> >> wrote: >> >>> Not sure, but you can create that path in all workers and put that file >>> in it. >>> >>> Thanks >>> Best Regards >>> >>> On Thu, Mar 26, 2015 at 1:56 PM, ÐΞ€ρ@Ҝ (๏̯͡๏) <deepuj...@gmail.com> >>> wrote: >>> >>>> The Hive command >>>> >>>> LOAD DATA LOCAL INPATH >>>> '/home/dvasthimal/spark1.3/spark-1.3.0-bin-hadoop2.4/examples/src/main/resources/kv1.txt' >>>> INTO TABLE src_spark >>>> >>>> 1. LOCAL INPATH. if i push to HDFS then how will it work ? >>>> >>>> 2. I cant use sc.addFile, cause i want to run Hive (Spark SQL) queries. >>>> >>>> On Thu, Mar 26, 2015 at 1:41 PM, Akhil Das <ak...@sigmoidanalytics.com> >>>> wrote: >>>> >>>>> Now its clear that the workers are not having the file kv1.txt in >>>>> their local filesystem. You can try putting that in hdfs and use the URI >>>>> to >>>>> that file or try adding the file with sc.addFile >>>>> >>>>> Thanks >>>>> Best Regards >>>>> >>>>> On Thu, Mar 26, 2015 at 1:38 PM, ÐΞ€ρ@Ҝ (๏̯͡๏) <deepuj...@gmail.com> >>>>> wrote: >>>>> >>>>>> Does not work >>>>>> >>>>>> 15/03/26 01:07:05 INFO HiveMetaStore.audit: ugi=dvasthimal >>>>>> ip=unknown-ip-addr cmd=get_table : db=default tbl=src_spark >>>>>> 15/03/26 01:07:06 ERROR ql.Driver: FAILED: SemanticException Line >>>>>> 1:23 Invalid path >>>>>> ''/home/dvasthimal/spark1.3/spark-1.3.0-bin-hadoop2.4/examples/src/main/resources/kv1.txt'': >>>>>> No files matching path >>>>>> file:/home/dvasthimal/spark1.3/spark-1.3.0-bin-hadoop2.4/examples/src/main/resources/kv1.txt >>>>>> org.apache.hadoop.hive.ql.parse.SemanticException: Line 1:23 Invalid >>>>>> path >>>>>> ''/home/dvasthimal/spark1.3/spark-1.3.0-bin-hadoop2.4/examples/src/main/resources/kv1.txt'': >>>>>> No files matching path >>>>>> file:/home/dvasthimal/spark1.3/spark-1.3.0-bin-hadoop2.4/examples/src/main/resources/kv1.txt >>>>>> at >>>>>> org.apache.hadoop.hive.ql.parse.LoadSemanticAnalyzer.applyConstraints(LoadSemanticAnalyzer.java:142) >>>>>> at >>>>>> org.apache.hadoop.hive.ql.parse.LoadSemanticAnalyzer.analyzeInternal(LoadSemanticAnalyzer.java:233) >>>>>> at >>>>>> org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:327) >>>>>> at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:422) >>>>>> >>>>>> >>>>>> >>>>>> Does the input file needs to be passed to executor via -- jars ? >>>>>> >>>>>> On Thu, Mar 26, 2015 at 12:15 PM, Akhil Das < >>>>>> ak...@sigmoidanalytics.com> wrote: >>>>>> >>>>>>> Try to give the complete path to the file kv1.txt. >>>>>>> On 26 Mar 2015 11:48, "ÐΞ€ρ@Ҝ (๏̯͡๏)" <deepuj...@gmail.com> wrote: >>>>>>> >>>>>>>> I am now seeing this error. >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> 15/03/25 19:44:03 ERROR yarn.ApplicationMaster: User class threw >>>>>>>> exception: FAILED: SemanticException Line 1:23 Invalid path >>>>>>>> ''examples/src/main/resources/kv1.txt'': No files matching path >>>>>>>> file:/hadoop/10/scratch/local/usercache/dvasthimal/appcache/application_1426715280024_89893/container_1426715280024_89893_01_000002/examples/src/main/resources/kv1.txt >>>>>>>> >>>>>>>> org.apache.spark.sql.execution.QueryExecutionException: FAILED: >>>>>>>> SemanticException Line 1:23 Invalid path >>>>>>>> ''examples/src/main/resources/kv1.txt'': No files matching path >>>>>>>> file:/hadoop/10/scratch/local/usercache/dvasthimal/appcache/application_1426715280024_89893/container_1426715280024_89893_01_000002/examples/src/main/resources/kv1.txt >>>>>>>> >>>>>>>> at >>>>>>>> org.apache.spark.sql.hive.HiveContext.runHive(HiveContext.scala:312) >>>>>>>> >>>>>>>> at >>>>>>>> org.apache.spark.sql.hive.HiveContext.runSqlHive(HiveContext.scala:280) >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> -sh-4.1$ pwd >>>>>>>> >>>>>>>> /home/dvasthimal/spark1.3/spark-1.3.0-bin-hadoop2.4 >>>>>>>> >>>>>>>> -sh-4.1$ ls examples/src/main/resources/kv1.txt >>>>>>>> >>>>>>>> examples/src/main/resources/kv1.txt >>>>>>>> >>>>>>>> -sh-4.1$ >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> On Thu, Mar 26, 2015 at 8:08 AM, Zhan Zhang <zzh...@hortonworks.com >>>>>>>> > wrote: >>>>>>>> >>>>>>>>> You can do it in $SPARK_HOME/conf/spark-defaults.con >>>>>>>>> >>>>>>>>> spark.driver.extraJavaOptions -XX:MaxPermSize=512m >>>>>>>>> >>>>>>>>> Thanks. >>>>>>>>> >>>>>>>>> Zhan Zhang >>>>>>>>> >>>>>>>>> >>>>>>>>> On Mar 25, 2015, at 7:25 PM, ÐΞ€ρ@Ҝ (๏̯͡๏) <deepuj...@gmail.com> >>>>>>>>> wrote: >>>>>>>>> >>>>>>>>> Where and how do i pass this or other JVM argument ? >>>>>>>>> -XX:MaxPermSize=512m >>>>>>>>> >>>>>>>>> On Wed, Mar 25, 2015 at 11:36 PM, Zhan Zhang < >>>>>>>>> zzh...@hortonworks.com> wrote: >>>>>>>>> >>>>>>>>>> I solve this by increase the PermGen memory size in driver. >>>>>>>>>> >>>>>>>>>> -XX:MaxPermSize=512m >>>>>>>>>> >>>>>>>>>> Thanks. >>>>>>>>>> >>>>>>>>>> Zhan Zhang >>>>>>>>>> >>>>>>>>>> On Mar 25, 2015, at 10:54 AM, ÐΞ€ρ@Ҝ (๏̯͡๏) <deepuj...@gmail.com> >>>>>>>>>> wrote: >>>>>>>>>> >>>>>>>>>> I am facing same issue, posted a new thread. Please respond. >>>>>>>>>> >>>>>>>>>> On Wed, Jan 14, 2015 at 4:38 AM, Zhan Zhang < >>>>>>>>>> zzh...@hortonworks.com> wrote: >>>>>>>>>> >>>>>>>>>>> Hi Folks, >>>>>>>>>>> >>>>>>>>>>> I am trying to run hive context in yarn-cluster mode, but met >>>>>>>>>>> some error. Does anybody know what cause the issue. >>>>>>>>>>> >>>>>>>>>>> I use following cmd to build the distribution: >>>>>>>>>>> >>>>>>>>>>> ./make-distribution.sh -Phive -Phive-thriftserver -Pyarn >>>>>>>>>>> -Phadoop-2.4 >>>>>>>>>>> >>>>>>>>>>> 15/01/13 17:59:42 INFO cluster.YarnClusterScheduler: >>>>>>>>>>> YarnClusterScheduler.postStartHook done >>>>>>>>>>> 15/01/13 17:59:42 INFO storage.BlockManagerMasterActor: >>>>>>>>>>> Registering block manager cn122-10.l42scl.hortonworks.com:56157 >>>>>>>>>>> with 1589.8 MB RAM, BlockManagerId(2, >>>>>>>>>>> cn122-10.l42scl.hortonworks.com, 56157) >>>>>>>>>>> 15/01/13 17:59:43 INFO parse.ParseDriver: Parsing command: >>>>>>>>>>> CREATE TABLE IF NOT EXISTS src (key INT, value STRING) >>>>>>>>>>> 15/01/13 17:59:43 INFO parse.ParseDriver: Parse Completed >>>>>>>>>>> 15/01/13 17:59:44 INFO metastore.HiveMetaStore: 0: Opening raw >>>>>>>>>>> store with implemenation >>>>>>>>>>> class:org.apache.hadoop.hive.metastore.ObjectStore >>>>>>>>>>> 15/01/13 17:59:44 INFO metastore.ObjectStore: ObjectStore, >>>>>>>>>>> initialize called >>>>>>>>>>> 15/01/13 17:59:44 INFO DataNucleus.Persistence: Property >>>>>>>>>>> datanucleus.cache.level2 unknown - will be ignored >>>>>>>>>>> 15/01/13 17:59:44 INFO DataNucleus.Persistence: Property >>>>>>>>>>> hive.metastore.integral.jdo.pushdown unknown - will be ignored >>>>>>>>>>> 15/01/13 17:59:44 WARN DataNucleus.Connection: BoneCP specified >>>>>>>>>>> but not present in CLASSPATH (or one of dependencies) >>>>>>>>>>> 15/01/13 17:59:44 WARN DataNucleus.Connection: BoneCP specified >>>>>>>>>>> but not present in CLASSPATH (or one of dependencies) >>>>>>>>>>> 15/01/13 17:59:52 INFO metastore.ObjectStore: Setting MetaStore >>>>>>>>>>> object pin classes with >>>>>>>>>>> hive.metastore.cache.pinobjtypes="Table,StorageDescriptor,SerDeInfo,Partition,Database,Type,FieldSchema,Order" >>>>>>>>>>> 15/01/13 17:59:52 INFO metastore.MetaStoreDirectSql: MySQL check >>>>>>>>>>> failed, assuming we are not on mysql: Lexical error at line 1, >>>>>>>>>>> column 5. >>>>>>>>>>> Encountered: "@" (64), after : "". >>>>>>>>>>> 15/01/13 17:59:53 INFO DataNucleus.Datastore: The class >>>>>>>>>>> "org.apache.hadoop.hive.metastore.model.MFieldSchema" is tagged as >>>>>>>>>>> "embedded-only" so does not have its own datastore table. >>>>>>>>>>> 15/01/13 17:59:53 INFO DataNucleus.Datastore: The class >>>>>>>>>>> "org.apache.hadoop.hive.metastore.model.MOrder" is tagged as >>>>>>>>>>> "embedded-only" so does not have its own datastore table. >>>>>>>>>>> 15/01/13 17:59:59 INFO DataNucleus.Datastore: The class >>>>>>>>>>> "org.apache.hadoop.hive.metastore.model.MFieldSchema" is tagged as >>>>>>>>>>> "embedded-only" so does not have its own datastore table. >>>>>>>>>>> 15/01/13 17:59:59 INFO DataNucleus.Datastore: The class >>>>>>>>>>> "org.apache.hadoop.hive.metastore.model.MOrder" is tagged as >>>>>>>>>>> "embedded-only" so does not have its own datastore table. >>>>>>>>>>> 15/01/13 18:00:00 INFO metastore.ObjectStore: Initialized >>>>>>>>>>> ObjectStore >>>>>>>>>>> 15/01/13 18:00:00 WARN metastore.ObjectStore: Version >>>>>>>>>>> information not found in metastore. >>>>>>>>>>> hive.metastore.schema.verification is >>>>>>>>>>> not enabled so recording the schema version 0.13.1aa >>>>>>>>>>> 15/01/13 18:00:01 INFO metastore.HiveMetaStore: Added admin role >>>>>>>>>>> in metastore >>>>>>>>>>> 15/01/13 18:00:01 INFO metastore.HiveMetaStore: Added public >>>>>>>>>>> role in metastore >>>>>>>>>>> 15/01/13 18:00:01 INFO metastore.HiveMetaStore: No user is added >>>>>>>>>>> in admin role, since config is empty >>>>>>>>>>> 15/01/13 18:00:01 INFO session.SessionState: No Tez session >>>>>>>>>>> required at this point. hive.execution.engine=mr. >>>>>>>>>>> 15/01/13 18:00:02 INFO log.PerfLogger: <PERFLOG >>>>>>>>>>> method=Driver.run from=org.apache.hadoop.hive.ql.Driver> >>>>>>>>>>> 15/01/13 18:00:02 INFO log.PerfLogger: <PERFLOG >>>>>>>>>>> method=TimeToSubmit from=org.apache.hadoop.hive.ql.Driver> >>>>>>>>>>> 15/01/13 18:00:02 INFO ql.Driver: Concurrency mode is disabled, >>>>>>>>>>> not creating a lock manager >>>>>>>>>>> 15/01/13 18:00:02 INFO log.PerfLogger: <PERFLOG method=compile >>>>>>>>>>> from=org.apache.hadoop.hive.ql.Driver> >>>>>>>>>>> 15/01/13 18:00:03 INFO log.PerfLogger: <PERFLOG method=parse >>>>>>>>>>> from=org.apache.hadoop.hive.ql.Driver> >>>>>>>>>>> 15/01/13 18:00:03 INFO parse.ParseDriver: Parsing command: >>>>>>>>>>> CREATE TABLE IF NOT EXISTS src (key INT, value STRING) >>>>>>>>>>> 15/01/13 18:00:03 INFO parse.ParseDriver: Parse Completed >>>>>>>>>>> 15/01/13 18:00:03 INFO log.PerfLogger: </PERFLOG method=parse >>>>>>>>>>> start=1421190003030 end=1421190003031 duration=1 >>>>>>>>>>> from=org.apache.hadoop.hive.ql.Driver> >>>>>>>>>>> 15/01/13 18:00:03 INFO log.PerfLogger: <PERFLOG >>>>>>>>>>> method=semanticAnalyze from=org.apache.hadoop.hive.ql.Driver> >>>>>>>>>>> 15/01/13 18:00:03 INFO parse.SemanticAnalyzer: Starting Semantic >>>>>>>>>>> Analysis >>>>>>>>>>> 15/01/13 18:00:03 INFO parse.SemanticAnalyzer: Creating table >>>>>>>>>>> src position=27 >>>>>>>>>>> 15/01/13 18:00:03 INFO metastore.HiveMetaStore: 0: get_table : >>>>>>>>>>> db=default tbl=src >>>>>>>>>>> 15/01/13 18:00:03 INFO HiveMetaStore.audit: ugi=zzhang >>>>>>>>>>> ip=unknown-ip-addr cmd=get_table : db=default tbl=src >>>>>>>>>>> 15/01/13 18:00:03 INFO metastore.HiveMetaStore: 0: get_database: >>>>>>>>>>> default >>>>>>>>>>> 15/01/13 18:00:03 INFO HiveMetaStore.audit: ugi=zzhang >>>>>>>>>>> ip=unknown-ip-addr cmd=get_database: default >>>>>>>>>>> 15/01/13 18:00:03 INFO ql.Driver: Semantic Analysis Completed >>>>>>>>>>> 15/01/13 18:00:03 INFO log.PerfLogger: </PERFLOG >>>>>>>>>>> method=semanticAnalyze start=1421190003031 end=1421190003406 >>>>>>>>>>> duration=375 >>>>>>>>>>> from=org.apache.hadoop.hive.ql.Driver> >>>>>>>>>>> 15/01/13 18:00:03 INFO ql.Driver: Returning Hive schema: >>>>>>>>>>> Schema(fieldSchemas:null, properties:null) >>>>>>>>>>> 15/01/13 18:00:03 INFO log.PerfLogger: </PERFLOG method=compile >>>>>>>>>>> start=1421190002998 end=1421190003416 duration=418 >>>>>>>>>>> from=org.apache.hadoop.hive.ql.Driver> >>>>>>>>>>> 15/01/13 18:00:03 INFO log.PerfLogger: <PERFLOG >>>>>>>>>>> method=Driver.execute from=org.apache.hadoop.hive.ql.Driver> >>>>>>>>>>> 15/01/13 18:00:03 INFO ql.Driver: Starting command: CREATE TABLE >>>>>>>>>>> IF NOT EXISTS src (key INT, value STRING) >>>>>>>>>>> 15/01/13 18:00:03 INFO log.PerfLogger: </PERFLOG >>>>>>>>>>> method=TimeToSubmit start=1421190002995 end=1421190003421 >>>>>>>>>>> duration=426 >>>>>>>>>>> from=org.apache.hadoop.hive.ql.Driver> >>>>>>>>>>> 15/01/13 18:00:03 INFO log.PerfLogger: <PERFLOG method=runTasks >>>>>>>>>>> from=org.apache.hadoop.hive.ql.Driver> >>>>>>>>>>> 15/01/13 18:00:03 INFO log.PerfLogger: <PERFLOG >>>>>>>>>>> method=task.DDL.Stage-0 from=org.apache.hadoop.hive.ql.Driver> >>>>>>>>>>> 15/01/13 18:00:03 INFO exec.DDLTask: Default to LazySimpleSerDe >>>>>>>>>>> for table src >>>>>>>>>>> 15/01/13 18:00:05 INFO log.PerfLogger: </PERFLOG >>>>>>>>>>> method=Driver.execute start=1421190003416 end=1421190005498 >>>>>>>>>>> duration=2082 >>>>>>>>>>> from=org.apache.hadoop.hive.ql.Driver> >>>>>>>>>>> Exception in thread "Driver" >>>>>>>>>>> Exception: java.lang.OutOfMemoryError thrown from the >>>>>>>>>>> UncaughtExceptionHandler in thread "Driver" >>>>>>>>>>> -- >>>>>>>>>>> CONFIDENTIALITY NOTICE >>>>>>>>>>> NOTICE: This message is intended for the use of the individual >>>>>>>>>>> or entity to >>>>>>>>>>> which it is addressed and may contain information that is >>>>>>>>>>> confidential, >>>>>>>>>>> privileged and exempt from disclosure under applicable law. If >>>>>>>>>>> the reader >>>>>>>>>>> of this message is not the intended recipient, you are hereby >>>>>>>>>>> notified that >>>>>>>>>>> any printing, copying, dissemination, distribution, disclosure or >>>>>>>>>>> forwarding of this communication is strictly prohibited. If you >>>>>>>>>>> have >>>>>>>>>>> received this communication in error, please contact the sender >>>>>>>>>>> immediately >>>>>>>>>>> and delete it from your system. Thank You. >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> --------------------------------------------------------------------- >>>>>>>>>>> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org >>>>>>>>>>> For additional commands, e-mail: user-h...@spark.apache.org >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> -- >>>>>>>>>> Deepak >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> -- >>>>>>>>> Deepak >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> -- >>>>>>>> Deepak >>>>>>>> >>>>>>>> >>>>>> >>>>>> >>>>>> -- >>>>>> Deepak >>>>>> >>>>>> >>>>> >>>> >>>> >>>> -- >>>> Deepak >>>> >>>> >>> >> >> >> -- >> Deepak >> >> >