Does not work 15/03/26 01:07:05 INFO HiveMetaStore.audit: ugi=dvasthimal ip=unknown-ip-addr cmd=get_table : db=default tbl=src_spark 15/03/26 01:07:06 ERROR ql.Driver: FAILED: SemanticException Line 1:23 Invalid path ''/home/dvasthimal/spark1.3/spark-1.3.0-bin-hadoop2.4/examples/src/main/resources/kv1.txt'': No files matching path file:/home/dvasthimal/spark1.3/spark-1.3.0-bin-hadoop2.4/examples/src/main/resources/kv1.txt org.apache.hadoop.hive.ql.parse.SemanticException: Line 1:23 Invalid path ''/home/dvasthimal/spark1.3/spark-1.3.0-bin-hadoop2.4/examples/src/main/resources/kv1.txt'': No files matching path file:/home/dvasthimal/spark1.3/spark-1.3.0-bin-hadoop2.4/examples/src/main/resources/kv1.txt at org.apache.hadoop.hive.ql.parse.LoadSemanticAnalyzer.applyConstraints(LoadSemanticAnalyzer.java:142) at org.apache.hadoop.hive.ql.parse.LoadSemanticAnalyzer.analyzeInternal(LoadSemanticAnalyzer.java:233) at org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:327) at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:422)
Does the input file needs to be passed to executor via -- jars ? On Thu, Mar 26, 2015 at 12:15 PM, Akhil Das <ak...@sigmoidanalytics.com> wrote: > Try to give the complete path to the file kv1.txt. > On 26 Mar 2015 11:48, "ÐΞ€ρ@Ҝ (๏̯͡๏)" <deepuj...@gmail.com> wrote: > >> I am now seeing this error. >> >> >> >> >> >> 15/03/25 19:44:03 ERROR yarn.ApplicationMaster: User class threw >> exception: FAILED: SemanticException Line 1:23 Invalid path >> ''examples/src/main/resources/kv1.txt'': No files matching path >> file:/hadoop/10/scratch/local/usercache/dvasthimal/appcache/application_1426715280024_89893/container_1426715280024_89893_01_000002/examples/src/main/resources/kv1.txt >> >> org.apache.spark.sql.execution.QueryExecutionException: FAILED: >> SemanticException Line 1:23 Invalid path >> ''examples/src/main/resources/kv1.txt'': No files matching path >> file:/hadoop/10/scratch/local/usercache/dvasthimal/appcache/application_1426715280024_89893/container_1426715280024_89893_01_000002/examples/src/main/resources/kv1.txt >> >> at org.apache.spark.sql.hive.HiveContext.runHive(HiveContext.scala:312) >> >> at org.apache.spark.sql.hive.HiveContext.runSqlHive(HiveContext.scala:280) >> >> >> >> >> -sh-4.1$ pwd >> >> /home/dvasthimal/spark1.3/spark-1.3.0-bin-hadoop2.4 >> >> -sh-4.1$ ls examples/src/main/resources/kv1.txt >> >> examples/src/main/resources/kv1.txt >> >> -sh-4.1$ >> >> >> >> On Thu, Mar 26, 2015 at 8:08 AM, Zhan Zhang <zzh...@hortonworks.com> >> wrote: >> >>> You can do it in $SPARK_HOME/conf/spark-defaults.con >>> >>> spark.driver.extraJavaOptions -XX:MaxPermSize=512m >>> >>> Thanks. >>> >>> Zhan Zhang >>> >>> >>> On Mar 25, 2015, at 7:25 PM, ÐΞ€ρ@Ҝ (๏̯͡๏) <deepuj...@gmail.com> wrote: >>> >>> Where and how do i pass this or other JVM argument ? >>> -XX:MaxPermSize=512m >>> >>> On Wed, Mar 25, 2015 at 11:36 PM, Zhan Zhang <zzh...@hortonworks.com> >>> wrote: >>> >>>> I solve this by increase the PermGen memory size in driver. >>>> >>>> -XX:MaxPermSize=512m >>>> >>>> Thanks. >>>> >>>> Zhan Zhang >>>> >>>> On Mar 25, 2015, at 10:54 AM, ÐΞ€ρ@Ҝ (๏̯͡๏) <deepuj...@gmail.com> >>>> wrote: >>>> >>>> I am facing same issue, posted a new thread. Please respond. >>>> >>>> On Wed, Jan 14, 2015 at 4:38 AM, Zhan Zhang <zzh...@hortonworks.com> >>>> wrote: >>>> >>>>> Hi Folks, >>>>> >>>>> I am trying to run hive context in yarn-cluster mode, but met some >>>>> error. Does anybody know what cause the issue. >>>>> >>>>> I use following cmd to build the distribution: >>>>> >>>>> ./make-distribution.sh -Phive -Phive-thriftserver -Pyarn >>>>> -Phadoop-2.4 >>>>> >>>>> 15/01/13 17:59:42 INFO cluster.YarnClusterScheduler: >>>>> YarnClusterScheduler.postStartHook done >>>>> 15/01/13 17:59:42 INFO storage.BlockManagerMasterActor: Registering >>>>> block manager cn122-10.l42scl.hortonworks.com:56157 with 1589.8 MB >>>>> RAM, BlockManagerId(2, cn122-10.l42scl.hortonworks.com, 56157) >>>>> 15/01/13 17:59:43 INFO parse.ParseDriver: Parsing command: CREATE >>>>> TABLE IF NOT EXISTS src (key INT, value STRING) >>>>> 15/01/13 17:59:43 INFO parse.ParseDriver: Parse Completed >>>>> 15/01/13 17:59:44 INFO metastore.HiveMetaStore: 0: Opening raw store >>>>> with implemenation class:org.apache.hadoop.hive.metastore.ObjectStore >>>>> 15/01/13 17:59:44 INFO metastore.ObjectStore: ObjectStore, initialize >>>>> called >>>>> 15/01/13 17:59:44 INFO DataNucleus.Persistence: Property >>>>> datanucleus.cache.level2 unknown - will be ignored >>>>> 15/01/13 17:59:44 INFO DataNucleus.Persistence: Property >>>>> hive.metastore.integral.jdo.pushdown unknown - will be ignored >>>>> 15/01/13 17:59:44 WARN DataNucleus.Connection: BoneCP specified but >>>>> not present in CLASSPATH (or one of dependencies) >>>>> 15/01/13 17:59:44 WARN DataNucleus.Connection: BoneCP specified but >>>>> not present in CLASSPATH (or one of dependencies) >>>>> 15/01/13 17:59:52 INFO metastore.ObjectStore: Setting MetaStore object >>>>> pin classes with >>>>> hive.metastore.cache.pinobjtypes="Table,StorageDescriptor,SerDeInfo,Partition,Database,Type,FieldSchema,Order" >>>>> 15/01/13 17:59:52 INFO metastore.MetaStoreDirectSql: MySQL check >>>>> failed, assuming we are not on mysql: Lexical error at line 1, column 5. >>>>> Encountered: "@" (64), after : "". >>>>> 15/01/13 17:59:53 INFO DataNucleus.Datastore: The class >>>>> "org.apache.hadoop.hive.metastore.model.MFieldSchema" is tagged as >>>>> "embedded-only" so does not have its own datastore table. >>>>> 15/01/13 17:59:53 INFO DataNucleus.Datastore: The class >>>>> "org.apache.hadoop.hive.metastore.model.MOrder" is tagged as >>>>> "embedded-only" so does not have its own datastore table. >>>>> 15/01/13 17:59:59 INFO DataNucleus.Datastore: The class >>>>> "org.apache.hadoop.hive.metastore.model.MFieldSchema" is tagged as >>>>> "embedded-only" so does not have its own datastore table. >>>>> 15/01/13 17:59:59 INFO DataNucleus.Datastore: The class >>>>> "org.apache.hadoop.hive.metastore.model.MOrder" is tagged as >>>>> "embedded-only" so does not have its own datastore table. >>>>> 15/01/13 18:00:00 INFO metastore.ObjectStore: Initialized ObjectStore >>>>> 15/01/13 18:00:00 WARN metastore.ObjectStore: Version information not >>>>> found in metastore. hive.metastore.schema.verification is not enabled so >>>>> recording the schema version 0.13.1aa >>>>> 15/01/13 18:00:01 INFO metastore.HiveMetaStore: Added admin role in >>>>> metastore >>>>> 15/01/13 18:00:01 INFO metastore.HiveMetaStore: Added public role in >>>>> metastore >>>>> 15/01/13 18:00:01 INFO metastore.HiveMetaStore: No user is added in >>>>> admin role, since config is empty >>>>> 15/01/13 18:00:01 INFO session.SessionState: No Tez session required >>>>> at this point. hive.execution.engine=mr. >>>>> 15/01/13 18:00:02 INFO log.PerfLogger: <PERFLOG method=Driver.run >>>>> from=org.apache.hadoop.hive.ql.Driver> >>>>> 15/01/13 18:00:02 INFO log.PerfLogger: <PERFLOG method=TimeToSubmit >>>>> from=org.apache.hadoop.hive.ql.Driver> >>>>> 15/01/13 18:00:02 INFO ql.Driver: Concurrency mode is disabled, not >>>>> creating a lock manager >>>>> 15/01/13 18:00:02 INFO log.PerfLogger: <PERFLOG method=compile >>>>> from=org.apache.hadoop.hive.ql.Driver> >>>>> 15/01/13 18:00:03 INFO log.PerfLogger: <PERFLOG method=parse >>>>> from=org.apache.hadoop.hive.ql.Driver> >>>>> 15/01/13 18:00:03 INFO parse.ParseDriver: Parsing command: CREATE >>>>> TABLE IF NOT EXISTS src (key INT, value STRING) >>>>> 15/01/13 18:00:03 INFO parse.ParseDriver: Parse Completed >>>>> 15/01/13 18:00:03 INFO log.PerfLogger: </PERFLOG method=parse >>>>> start=1421190003030 end=1421190003031 duration=1 >>>>> from=org.apache.hadoop.hive.ql.Driver> >>>>> 15/01/13 18:00:03 INFO log.PerfLogger: <PERFLOG method=semanticAnalyze >>>>> from=org.apache.hadoop.hive.ql.Driver> >>>>> 15/01/13 18:00:03 INFO parse.SemanticAnalyzer: Starting Semantic >>>>> Analysis >>>>> 15/01/13 18:00:03 INFO parse.SemanticAnalyzer: Creating table src >>>>> position=27 >>>>> 15/01/13 18:00:03 INFO metastore.HiveMetaStore: 0: get_table : >>>>> db=default tbl=src >>>>> 15/01/13 18:00:03 INFO HiveMetaStore.audit: ugi=zzhang >>>>> ip=unknown-ip-addr cmd=get_table : db=default tbl=src >>>>> 15/01/13 18:00:03 INFO metastore.HiveMetaStore: 0: get_database: >>>>> default >>>>> 15/01/13 18:00:03 INFO HiveMetaStore.audit: ugi=zzhang >>>>> ip=unknown-ip-addr cmd=get_database: default >>>>> 15/01/13 18:00:03 INFO ql.Driver: Semantic Analysis Completed >>>>> 15/01/13 18:00:03 INFO log.PerfLogger: </PERFLOG >>>>> method=semanticAnalyze start=1421190003031 end=1421190003406 duration=375 >>>>> from=org.apache.hadoop.hive.ql.Driver> >>>>> 15/01/13 18:00:03 INFO ql.Driver: Returning Hive schema: >>>>> Schema(fieldSchemas:null, properties:null) >>>>> 15/01/13 18:00:03 INFO log.PerfLogger: </PERFLOG method=compile >>>>> start=1421190002998 end=1421190003416 duration=418 >>>>> from=org.apache.hadoop.hive.ql.Driver> >>>>> 15/01/13 18:00:03 INFO log.PerfLogger: <PERFLOG method=Driver.execute >>>>> from=org.apache.hadoop.hive.ql.Driver> >>>>> 15/01/13 18:00:03 INFO ql.Driver: Starting command: CREATE TABLE IF >>>>> NOT EXISTS src (key INT, value STRING) >>>>> 15/01/13 18:00:03 INFO log.PerfLogger: </PERFLOG method=TimeToSubmit >>>>> start=1421190002995 end=1421190003421 duration=426 >>>>> from=org.apache.hadoop.hive.ql.Driver> >>>>> 15/01/13 18:00:03 INFO log.PerfLogger: <PERFLOG method=runTasks >>>>> from=org.apache.hadoop.hive.ql.Driver> >>>>> 15/01/13 18:00:03 INFO log.PerfLogger: <PERFLOG >>>>> method=task.DDL.Stage-0 from=org.apache.hadoop.hive.ql.Driver> >>>>> 15/01/13 18:00:03 INFO exec.DDLTask: Default to LazySimpleSerDe for >>>>> table src >>>>> 15/01/13 18:00:05 INFO log.PerfLogger: </PERFLOG method=Driver.execute >>>>> start=1421190003416 end=1421190005498 duration=2082 >>>>> from=org.apache.hadoop.hive.ql.Driver> >>>>> Exception in thread "Driver" >>>>> Exception: java.lang.OutOfMemoryError thrown from the >>>>> UncaughtExceptionHandler in thread "Driver" >>>>> -- >>>>> CONFIDENTIALITY NOTICE >>>>> NOTICE: This message is intended for the use of the individual or >>>>> entity to >>>>> which it is addressed and may contain information that is confidential, >>>>> privileged and exempt from disclosure under applicable law. If the >>>>> reader >>>>> of this message is not the intended recipient, you are hereby notified >>>>> that >>>>> any printing, copying, dissemination, distribution, disclosure or >>>>> forwarding of this communication is strictly prohibited. If you have >>>>> received this communication in error, please contact the sender >>>>> immediately >>>>> and delete it from your system. Thank You. >>>>> >>>>> --------------------------------------------------------------------- >>>>> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org >>>>> For additional commands, e-mail: user-h...@spark.apache.org >>>>> >>>>> >>>> >>>> >>>> -- >>>> Deepak >>>> >>>> >>>> >>> >>> >>> -- >>> Deepak >>> >>> >>> >> >> >> -- >> Deepak >> >> -- Deepak