Try to give the complete path to the file kv1.txt. On 26 Mar 2015 11:48, "ÐΞ€ρ@Ҝ (๏̯͡๏)" <deepuj...@gmail.com> wrote:
> I am now seeing this error. > > > > > > 15/03/25 19:44:03 ERROR yarn.ApplicationMaster: User class threw > exception: FAILED: SemanticException Line 1:23 Invalid path > ''examples/src/main/resources/kv1.txt'': No files matching path > file:/hadoop/10/scratch/local/usercache/dvasthimal/appcache/application_1426715280024_89893/container_1426715280024_89893_01_000002/examples/src/main/resources/kv1.txt > > org.apache.spark.sql.execution.QueryExecutionException: FAILED: > SemanticException Line 1:23 Invalid path > ''examples/src/main/resources/kv1.txt'': No files matching path > file:/hadoop/10/scratch/local/usercache/dvasthimal/appcache/application_1426715280024_89893/container_1426715280024_89893_01_000002/examples/src/main/resources/kv1.txt > > at org.apache.spark.sql.hive.HiveContext.runHive(HiveContext.scala:312) > > at org.apache.spark.sql.hive.HiveContext.runSqlHive(HiveContext.scala:280) > > > > > -sh-4.1$ pwd > > /home/dvasthimal/spark1.3/spark-1.3.0-bin-hadoop2.4 > > -sh-4.1$ ls examples/src/main/resources/kv1.txt > > examples/src/main/resources/kv1.txt > > -sh-4.1$ > > > > On Thu, Mar 26, 2015 at 8:08 AM, Zhan Zhang <zzh...@hortonworks.com> > wrote: > >> You can do it in $SPARK_HOME/conf/spark-defaults.con >> >> spark.driver.extraJavaOptions -XX:MaxPermSize=512m >> >> Thanks. >> >> Zhan Zhang >> >> >> On Mar 25, 2015, at 7:25 PM, ÐΞ€ρ@Ҝ (๏̯͡๏) <deepuj...@gmail.com> wrote: >> >> Where and how do i pass this or other JVM argument ? >> -XX:MaxPermSize=512m >> >> On Wed, Mar 25, 2015 at 11:36 PM, Zhan Zhang <zzh...@hortonworks.com> >> wrote: >> >>> I solve this by increase the PermGen memory size in driver. >>> >>> -XX:MaxPermSize=512m >>> >>> Thanks. >>> >>> Zhan Zhang >>> >>> On Mar 25, 2015, at 10:54 AM, ÐΞ€ρ@Ҝ (๏̯͡๏) <deepuj...@gmail.com> >>> wrote: >>> >>> I am facing same issue, posted a new thread. Please respond. >>> >>> On Wed, Jan 14, 2015 at 4:38 AM, Zhan Zhang <zzh...@hortonworks.com> >>> wrote: >>> >>>> Hi Folks, >>>> >>>> I am trying to run hive context in yarn-cluster mode, but met some >>>> error. Does anybody know what cause the issue. >>>> >>>> I use following cmd to build the distribution: >>>> >>>> ./make-distribution.sh -Phive -Phive-thriftserver -Pyarn -Phadoop-2.4 >>>> >>>> 15/01/13 17:59:42 INFO cluster.YarnClusterScheduler: >>>> YarnClusterScheduler.postStartHook done >>>> 15/01/13 17:59:42 INFO storage.BlockManagerMasterActor: Registering >>>> block manager cn122-10.l42scl.hortonworks.com:56157 with 1589.8 MB >>>> RAM, BlockManagerId(2, cn122-10.l42scl.hortonworks.com, 56157) >>>> 15/01/13 17:59:43 INFO parse.ParseDriver: Parsing command: CREATE TABLE >>>> IF NOT EXISTS src (key INT, value STRING) >>>> 15/01/13 17:59:43 INFO parse.ParseDriver: Parse Completed >>>> 15/01/13 17:59:44 INFO metastore.HiveMetaStore: 0: Opening raw store >>>> with implemenation class:org.apache.hadoop.hive.metastore.ObjectStore >>>> 15/01/13 17:59:44 INFO metastore.ObjectStore: ObjectStore, initialize >>>> called >>>> 15/01/13 17:59:44 INFO DataNucleus.Persistence: Property >>>> datanucleus.cache.level2 unknown - will be ignored >>>> 15/01/13 17:59:44 INFO DataNucleus.Persistence: Property >>>> hive.metastore.integral.jdo.pushdown unknown - will be ignored >>>> 15/01/13 17:59:44 WARN DataNucleus.Connection: BoneCP specified but not >>>> present in CLASSPATH (or one of dependencies) >>>> 15/01/13 17:59:44 WARN DataNucleus.Connection: BoneCP specified but not >>>> present in CLASSPATH (or one of dependencies) >>>> 15/01/13 17:59:52 INFO metastore.ObjectStore: Setting MetaStore object >>>> pin classes with >>>> hive.metastore.cache.pinobjtypes="Table,StorageDescriptor,SerDeInfo,Partition,Database,Type,FieldSchema,Order" >>>> 15/01/13 17:59:52 INFO metastore.MetaStoreDirectSql: MySQL check >>>> failed, assuming we are not on mysql: Lexical error at line 1, column 5. >>>> Encountered: "@" (64), after : "". >>>> 15/01/13 17:59:53 INFO DataNucleus.Datastore: The class >>>> "org.apache.hadoop.hive.metastore.model.MFieldSchema" is tagged as >>>> "embedded-only" so does not have its own datastore table. >>>> 15/01/13 17:59:53 INFO DataNucleus.Datastore: The class >>>> "org.apache.hadoop.hive.metastore.model.MOrder" is tagged as >>>> "embedded-only" so does not have its own datastore table. >>>> 15/01/13 17:59:59 INFO DataNucleus.Datastore: The class >>>> "org.apache.hadoop.hive.metastore.model.MFieldSchema" is tagged as >>>> "embedded-only" so does not have its own datastore table. >>>> 15/01/13 17:59:59 INFO DataNucleus.Datastore: The class >>>> "org.apache.hadoop.hive.metastore.model.MOrder" is tagged as >>>> "embedded-only" so does not have its own datastore table. >>>> 15/01/13 18:00:00 INFO metastore.ObjectStore: Initialized ObjectStore >>>> 15/01/13 18:00:00 WARN metastore.ObjectStore: Version information not >>>> found in metastore. hive.metastore.schema.verification is not enabled so >>>> recording the schema version 0.13.1aa >>>> 15/01/13 18:00:01 INFO metastore.HiveMetaStore: Added admin role in >>>> metastore >>>> 15/01/13 18:00:01 INFO metastore.HiveMetaStore: Added public role in >>>> metastore >>>> 15/01/13 18:00:01 INFO metastore.HiveMetaStore: No user is added in >>>> admin role, since config is empty >>>> 15/01/13 18:00:01 INFO session.SessionState: No Tez session required at >>>> this point. hive.execution.engine=mr. >>>> 15/01/13 18:00:02 INFO log.PerfLogger: <PERFLOG method=Driver.run >>>> from=org.apache.hadoop.hive.ql.Driver> >>>> 15/01/13 18:00:02 INFO log.PerfLogger: <PERFLOG method=TimeToSubmit >>>> from=org.apache.hadoop.hive.ql.Driver> >>>> 15/01/13 18:00:02 INFO ql.Driver: Concurrency mode is disabled, not >>>> creating a lock manager >>>> 15/01/13 18:00:02 INFO log.PerfLogger: <PERFLOG method=compile >>>> from=org.apache.hadoop.hive.ql.Driver> >>>> 15/01/13 18:00:03 INFO log.PerfLogger: <PERFLOG method=parse >>>> from=org.apache.hadoop.hive.ql.Driver> >>>> 15/01/13 18:00:03 INFO parse.ParseDriver: Parsing command: CREATE TABLE >>>> IF NOT EXISTS src (key INT, value STRING) >>>> 15/01/13 18:00:03 INFO parse.ParseDriver: Parse Completed >>>> 15/01/13 18:00:03 INFO log.PerfLogger: </PERFLOG method=parse >>>> start=1421190003030 end=1421190003031 duration=1 >>>> from=org.apache.hadoop.hive.ql.Driver> >>>> 15/01/13 18:00:03 INFO log.PerfLogger: <PERFLOG method=semanticAnalyze >>>> from=org.apache.hadoop.hive.ql.Driver> >>>> 15/01/13 18:00:03 INFO parse.SemanticAnalyzer: Starting Semantic >>>> Analysis >>>> 15/01/13 18:00:03 INFO parse.SemanticAnalyzer: Creating table src >>>> position=27 >>>> 15/01/13 18:00:03 INFO metastore.HiveMetaStore: 0: get_table : >>>> db=default tbl=src >>>> 15/01/13 18:00:03 INFO HiveMetaStore.audit: ugi=zzhang >>>> ip=unknown-ip-addr cmd=get_table : db=default tbl=src >>>> 15/01/13 18:00:03 INFO metastore.HiveMetaStore: 0: get_database: default >>>> 15/01/13 18:00:03 INFO HiveMetaStore.audit: ugi=zzhang >>>> ip=unknown-ip-addr cmd=get_database: default >>>> 15/01/13 18:00:03 INFO ql.Driver: Semantic Analysis Completed >>>> 15/01/13 18:00:03 INFO log.PerfLogger: </PERFLOG method=semanticAnalyze >>>> start=1421190003031 end=1421190003406 duration=375 >>>> from=org.apache.hadoop.hive.ql.Driver> >>>> 15/01/13 18:00:03 INFO ql.Driver: Returning Hive schema: >>>> Schema(fieldSchemas:null, properties:null) >>>> 15/01/13 18:00:03 INFO log.PerfLogger: </PERFLOG method=compile >>>> start=1421190002998 end=1421190003416 duration=418 >>>> from=org.apache.hadoop.hive.ql.Driver> >>>> 15/01/13 18:00:03 INFO log.PerfLogger: <PERFLOG method=Driver.execute >>>> from=org.apache.hadoop.hive.ql.Driver> >>>> 15/01/13 18:00:03 INFO ql.Driver: Starting command: CREATE TABLE IF NOT >>>> EXISTS src (key INT, value STRING) >>>> 15/01/13 18:00:03 INFO log.PerfLogger: </PERFLOG method=TimeToSubmit >>>> start=1421190002995 end=1421190003421 duration=426 >>>> from=org.apache.hadoop.hive.ql.Driver> >>>> 15/01/13 18:00:03 INFO log.PerfLogger: <PERFLOG method=runTasks >>>> from=org.apache.hadoop.hive.ql.Driver> >>>> 15/01/13 18:00:03 INFO log.PerfLogger: <PERFLOG method=task.DDL.Stage-0 >>>> from=org.apache.hadoop.hive.ql.Driver> >>>> 15/01/13 18:00:03 INFO exec.DDLTask: Default to LazySimpleSerDe for >>>> table src >>>> 15/01/13 18:00:05 INFO log.PerfLogger: </PERFLOG method=Driver.execute >>>> start=1421190003416 end=1421190005498 duration=2082 >>>> from=org.apache.hadoop.hive.ql.Driver> >>>> Exception in thread "Driver" >>>> Exception: java.lang.OutOfMemoryError thrown from the >>>> UncaughtExceptionHandler in thread "Driver" >>>> -- >>>> CONFIDENTIALITY NOTICE >>>> NOTICE: This message is intended for the use of the individual or >>>> entity to >>>> which it is addressed and may contain information that is confidential, >>>> privileged and exempt from disclosure under applicable law. If the >>>> reader >>>> of this message is not the intended recipient, you are hereby notified >>>> that >>>> any printing, copying, dissemination, distribution, disclosure or >>>> forwarding of this communication is strictly prohibited. If you have >>>> received this communication in error, please contact the sender >>>> immediately >>>> and delete it from your system. Thank You. >>>> >>>> --------------------------------------------------------------------- >>>> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org >>>> For additional commands, e-mail: user-h...@spark.apache.org >>>> >>>> >>> >>> >>> -- >>> Deepak >>> >>> >>> >> >> >> -- >> Deepak >> >> >> > > > -- > Deepak > >