Re: Error when compiling spark in IDEA and best practice to use IDE?
After sbt/sbt gen-diea, do not import as an SBT project but choose open project and point it to the spark folder. -Xiangrui On Tue, Apr 8, 2014 at 10:45 PM, Sean Owen so...@cloudera.com wrote: I let IntelliJ read the Maven build directly and that works fine. -- Sean Owen | Director, Data Science | London On Wed, Apr 9, 2014 at 6:14 AM, Dong Mo monted...@gmail.com wrote: Dear list, SBT compiles fine, but when I do the following: sbt/sbt gen-idea import project as SBT project to IDEA 13.1 Make Project and these errors show up: Error:(28, 8) object FileContext is not a member of package org.apache.hadoop.fs import org.apache.hadoop.fs.{FileContext, FileStatus, FileSystem, Path, FileUtil} ^ Error:(31, 8) object Master is not a member of package org.apache.hadoop.mapred import org.apache.hadoop.mapred.Master ^ Error:(34, 26) object yarn is not a member of package org.apache.hadoop import org.apache.hadoop.yarn.api._ ^ Error:(35, 26) object yarn is not a member of package org.apache.hadoop import org.apache.hadoop.yarn.api.ApplicationConstants.Environment ^ Error:(36, 26) object yarn is not a member of package org.apache.hadoop import org.apache.hadoop.yarn.api.protocolrecords._ ^ Error:(37, 26) object yarn is not a member of package org.apache.hadoop import org.apache.hadoop.yarn.api.records._ ^ Error:(38, 26) object yarn is not a member of package org.apache.hadoop import org.apache.hadoop.yarn.client.YarnClientImpl ^ Error:(39, 26) object yarn is not a member of package org.apache.hadoop import org.apache.hadoop.yarn.conf.YarnConfiguration ^ Error:(40, 26) object yarn is not a member of package org.apache.hadoop import org.apache.hadoop.yarn.ipc.YarnRPC ^ Error:(41, 26) object yarn is not a member of package org.apache.hadoop import org.apache.hadoop.yarn.util.{Apps, Records} ^ Error:(49, 11) not found: type YarnClientImpl extends YarnClientImpl with Logging { ^ Error:(48, 20) not found: type ClientArguments class Client(args: ClientArguments, conf: Configuration, sparkConf: SparkConf) ^ Error:(51, 18) not found: type ClientArguments def this(args: ClientArguments, sparkConf: SparkConf) = ^ Error:(54, 18) not found: type ClientArguments def this(args: ClientArguments) = this(args, new SparkConf()) ^ Error:(56, 12) not found: type YarnRPC var rpc: YarnRPC = YarnRPC.create(conf) ^ Error:(56, 22) not found: value YarnRPC var rpc: YarnRPC = YarnRPC.create(conf) ^ Error:(57, 17) not found: type YarnConfiguration val yarnConf: YarnConfiguration = new YarnConfiguration(conf) ^ Error:(57, 41) not found: type YarnConfiguration val yarnConf: YarnConfiguration = new YarnConfiguration(conf) ^ Error:(58, 59) value getCredentials is not a member of org.apache.hadoop.security.UserGroupInformation val credentials = UserGroupInformation.getCurrentUser().getCredentials() ^ Error:(60, 34) not found: type ClientDistributedCacheManager private val distCacheMgr = new ClientDistributedCacheManager() ^ Error:(72, 5) not found: value init init(yarnConf) ^ Error:(73, 5) not found: value start start() ^ Error:(76, 24) value getNewApplication is not a member of org.apache.spark.Logging val newApp = super.getNewApplication() ^ Error:(137, 35) not found: type GetNewApplicationResponse def verifyClusterResources(app: GetNewApplicationResponse) = { ^ Error:(156, 65) not found: type ApplicationSubmissionContext def createApplicationSubmissionContext(appId: ApplicationId): ApplicationSubmissionContext = { ^ Error:(156, 49) not found: type ApplicationId def createApplicationSubmissionContext(appId: ApplicationId): ApplicationSubmissionContext = { ^ Error:(118, 31) not found: type ApplicationId def getAppStagingDir(appId: ApplicationId): String = { ^ Error:(224, 69) not found: type LocalResource def prepareLocalResources(appStagingDir: String): HashMap[String, LocalResource] = { ^ Error:(307, 39) not found: type LocalResource localResources: HashMap[String, LocalResource], ^ Error:(343, 38) not found: type ContainerLaunchContext env: HashMap[String, String]): ContainerLaunchContext = { ^
Error when compiling spark in IDEA and best practice to use IDE?
Dear list, SBT compiles fine, but when I do the following: sbt/sbt gen-idea import project as SBT project to IDEA 13.1 Make Project and these errors show up: Error:(28, 8) object FileContext is not a member of package org.apache.hadoop.fs import org.apache.hadoop.fs.{FileContext, FileStatus, FileSystem, Path, FileUtil} ^ Error:(31, 8) object Master is not a member of package org.apache.hadoop.mapred import org.apache.hadoop.mapred.Master ^ Error:(34, 26) object yarn is not a member of package org.apache.hadoop import org.apache.hadoop.yarn.api._ ^ Error:(35, 26) object yarn is not a member of package org.apache.hadoop import org.apache.hadoop.yarn.api.ApplicationConstants.Environment ^ Error:(36, 26) object yarn is not a member of package org.apache.hadoop import org.apache.hadoop.yarn.api.protocolrecords._ ^ Error:(37, 26) object yarn is not a member of package org.apache.hadoop import org.apache.hadoop.yarn.api.records._ ^ Error:(38, 26) object yarn is not a member of package org.apache.hadoop import org.apache.hadoop.yarn.client.YarnClientImpl ^ Error:(39, 26) object yarn is not a member of package org.apache.hadoop import org.apache.hadoop.yarn.conf.YarnConfiguration ^ Error:(40, 26) object yarn is not a member of package org.apache.hadoop import org.apache.hadoop.yarn.ipc.YarnRPC ^ Error:(41, 26) object yarn is not a member of package org.apache.hadoop import org.apache.hadoop.yarn.util.{Apps, Records} ^ Error:(49, 11) not found: type YarnClientImpl extends YarnClientImpl with Logging { ^ Error:(48, 20) not found: type ClientArguments class Client(args: ClientArguments, conf: Configuration, sparkConf: SparkConf) ^ Error:(51, 18) not found: type ClientArguments def this(args: ClientArguments, sparkConf: SparkConf) = ^ Error:(54, 18) not found: type ClientArguments def this(args: ClientArguments) = this(args, new SparkConf()) ^ Error:(56, 12) not found: type YarnRPC var rpc: YarnRPC = YarnRPC.create(conf) ^ Error:(56, 22) not found: value YarnRPC var rpc: YarnRPC = YarnRPC.create(conf) ^ Error:(57, 17) not found: type YarnConfiguration val yarnConf: YarnConfiguration = new YarnConfiguration(conf) ^ Error:(57, 41) not found: type YarnConfiguration val yarnConf: YarnConfiguration = new YarnConfiguration(conf) ^ Error:(58, 59) value getCredentials is not a member of org.apache.hadoop.security.UserGroupInformation val credentials = UserGroupInformation.getCurrentUser().getCredentials() ^ Error:(60, 34) not found: type ClientDistributedCacheManager private val distCacheMgr = new ClientDistributedCacheManager() ^ Error:(72, 5) not found: value init init(yarnConf) ^ Error:(73, 5) not found: value start start() ^ Error:(76, 24) value getNewApplication is not a member of org.apache.spark.Logging val newApp = super.getNewApplication() ^ Error:(137, 35) not found: type GetNewApplicationResponse def verifyClusterResources(app: GetNewApplicationResponse) = { ^ Error:(156, 65) not found: type ApplicationSubmissionContext def createApplicationSubmissionContext(appId: ApplicationId): ApplicationSubmissionContext = { ^ Error:(156, 49) not found: type ApplicationId def createApplicationSubmissionContext(appId: ApplicationId): ApplicationSubmissionContext = { ^ Error:(118, 31) not found: type ApplicationId def getAppStagingDir(appId: ApplicationId): String = { ^ Error:(224, 69) not found: type LocalResource def prepareLocalResources(appStagingDir: String): HashMap[String, LocalResource] = { ^ Error:(307, 39) not found: type LocalResource localResources: HashMap[String, LocalResource], ^ Error:(343, 38) not found: type ContainerLaunchContext env: HashMap[String, String]): ContainerLaunchContext = { ^ Error:(341, 15) not found: type GetNewApplicationResponse newApp: GetNewApplicationResponse, ^ Error:(342, 39) not found: type LocalResource localResources: HashMap[String, LocalResource], ^ Error:(426, 11) value submitApplication is not a member of org.apache.spark.Logging super.submitApplication(appContext) ^ Error:(423, 29) not found: type ApplicationSubmissionContext def submitApp(appContext:
Re: Error when compiling spark in IDEA and best practice to use IDE?
Hi Dong, This is pretty much what I did. I run into the same issue you have. Since I'm not developing yarn related stuff, I just excluded those two yarn related project from intellji, and it works. PS, you may need to exclude java8 project as well now. Sincerely, DB Tsai --- My Blog: https://www.dbtsai.com LinkedIn: https://www.linkedin.com/in/dbtsai On Tue, Apr 8, 2014 at 10:14 PM, Dong Mo monted...@gmail.com wrote: Dear list, SBT compiles fine, but when I do the following: sbt/sbt gen-idea import project as SBT project to IDEA 13.1 Make Project and these errors show up: Error:(28, 8) object FileContext is not a member of package org.apache.hadoop.fs import org.apache.hadoop.fs.{FileContext, FileStatus, FileSystem, Path, FileUtil} ^ Error:(31, 8) object Master is not a member of package org.apache.hadoop.mapred import org.apache.hadoop.mapred.Master ^ Error:(34, 26) object yarn is not a member of package org.apache.hadoop import org.apache.hadoop.yarn.api._ ^ Error:(35, 26) object yarn is not a member of package org.apache.hadoop import org.apache.hadoop.yarn.api.ApplicationConstants.Environment ^ Error:(36, 26) object yarn is not a member of package org.apache.hadoop import org.apache.hadoop.yarn.api.protocolrecords._ ^ Error:(37, 26) object yarn is not a member of package org.apache.hadoop import org.apache.hadoop.yarn.api.records._ ^ Error:(38, 26) object yarn is not a member of package org.apache.hadoop import org.apache.hadoop.yarn.client.YarnClientImpl ^ Error:(39, 26) object yarn is not a member of package org.apache.hadoop import org.apache.hadoop.yarn.conf.YarnConfiguration ^ Error:(40, 26) object yarn is not a member of package org.apache.hadoop import org.apache.hadoop.yarn.ipc.YarnRPC ^ Error:(41, 26) object yarn is not a member of package org.apache.hadoop import org.apache.hadoop.yarn.util.{Apps, Records} ^ Error:(49, 11) not found: type YarnClientImpl extends YarnClientImpl with Logging { ^ Error:(48, 20) not found: type ClientArguments class Client(args: ClientArguments, conf: Configuration, sparkConf: SparkConf) ^ Error:(51, 18) not found: type ClientArguments def this(args: ClientArguments, sparkConf: SparkConf) = ^ Error:(54, 18) not found: type ClientArguments def this(args: ClientArguments) = this(args, new SparkConf()) ^ Error:(56, 12) not found: type YarnRPC var rpc: YarnRPC = YarnRPC.create(conf) ^ Error:(56, 22) not found: value YarnRPC var rpc: YarnRPC = YarnRPC.create(conf) ^ Error:(57, 17) not found: type YarnConfiguration val yarnConf: YarnConfiguration = new YarnConfiguration(conf) ^ Error:(57, 41) not found: type YarnConfiguration val yarnConf: YarnConfiguration = new YarnConfiguration(conf) ^ Error:(58, 59) value getCredentials is not a member of org.apache.hadoop.security.UserGroupInformation val credentials = UserGroupInformation.getCurrentUser().getCredentials() ^ Error:(60, 34) not found: type ClientDistributedCacheManager private val distCacheMgr = new ClientDistributedCacheManager() ^ Error:(72, 5) not found: value init init(yarnConf) ^ Error:(73, 5) not found: value start start() ^ Error:(76, 24) value getNewApplication is not a member of org.apache.spark.Logging val newApp = super.getNewApplication() ^ Error:(137, 35) not found: type GetNewApplicationResponse def verifyClusterResources(app: GetNewApplicationResponse) = { ^ Error:(156, 65) not found: type ApplicationSubmissionContext def createApplicationSubmissionContext(appId: ApplicationId): ApplicationSubmissionContext = { ^ Error:(156, 49) not found: type ApplicationId def createApplicationSubmissionContext(appId: ApplicationId): ApplicationSubmissionContext = { ^ Error:(118, 31) not found: type ApplicationId def getAppStagingDir(appId: ApplicationId): String = { ^ Error:(224, 69) not found: type LocalResource def prepareLocalResources(appStagingDir: String): HashMap[String, LocalResource] = { ^ Error:(307, 39) not found: type LocalResource localResources: HashMap[String, LocalResource], ^ Error:(343, 38) not found: type ContainerLaunchContext
Re: Error when compiling spark in IDEA and best practice to use IDE?
I let IntelliJ read the Maven build directly and that works fine. -- Sean Owen | Director, Data Science | London On Wed, Apr 9, 2014 at 6:14 AM, Dong Mo monted...@gmail.com wrote: Dear list, SBT compiles fine, but when I do the following: sbt/sbt gen-idea import project as SBT project to IDEA 13.1 Make Project and these errors show up: Error:(28, 8) object FileContext is not a member of package org.apache.hadoop.fs import org.apache.hadoop.fs.{FileContext, FileStatus, FileSystem, Path, FileUtil} ^ Error:(31, 8) object Master is not a member of package org.apache.hadoop.mapred import org.apache.hadoop.mapred.Master ^ Error:(34, 26) object yarn is not a member of package org.apache.hadoop import org.apache.hadoop.yarn.api._ ^ Error:(35, 26) object yarn is not a member of package org.apache.hadoop import org.apache.hadoop.yarn.api.ApplicationConstants.Environment ^ Error:(36, 26) object yarn is not a member of package org.apache.hadoop import org.apache.hadoop.yarn.api.protocolrecords._ ^ Error:(37, 26) object yarn is not a member of package org.apache.hadoop import org.apache.hadoop.yarn.api.records._ ^ Error:(38, 26) object yarn is not a member of package org.apache.hadoop import org.apache.hadoop.yarn.client.YarnClientImpl ^ Error:(39, 26) object yarn is not a member of package org.apache.hadoop import org.apache.hadoop.yarn.conf.YarnConfiguration ^ Error:(40, 26) object yarn is not a member of package org.apache.hadoop import org.apache.hadoop.yarn.ipc.YarnRPC ^ Error:(41, 26) object yarn is not a member of package org.apache.hadoop import org.apache.hadoop.yarn.util.{Apps, Records} ^ Error:(49, 11) not found: type YarnClientImpl extends YarnClientImpl with Logging { ^ Error:(48, 20) not found: type ClientArguments class Client(args: ClientArguments, conf: Configuration, sparkConf: SparkConf) ^ Error:(51, 18) not found: type ClientArguments def this(args: ClientArguments, sparkConf: SparkConf) = ^ Error:(54, 18) not found: type ClientArguments def this(args: ClientArguments) = this(args, new SparkConf()) ^ Error:(56, 12) not found: type YarnRPC var rpc: YarnRPC = YarnRPC.create(conf) ^ Error:(56, 22) not found: value YarnRPC var rpc: YarnRPC = YarnRPC.create(conf) ^ Error:(57, 17) not found: type YarnConfiguration val yarnConf: YarnConfiguration = new YarnConfiguration(conf) ^ Error:(57, 41) not found: type YarnConfiguration val yarnConf: YarnConfiguration = new YarnConfiguration(conf) ^ Error:(58, 59) value getCredentials is not a member of org.apache.hadoop.security.UserGroupInformation val credentials = UserGroupInformation.getCurrentUser().getCredentials() ^ Error:(60, 34) not found: type ClientDistributedCacheManager private val distCacheMgr = new ClientDistributedCacheManager() ^ Error:(72, 5) not found: value init init(yarnConf) ^ Error:(73, 5) not found: value start start() ^ Error:(76, 24) value getNewApplication is not a member of org.apache.spark.Logging val newApp = super.getNewApplication() ^ Error:(137, 35) not found: type GetNewApplicationResponse def verifyClusterResources(app: GetNewApplicationResponse) = { ^ Error:(156, 65) not found: type ApplicationSubmissionContext def createApplicationSubmissionContext(appId: ApplicationId): ApplicationSubmissionContext = { ^ Error:(156, 49) not found: type ApplicationId def createApplicationSubmissionContext(appId: ApplicationId): ApplicationSubmissionContext = { ^ Error:(118, 31) not found: type ApplicationId def getAppStagingDir(appId: ApplicationId): String = { ^ Error:(224, 69) not found: type LocalResource def prepareLocalResources(appStagingDir: String): HashMap[String, LocalResource] = { ^ Error:(307, 39) not found: type LocalResource localResources: HashMap[String, LocalResource], ^ Error:(343, 38) not found: type ContainerLaunchContext env: HashMap[String, String]): ContainerLaunchContext = { ^ Error:(341, 15) not found: type GetNewApplicationResponse newApp: GetNewApplicationResponse, ^ Error:(342, 39) not found: type LocalResource localResources: