After sbt/sbt gen-diea, do not import as an SBT project but choose "open project" and point it to the spark folder. -Xiangrui
On Tue, Apr 8, 2014 at 10:45 PM, Sean Owen <so...@cloudera.com> wrote: > I let IntelliJ read the Maven build directly and that works fine. > -- > Sean Owen | Director, Data Science | London > > > On Wed, Apr 9, 2014 at 6:14 AM, Dong Mo <monted...@gmail.com> wrote: >> Dear list, >> >> SBT compiles fine, but when I do the following: >> sbt/sbt gen-idea >> import project as SBT project to IDEA 13.1 >> Make Project >> and these errors show up: >> >> Error:(28, 8) object FileContext is not a member of package >> org.apache.hadoop.fs >> import org.apache.hadoop.fs.{FileContext, FileStatus, FileSystem, Path, >> FileUtil} >> ^ >> Error:(31, 8) object Master is not a member of package >> org.apache.hadoop.mapred >> import org.apache.hadoop.mapred.Master >> ^ >> Error:(34, 26) object yarn is not a member of package org.apache.hadoop >> import org.apache.hadoop.yarn.api._ >> ^ >> Error:(35, 26) object yarn is not a member of package org.apache.hadoop >> import org.apache.hadoop.yarn.api.ApplicationConstants.Environment >> ^ >> Error:(36, 26) object yarn is not a member of package org.apache.hadoop >> import org.apache.hadoop.yarn.api.protocolrecords._ >> ^ >> Error:(37, 26) object yarn is not a member of package org.apache.hadoop >> import org.apache.hadoop.yarn.api.records._ >> ^ >> Error:(38, 26) object yarn is not a member of package org.apache.hadoop >> import org.apache.hadoop.yarn.client.YarnClientImpl >> ^ >> Error:(39, 26) object yarn is not a member of package org.apache.hadoop >> import org.apache.hadoop.yarn.conf.YarnConfiguration >> ^ >> Error:(40, 26) object yarn is not a member of package org.apache.hadoop >> import org.apache.hadoop.yarn.ipc.YarnRPC >> ^ >> Error:(41, 26) object yarn is not a member of package org.apache.hadoop >> import org.apache.hadoop.yarn.util.{Apps, Records} >> ^ >> Error:(49, 11) not found: type YarnClientImpl >> extends YarnClientImpl with Logging { >> ^ >> Error:(48, 20) not found: type ClientArguments >> class Client(args: ClientArguments, conf: Configuration, sparkConf: >> SparkConf) >> ^ >> Error:(51, 18) not found: type ClientArguments >> def this(args: ClientArguments, sparkConf: SparkConf) = >> ^ >> Error:(54, 18) not found: type ClientArguments >> def this(args: ClientArguments) = this(args, new SparkConf()) >> ^ >> Error:(56, 12) not found: type YarnRPC >> var rpc: YarnRPC = YarnRPC.create(conf) >> ^ >> Error:(56, 22) not found: value YarnRPC >> var rpc: YarnRPC = YarnRPC.create(conf) >> ^ >> Error:(57, 17) not found: type YarnConfiguration >> val yarnConf: YarnConfiguration = new YarnConfiguration(conf) >> ^ >> Error:(57, 41) not found: type YarnConfiguration >> val yarnConf: YarnConfiguration = new YarnConfiguration(conf) >> ^ >> Error:(58, 59) value getCredentials is not a member of >> org.apache.hadoop.security.UserGroupInformation >> val credentials = UserGroupInformation.getCurrentUser().getCredentials() >> ^ >> Error:(60, 34) not found: type ClientDistributedCacheManager >> private val distCacheMgr = new ClientDistributedCacheManager() >> ^ >> Error:(72, 5) not found: value init >> init(yarnConf) >> ^ >> Error:(73, 5) not found: value start >> start() >> ^ >> Error:(76, 24) value getNewApplication is not a member of >> org.apache.spark.Logging >> val newApp = super.getNewApplication() >> ^ >> Error:(137, 35) not found: type GetNewApplicationResponse >> def verifyClusterResources(app: GetNewApplicationResponse) = { >> ^ >> Error:(156, 65) not found: type ApplicationSubmissionContext >> def createApplicationSubmissionContext(appId: ApplicationId): >> ApplicationSubmissionContext = { >> ^ >> Error:(156, 49) not found: type ApplicationId >> def createApplicationSubmissionContext(appId: ApplicationId): >> ApplicationSubmissionContext = { >> ^ >> Error:(118, 31) not found: type ApplicationId >> def getAppStagingDir(appId: ApplicationId): String = { >> ^ >> Error:(224, 69) not found: type LocalResource >> def prepareLocalResources(appStagingDir: String): HashMap[String, >> LocalResource] = { >> ^ >> Error:(307, 39) not found: type LocalResource >> localResources: HashMap[String, LocalResource], >> ^ >> Error:(343, 38) not found: type ContainerLaunchContext >> env: HashMap[String, String]): ContainerLaunchContext = { >> ^ >> Error:(341, 15) not found: type GetNewApplicationResponse >> newApp: GetNewApplicationResponse, >> ^ >> Error:(342, 39) not found: type LocalResource >> localResources: HashMap[String, LocalResource], >> ^ >> Error:(426, 11) value submitApplication is not a member of >> org.apache.spark.Logging >> super.submitApplication(appContext) >> ^ >> Error:(423, 29) not found: type ApplicationSubmissionContext >> def submitApp(appContext: ApplicationSubmissionContext) = { >> ^ >> Error:(429, 33) not found: type ApplicationId >> def monitorApplication(appId: ApplicationId): Boolean = { >> ^ >> Error:(123, 25) not found: type YarnClusterMetrics >> val clusterMetrics: YarnClusterMetrics = super.getYarnClusterMetrics >> ^ >> Error:(123, 52) value getYarnClusterMetrics is not a member of >> org.apache.spark.Logging >> val clusterMetrics: YarnClusterMetrics = super.getYarnClusterMetrics >> ^ >> Error:(127, 20) not found: type QueueInfo >> val queueInfo: QueueInfo = super.getQueueInfo(args.amQueue) >> ^ >> Error:(127, 38) value getQueueInfo is not a member of >> org.apache.spark.Logging >> val queueInfo: QueueInfo = super.getQueueInfo(args.amQueue) >> ^ >> Error:(158, 22) not found: value Records >> val appContext = >> Records.newRecord(classOf[ApplicationSubmissionContext]) >> ^ >> Error:(219, 14) not found: value FileContext >> val fc = FileContext.getFileContext(qualPath.toUri(), conf) >> ^ >> Error:(230, 29) not found: value Master >> val delegTokenRenewer = Master.getMasterPrincipal(conf) >> ^ >> Error:(242, 13) value addDelegationTokens is not a member of >> org.apache.hadoop.fs.FileSystem >> dstFs.addDelegationTokens(delegTokenRenewer, credentials) >> ^ >> Error:(244, 42) not found: type LocalResource >> val localResources = HashMap[String, LocalResource]() >> ^ >> Error:(302, 43) value addCredentials is not a member of >> org.apache.hadoop.security.UserGroupInformation >> UserGroupInformation.getCurrentUser().addCredentials(credentials) >> ^ >> Error:(323, 5) not found: value Apps >> Apps.setEnvFromInputString(env, System.getenv("SPARK_YARN_USER_ENV")) >> ^ >> Error:(330, 36) not found: type ClientArguments >> def userArgsToString(clientArgs: ClientArguments): String = { >> ^ >> Error:(345, 23) not found: value Records >> val amContainer = Records.newRecord(classOf[ContainerLaunchContext]) >> ^ >> Error:(363, 16) not found: value Environment >> new Path(Environment.PWD.$(), >> YarnConfiguration.DEFAULT_CONTAINER_TEMP_DIR) + " " >> ^ >> Error:(392, 21) not found: value Environment >> javaCommand = Environment.JAVA_HOME.$() + "/bin/java" >> ^ >> Error:(405, 16) not found: value ApplicationConstants >> " 1> " + ApplicationConstants.LOG_DIR_EXPANSION_VAR + "/stdout" + >> ^ >> Error:(410, 22) not found: value Records >> val capability = >> Records.newRecord(classOf[Resource]).asInstanceOf[Resource] >> ^ >> Error:(410, 72) not found: type Resource >> val capability = >> Records.newRecord(classOf[Resource]).asInstanceOf[Resource] >> ^ >> Error:(434, 26) value getApplicationReport is not a member of >> org.apache.spark.Logging >> val report = super.getApplicationReport(appId) >> ^ >> Error:(474, 20) not found: type ClientArguments >> val args = new ClientArguments(argStrings, sparkConf) >> ^ >> Error:(481, 31) not found: value YarnConfiguration >> for (c <- conf.getStrings(YarnConfiguration.YARN_APPLICATION_CLASSPATH)) >> { >> ^ >> Error:(487, 5) not found: value Apps >> Apps.addToEnvironment(env, Environment.CLASSPATH.name, >> Environment.PWD.$()) >> ^ >> Error:(490, 7) not found: value Apps >> Apps.addToEnvironment(env, Environment.CLASSPATH.name, >> Environment.PWD.$() + >> ^ >> Error:(496, 7) not found: value Apps >> Apps.addToEnvironment(env, Environment.CLASSPATH.name, >> Environment.PWD.$() + >> ^ >> Error:(499, 5) not found: value Apps >> Apps.addToEnvironment(env, Environment.CLASSPATH.name, >> Environment.PWD.$() + >> ^ >> Error:(504, 7) not found: value Apps >> Apps.addToEnvironment(env, Environment.CLASSPATH.name, >> Environment.PWD.$() + >> ^ >> Error:(507, 5) not found: value Apps >> Apps.addToEnvironment(env, Environment.CLASSPATH.name, >> Environment.PWD.$() + >> ^ >> >> Any idea what's causing them, and maybe I am not using the best practice to >> import Spark to IDE? >> >> I would appreciate any suggestion on the best practice to import Spark to >> any IDE. >> >> Thank you >> >> -Mo