I let IntelliJ read the Maven build directly and that works fine. -- Sean Owen | Director, Data Science | London
On Wed, Apr 9, 2014 at 6:14 AM, Dong Mo <monted...@gmail.com> wrote: > Dear list, > > SBT compiles fine, but when I do the following: > sbt/sbt gen-idea > import project as SBT project to IDEA 13.1 > Make Project > and these errors show up: > > Error:(28, 8) object FileContext is not a member of package > org.apache.hadoop.fs > import org.apache.hadoop.fs.{FileContext, FileStatus, FileSystem, Path, > FileUtil} > ^ > Error:(31, 8) object Master is not a member of package > org.apache.hadoop.mapred > import org.apache.hadoop.mapred.Master > ^ > Error:(34, 26) object yarn is not a member of package org.apache.hadoop > import org.apache.hadoop.yarn.api._ > ^ > Error:(35, 26) object yarn is not a member of package org.apache.hadoop > import org.apache.hadoop.yarn.api.ApplicationConstants.Environment > ^ > Error:(36, 26) object yarn is not a member of package org.apache.hadoop > import org.apache.hadoop.yarn.api.protocolrecords._ > ^ > Error:(37, 26) object yarn is not a member of package org.apache.hadoop > import org.apache.hadoop.yarn.api.records._ > ^ > Error:(38, 26) object yarn is not a member of package org.apache.hadoop > import org.apache.hadoop.yarn.client.YarnClientImpl > ^ > Error:(39, 26) object yarn is not a member of package org.apache.hadoop > import org.apache.hadoop.yarn.conf.YarnConfiguration > ^ > Error:(40, 26) object yarn is not a member of package org.apache.hadoop > import org.apache.hadoop.yarn.ipc.YarnRPC > ^ > Error:(41, 26) object yarn is not a member of package org.apache.hadoop > import org.apache.hadoop.yarn.util.{Apps, Records} > ^ > Error:(49, 11) not found: type YarnClientImpl > extends YarnClientImpl with Logging { > ^ > Error:(48, 20) not found: type ClientArguments > class Client(args: ClientArguments, conf: Configuration, sparkConf: > SparkConf) > ^ > Error:(51, 18) not found: type ClientArguments > def this(args: ClientArguments, sparkConf: SparkConf) = > ^ > Error:(54, 18) not found: type ClientArguments > def this(args: ClientArguments) = this(args, new SparkConf()) > ^ > Error:(56, 12) not found: type YarnRPC > var rpc: YarnRPC = YarnRPC.create(conf) > ^ > Error:(56, 22) not found: value YarnRPC > var rpc: YarnRPC = YarnRPC.create(conf) > ^ > Error:(57, 17) not found: type YarnConfiguration > val yarnConf: YarnConfiguration = new YarnConfiguration(conf) > ^ > Error:(57, 41) not found: type YarnConfiguration > val yarnConf: YarnConfiguration = new YarnConfiguration(conf) > ^ > Error:(58, 59) value getCredentials is not a member of > org.apache.hadoop.security.UserGroupInformation > val credentials = UserGroupInformation.getCurrentUser().getCredentials() > ^ > Error:(60, 34) not found: type ClientDistributedCacheManager > private val distCacheMgr = new ClientDistributedCacheManager() > ^ > Error:(72, 5) not found: value init > init(yarnConf) > ^ > Error:(73, 5) not found: value start > start() > ^ > Error:(76, 24) value getNewApplication is not a member of > org.apache.spark.Logging > val newApp = super.getNewApplication() > ^ > Error:(137, 35) not found: type GetNewApplicationResponse > def verifyClusterResources(app: GetNewApplicationResponse) = { > ^ > Error:(156, 65) not found: type ApplicationSubmissionContext > def createApplicationSubmissionContext(appId: ApplicationId): > ApplicationSubmissionContext = { > ^ > Error:(156, 49) not found: type ApplicationId > def createApplicationSubmissionContext(appId: ApplicationId): > ApplicationSubmissionContext = { > ^ > Error:(118, 31) not found: type ApplicationId > def getAppStagingDir(appId: ApplicationId): String = { > ^ > Error:(224, 69) not found: type LocalResource > def prepareLocalResources(appStagingDir: String): HashMap[String, > LocalResource] = { > ^ > Error:(307, 39) not found: type LocalResource > localResources: HashMap[String, LocalResource], > ^ > Error:(343, 38) not found: type ContainerLaunchContext > env: HashMap[String, String]): ContainerLaunchContext = { > ^ > Error:(341, 15) not found: type GetNewApplicationResponse > newApp: GetNewApplicationResponse, > ^ > Error:(342, 39) not found: type LocalResource > localResources: HashMap[String, LocalResource], > ^ > Error:(426, 11) value submitApplication is not a member of > org.apache.spark.Logging > super.submitApplication(appContext) > ^ > Error:(423, 29) not found: type ApplicationSubmissionContext > def submitApp(appContext: ApplicationSubmissionContext) = { > ^ > Error:(429, 33) not found: type ApplicationId > def monitorApplication(appId: ApplicationId): Boolean = { > ^ > Error:(123, 25) not found: type YarnClusterMetrics > val clusterMetrics: YarnClusterMetrics = super.getYarnClusterMetrics > ^ > Error:(123, 52) value getYarnClusterMetrics is not a member of > org.apache.spark.Logging > val clusterMetrics: YarnClusterMetrics = super.getYarnClusterMetrics > ^ > Error:(127, 20) not found: type QueueInfo > val queueInfo: QueueInfo = super.getQueueInfo(args.amQueue) > ^ > Error:(127, 38) value getQueueInfo is not a member of > org.apache.spark.Logging > val queueInfo: QueueInfo = super.getQueueInfo(args.amQueue) > ^ > Error:(158, 22) not found: value Records > val appContext = > Records.newRecord(classOf[ApplicationSubmissionContext]) > ^ > Error:(219, 14) not found: value FileContext > val fc = FileContext.getFileContext(qualPath.toUri(), conf) > ^ > Error:(230, 29) not found: value Master > val delegTokenRenewer = Master.getMasterPrincipal(conf) > ^ > Error:(242, 13) value addDelegationTokens is not a member of > org.apache.hadoop.fs.FileSystem > dstFs.addDelegationTokens(delegTokenRenewer, credentials) > ^ > Error:(244, 42) not found: type LocalResource > val localResources = HashMap[String, LocalResource]() > ^ > Error:(302, 43) value addCredentials is not a member of > org.apache.hadoop.security.UserGroupInformation > UserGroupInformation.getCurrentUser().addCredentials(credentials) > ^ > Error:(323, 5) not found: value Apps > Apps.setEnvFromInputString(env, System.getenv("SPARK_YARN_USER_ENV")) > ^ > Error:(330, 36) not found: type ClientArguments > def userArgsToString(clientArgs: ClientArguments): String = { > ^ > Error:(345, 23) not found: value Records > val amContainer = Records.newRecord(classOf[ContainerLaunchContext]) > ^ > Error:(363, 16) not found: value Environment > new Path(Environment.PWD.$(), > YarnConfiguration.DEFAULT_CONTAINER_TEMP_DIR) + " " > ^ > Error:(392, 21) not found: value Environment > javaCommand = Environment.JAVA_HOME.$() + "/bin/java" > ^ > Error:(405, 16) not found: value ApplicationConstants > " 1> " + ApplicationConstants.LOG_DIR_EXPANSION_VAR + "/stdout" + > ^ > Error:(410, 22) not found: value Records > val capability = > Records.newRecord(classOf[Resource]).asInstanceOf[Resource] > ^ > Error:(410, 72) not found: type Resource > val capability = > Records.newRecord(classOf[Resource]).asInstanceOf[Resource] > ^ > Error:(434, 26) value getApplicationReport is not a member of > org.apache.spark.Logging > val report = super.getApplicationReport(appId) > ^ > Error:(474, 20) not found: type ClientArguments > val args = new ClientArguments(argStrings, sparkConf) > ^ > Error:(481, 31) not found: value YarnConfiguration > for (c <- conf.getStrings(YarnConfiguration.YARN_APPLICATION_CLASSPATH)) > { > ^ > Error:(487, 5) not found: value Apps > Apps.addToEnvironment(env, Environment.CLASSPATH.name, > Environment.PWD.$()) > ^ > Error:(490, 7) not found: value Apps > Apps.addToEnvironment(env, Environment.CLASSPATH.name, > Environment.PWD.$() + > ^ > Error:(496, 7) not found: value Apps > Apps.addToEnvironment(env, Environment.CLASSPATH.name, > Environment.PWD.$() + > ^ > Error:(499, 5) not found: value Apps > Apps.addToEnvironment(env, Environment.CLASSPATH.name, > Environment.PWD.$() + > ^ > Error:(504, 7) not found: value Apps > Apps.addToEnvironment(env, Environment.CLASSPATH.name, > Environment.PWD.$() + > ^ > Error:(507, 5) not found: value Apps > Apps.addToEnvironment(env, Environment.CLASSPATH.name, > Environment.PWD.$() + > ^ > > Any idea what's causing them, and maybe I am not using the best practice to > import Spark to IDE? > > I would appreciate any suggestion on the best practice to import Spark to > any IDE. > > Thank you > > -Mo