Dear list,

SBT compiles fine, but when I do the following:
sbt/sbt gen-idea
import project as SBT project to IDEA 13.1
Make Project
and these errors show up:

Error:(28, 8) object FileContext is not a member of package
org.apache.hadoop.fs
import org.apache.hadoop.fs.{FileContext, FileStatus, FileSystem, Path,
FileUtil}
       ^
Error:(31, 8) object Master is not a member of package
org.apache.hadoop.mapred
import org.apache.hadoop.mapred.Master
       ^
Error:(34, 26) object yarn is not a member of package org.apache.hadoop
import org.apache.hadoop.yarn.api._
                         ^
Error:(35, 26) object yarn is not a member of package org.apache.hadoop
import org.apache.hadoop.yarn.api.ApplicationConstants.Environment
                         ^
Error:(36, 26) object yarn is not a member of package org.apache.hadoop
import org.apache.hadoop.yarn.api.protocolrecords._
                         ^
Error:(37, 26) object yarn is not a member of package org.apache.hadoop
import org.apache.hadoop.yarn.api.records._
                         ^
Error:(38, 26) object yarn is not a member of package org.apache.hadoop
import org.apache.hadoop.yarn.client.YarnClientImpl
                         ^
Error:(39, 26) object yarn is not a member of package org.apache.hadoop
import org.apache.hadoop.yarn.conf.YarnConfiguration
                         ^
Error:(40, 26) object yarn is not a member of package org.apache.hadoop
import org.apache.hadoop.yarn.ipc.YarnRPC
                         ^
Error:(41, 26) object yarn is not a member of package org.apache.hadoop
import org.apache.hadoop.yarn.util.{Apps, Records}
                         ^
Error:(49, 11) not found: type YarnClientImpl
  extends YarnClientImpl with Logging {
          ^
Error:(48, 20) not found: type ClientArguments
class Client(args: ClientArguments, conf: Configuration, sparkConf:
SparkConf)
                   ^
Error:(51, 18) not found: type ClientArguments
  def this(args: ClientArguments, sparkConf: SparkConf) =
                 ^
Error:(54, 18) not found: type ClientArguments
  def this(args: ClientArguments) = this(args, new SparkConf())
                 ^
Error:(56, 12) not found: type YarnRPC
  var rpc: YarnRPC = YarnRPC.create(conf)
           ^
Error:(56, 22) not found: value YarnRPC
  var rpc: YarnRPC = YarnRPC.create(conf)
                     ^
Error:(57, 17) not found: type YarnConfiguration
  val yarnConf: YarnConfiguration = new YarnConfiguration(conf)
                ^
Error:(57, 41) not found: type YarnConfiguration
  val yarnConf: YarnConfiguration = new YarnConfiguration(conf)
                                        ^
Error:(58, 59) value getCredentials is not a member of
org.apache.hadoop.security.UserGroupInformation
  val credentials = UserGroupInformation.getCurrentUser().getCredentials()
                                                          ^
Error:(60, 34) not found: type ClientDistributedCacheManager
  private val distCacheMgr = new ClientDistributedCacheManager()
                                 ^
Error:(72, 5) not found: value init
    init(yarnConf)
    ^
Error:(73, 5) not found: value start
    start()
    ^
Error:(76, 24) value getNewApplication is not a member of
org.apache.spark.Logging
    val newApp = super.getNewApplication()
                       ^
Error:(137, 35) not found: type GetNewApplicationResponse
  def verifyClusterResources(app: GetNewApplicationResponse) = {
                                  ^
Error:(156, 65) not found: type ApplicationSubmissionContext
  def createApplicationSubmissionContext(appId: ApplicationId):
ApplicationSubmissionContext = {
                                                                ^
Error:(156, 49) not found: type ApplicationId
  def createApplicationSubmissionContext(appId: ApplicationId):
ApplicationSubmissionContext = {
                                                ^
Error:(118, 31) not found: type ApplicationId
  def getAppStagingDir(appId: ApplicationId): String = {
                              ^
Error:(224, 69) not found: type LocalResource
  def prepareLocalResources(appStagingDir: String): HashMap[String,
LocalResource] = {
                                                                    ^
Error:(307, 39) not found: type LocalResource
      localResources: HashMap[String, LocalResource],
                                      ^
Error:(343, 38) not found: type ContainerLaunchContext
      env: HashMap[String, String]): ContainerLaunchContext = {
                                     ^
Error:(341, 15) not found: type GetNewApplicationResponse
      newApp: GetNewApplicationResponse,
              ^
Error:(342, 39) not found: type LocalResource
      localResources: HashMap[String, LocalResource],
                                      ^
Error:(426, 11) value submitApplication is not a member of
org.apache.spark.Logging
    super.submitApplication(appContext)
          ^
Error:(423, 29) not found: type ApplicationSubmissionContext
  def submitApp(appContext: ApplicationSubmissionContext) = {
                            ^
Error:(429, 33) not found: type ApplicationId
  def monitorApplication(appId: ApplicationId): Boolean = {
                                ^
Error:(123, 25) not found: type YarnClusterMetrics
    val clusterMetrics: YarnClusterMetrics = super.getYarnClusterMetrics
                        ^
Error:(123, 52) value getYarnClusterMetrics is not a member of
org.apache.spark.Logging
    val clusterMetrics: YarnClusterMetrics = super.getYarnClusterMetrics
                                                   ^
Error:(127, 20) not found: type QueueInfo
    val queueInfo: QueueInfo = super.getQueueInfo(args.amQueue)
                   ^
Error:(127, 38) value getQueueInfo is not a member of
org.apache.spark.Logging
    val queueInfo: QueueInfo = super.getQueueInfo(args.amQueue)
                                     ^
Error:(158, 22) not found: value Records
    val appContext =
Records.newRecord(classOf[ApplicationSubmissionContext])
                     ^
Error:(219, 14) not found: value FileContext
    val fc = FileContext.getFileContext(qualPath.toUri(), conf)
             ^
Error:(230, 29) not found: value Master
    val delegTokenRenewer = Master.getMasterPrincipal(conf)
                            ^
Error:(242, 13) value addDelegationTokens is not a member of
org.apache.hadoop.fs.FileSystem
      dstFs.addDelegationTokens(delegTokenRenewer, credentials)
            ^
Error:(244, 42) not found: type LocalResource
    val localResources = HashMap[String, LocalResource]()
                                         ^
Error:(302, 43) value addCredentials is not a member of
org.apache.hadoop.security.UserGroupInformation
    UserGroupInformation.getCurrentUser().addCredentials(credentials)
                                          ^
Error:(323, 5) not found: value Apps
    Apps.setEnvFromInputString(env, System.getenv("SPARK_YARN_USER_ENV"))
    ^
Error:(330, 36) not found: type ClientArguments
  def userArgsToString(clientArgs: ClientArguments): String = {
                                   ^
Error:(345, 23) not found: value Records
    val amContainer = Records.newRecord(classOf[ContainerLaunchContext])
                      ^
Error:(363, 16) not found: value Environment
      new Path(Environment.PWD.$(),
YarnConfiguration.DEFAULT_CONTAINER_TEMP_DIR) + " "
               ^
Error:(392, 21) not found: value Environment
      javaCommand = Environment.JAVA_HOME.$() + "/bin/java"
                    ^
Error:(405, 16) not found: value ApplicationConstants
      " 1> " + ApplicationConstants.LOG_DIR_EXPANSION_VAR + "/stdout" +
               ^
Error:(410, 22) not found: value Records
    val capability =
Records.newRecord(classOf[Resource]).asInstanceOf[Resource]
                     ^
Error:(410, 72) not found: type Resource
    val capability =
Records.newRecord(classOf[Resource]).asInstanceOf[Resource]
                                                                       ^
Error:(434, 26) value getApplicationReport is not a member of
org.apache.spark.Logging
      val report = super.getApplicationReport(appId)
                         ^
Error:(474, 20) not found: type ClientArguments
    val args = new ClientArguments(argStrings, sparkConf)
                   ^
Error:(481, 31) not found: value YarnConfiguration
    for (c <-
conf.getStrings(YarnConfiguration.YARN_APPLICATION_CLASSPATH)) {
                              ^
Error:(487, 5) not found: value Apps
    Apps.addToEnvironment(env, Environment.CLASSPATH.name,
Environment.PWD.$())
    ^
Error:(490, 7) not found: value Apps
      Apps.addToEnvironment(env, Environment.CLASSPATH.name,
Environment.PWD.$() +
      ^
Error:(496, 7) not found: value Apps
      Apps.addToEnvironment(env, Environment.CLASSPATH.name,
Environment.PWD.$() +
      ^
Error:(499, 5) not found: value Apps
    Apps.addToEnvironment(env, Environment.CLASSPATH.name,
Environment.PWD.$() +
    ^
Error:(504, 7) not found: value Apps
      Apps.addToEnvironment(env, Environment.CLASSPATH.name,
Environment.PWD.$() +
      ^
Error:(507, 5) not found: value Apps
    Apps.addToEnvironment(env, Environment.CLASSPATH.name,
Environment.PWD.$() +
    ^

Any idea what's causing them, and maybe I am not using the best practice to
import Spark to IDE?

I would appreciate any suggestion on the best practice to import Spark to
any IDE.

Thank you

-Mo

Reply via email to