Trying to build with sbt with the following dependencies

libraryDependencies += "org.apache.spark" %% "spark-core" % "1.5.1" %
"provided"
libraryDependencies += "org.apache.spark" %% "spark-sql" % "1.5.1"  %
"provided"
libraryDependencies += "org.apache.spark" %% "spark-hive" % "1.5.1" %
"provided"
libraryDependencies += "junit" % "junit" % "4.12"
libraryDependencies += "org.scala-sbt" % "test-interface" % "1.0"
libraryDependencies += "org.apache.spark" %% "spark-streaming" % "1.6.1" %
"provided"
libraryDependencies += "org.apache.spark" %% "spark-streaming-kafka" %
"1.6.1"
libraryDependencies += "org.scalactic" %% "scalactic" % "2.2.6"
libraryDependencies += "org.scalatest" %% "scalatest" % "2.2.6"
libraryDependencies += "org.apache.spark" % "spark-core_2.10" % "1.5.1"
libraryDependencies += "org.apache.spark" %
"spark-streaming-kafka-assembly_2.10" % "1.6.1"


[error]
/data6/hduser/scala/CEP_assembly/src/main/scala/myPackage/CEP_assemly.scala:35:
object SparkFunSuite is not a member of package org.apache.spark
[error] import org.apache.spark.{SparkConf, SparkContext, SparkFunSuite}
[error]        ^
[error]
/data6/hduser/scala/CEP_assembly/src/main/scala/myPackage/CEP_assemly.scala:47:
not found: type SparkFunSuite
[error]   extends SparkFunSuite
[error]           ^
[error]
/data6/hduser/scala/CEP_assembly/src/main/scala/myPackage/CEP_assemly.scala:88:
package test is not a value
[error]   test("basic stream receiving with multiple topics and smallest
starting offset") {
[error]   ^
[error]
/data6/hduser/scala/CEP_assembly/src/main/scala/myPackage/CEP_assemly.scala:146:
package test is not a value
[error]   test("receiving from largest starting offset") {
[error]   ^
[error]
/data6/hduser/scala/CEP_assembly/src/main/scala/myPackage/CEP_assemly.scala:191:
package test is not a value
[error]   test("creating stream by offset") {
[error]   ^
[error]
/data6/hduser/scala/CEP_assembly/src/main/scala/myPackage/CEP_assemly.scala:237:
package test is not a value
[error]   test("offset recovery") {
[error]   ^
[error]
/data6/hduser/scala/CEP_assembly/src/main/scala/myPackage/CEP_assemly.scala:319:
package test is not a value
[error]   test("Direct Kafka stream report input information") {
[error]   ^
[error]
/data6/hduser/scala/CEP_assembly/src/main/scala/myPackage/CEP_assemly.scala:358:
package test is not a value
[error]   test("maxMessagesPerPartition with backpressure disabled") {
[error]   ^
[error]
/data6/hduser/scala/CEP_assembly/src/main/scala/myPackage/CEP_assemly.scala:367:
package test is not a value
[error]   test("maxMessagesPerPartition with no lag") {
[error]   ^
[error]
/data6/hduser/scala/CEP_assembly/src/main/scala/myPackage/CEP_assemly.scala:376:
package test is not a value
[error]   test("maxMessagesPerPartition respects max rate") {
[error]   ^
[error]
/data6/hduser/scala/CEP_assembly/src/main/scala/myPackage/CEP_assemly.scala:386:
package test is not a value
[error]   test("using rate controller") {
[error]   ^
[error]
/data6/hduser/scala/CEP_assembly/src/main/scala/myPackage/CEP_assemly.scala:531:
object WindowState is not a member of package
org.apache.spark.streaming.dstream
[error] import org.apache.spark.streaming.dstream.WindowState
[error]        ^
[error]
/data6/hduser/scala/CEP_assembly/src/main/scala/myPackage/CEP_assemly.scala:578:
not found: type WindowState
[error]     def rise(in: Tick, ew: WindowState): Boolean = {
[error]                            ^
[error]
/data6/hduser/scala/CEP_assembly/src/main/scala/myPackage/CEP_assemly.scala:582:
not found: type WindowState
[error]     def drop(in: Tick, ew: WindowState): Boolean = {
[error]                            ^
[error]
/data6/hduser/scala/CEP_assembly/src/main/scala/myPackage/CEP_assemly.scala:586:
not found: type WindowState
[error]     def deep(in: Tick, ew: WindowState): Boolean = {
[error]                            ^
[error]
/data6/hduser/scala/CEP_assembly/src/main/scala/myPackage/CEP_assemly.scala:592:
not found: type WindowState
[error]     val predicateMapping: Map[String, (Tick, WindowState) =>
Boolean] =
[error]                                              ^
[error]
/data6/hduser/scala/CEP_assembly/src/main/scala/myPackage/CEP_assemly.scala:595:
value patternMatchByKeyAndWindow is not a member of
org.apache.spark.streaming.dstream.DStream[(String, Tick)]
[error]     val matches = kvTicks.patternMatchByKeyAndWindow("rise drop
[rise ]+ deep".r,
[error]                           ^
[error]
/data6/hduser/scala/CEP_assembly/src/main/scala/myPackage/CEP_assemly.scala:646:
not found: type WindowState
[error]     def rise(in: Tick, ew: WindowState): Boolean = {
[error]                            ^
[error]
/data6/hduser/scala/CEP_assembly/src/main/scala/myPackage/CEP_assemly.scala:650:
not found: type WindowState
[error]     def drop(in: Tick, ew: WindowState): Boolean = {
[error]                            ^
[error]
/data6/hduser/scala/CEP_assembly/src/main/scala/myPackage/CEP_assemly.scala:654:
not found: type WindowState
[error]     def deep(in: Tick, ew: WindowState): Boolean = {
[error]                            ^
[error]
/data6/hduser/scala/CEP_assembly/src/main/scala/myPackage/CEP_assemly.scala:660:
not found: type WindowState
[error]     val predicateMapping: Map[String, (Tick, WindowState) =>
Boolean] =
[error]                                              ^
[error]
/data6/hduser/scala/CEP_assembly/src/main/scala/myPackage/CEP_assemly.scala:663:
value patternMatchByWindow is not a member of
org.apache.spark.streaming.dstream.DStream[(Long, Tick)]
[error]     val matches = kvTicks.patternMatchByWindow("rise drop [rise ]+
deep".r,
[error]                           ^
[error] 22 errors found
[error] (compile:compileIncremental) Compilation failed


Dr Mich Talebzadeh



LinkedIn * 
https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw
<https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>*



http://talebzadehmich.wordpress.com



On 22 April 2016 at 14:53, Ted Yu <yuzhih...@gmail.com> wrote:

> Normally Logging would be included in spark-shell session since spark-core
> jar is imported by default:
>
> scala> import org.apache.spark.internal.Logging
> import org.apache.spark.internal.Logging
>
> See this JIRA:
>
> [SPARK-13928] Move org.apache.spark.Logging into
> org.apache.spark.internal.Logging
>
> In 1.6.x release, Logging was at org.apache.spark.Logging
>
> FYI
>
> On Fri, Apr 22, 2016 at 12:21 AM, Mich Talebzadeh <
> mich.talebza...@gmail.com> wrote:
>
>>
>> Hi,
>>
>> Anyone know which jar file has  import org.apache.spark.internal.Logging?
>>
>> I tried *spark-core_2.10-1.5.1.jar *
>>
>> but does not seem to work
>>
>> scala> import org.apache.spark.internal.Logging
>>
>> <console>:57: error: object internal is not a member of package
>> org.apache.spark
>>          import org.apache.spark.internal.Logging
>>
>> Thanks
>>
>> Dr Mich Talebzadeh
>>
>>
>>
>> LinkedIn * 
>> https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw
>> <https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>*
>>
>>
>>
>> http://talebzadehmich.wordpress.com
>>
>>
>>
>
>

Reply via email to