Re: Spark 1.6.2 version displayed as 1.6.1
This intrigued me as well. - Just for sure, I downloaded the 1.6.2 code and recompiled. - spark-shell and pyspark both show 1.6.2 as expected. Cheers On Mon, Jul 25, 2016 at 1:45 AM, Daniel Darabos < daniel.dara...@lynxanalytics.com> wrote: > Another possible explanation is that by accident you are still running > Spark 1.6.1. Which download are you using? This is what I see: > > $ ~/spark-1.6.2-bin-hadoop2.6/bin/spark-shell > log4j:WARN No appenders could be found for logger > (org.apache.hadoop.metrics2.lib.MutableMetricsFactory). > log4j:WARN Please initialize the log4j system properly. > log4j:WARN See http://logging.apache.org/log4j/1.2/faq.html#noconfig for > more info. > Using Spark's repl log4j profile: > org/apache/spark/log4j-defaults-repl.properties > To adjust logging level use sc.setLogLevel("INFO") > Welcome to > __ > / __/__ ___ _/ /__ > _\ \/ _ \/ _ `/ __/ '_/ >/___/ .__/\_,_/_/ /_/\_\ version 1.6.2 > /_/ > > > On Mon, Jul 25, 2016 at 7:45 AM, Sean Owenwrote: > >> Are you certain? looks like it was correct in the release: >> >> >> https://github.com/apache/spark/blob/v1.6.2/core/src/main/scala/org/apache/spark/package.scala >> >> >> >> On Mon, Jul 25, 2016 at 12:33 AM, Ascot Moss >> wrote: >> > Hi, >> > >> > I am trying to upgrade spark from 1.6.1 to 1.6.2, from 1.6.2 >> spark-shell, I >> > found the version is still displayed 1.6.1 >> > >> > Is this a minor typo/bug? >> > >> > Regards >> > >> > >> > >> > ### >> > >> > Welcome to >> > >> > __ >> > >> > / __/__ ___ _/ /__ >> > >> > _\ \/ _ \/ _ `/ __/ '_/ >> > >> >/___/ .__/\_,_/_/ /_/\_\ version 1.6.1 >> > >> > /_/ >> > >> > >> > >> >> - >> To unsubscribe e-mail: user-unsubscr...@spark.apache.org >> >> >
Re: Spark 1.6.2 version displayed as 1.6.1
Another possible explanation is that by accident you are still running Spark 1.6.1. Which download are you using? This is what I see: $ ~/spark-1.6.2-bin-hadoop2.6/bin/spark-shell log4j:WARN No appenders could be found for logger (org.apache.hadoop.metrics2.lib.MutableMetricsFactory). log4j:WARN Please initialize the log4j system properly. log4j:WARN See http://logging.apache.org/log4j/1.2/faq.html#noconfig for more info. Using Spark's repl log4j profile: org/apache/spark/log4j-defaults-repl.properties To adjust logging level use sc.setLogLevel("INFO") Welcome to __ / __/__ ___ _/ /__ _\ \/ _ \/ _ `/ __/ '_/ /___/ .__/\_,_/_/ /_/\_\ version 1.6.2 /_/ On Mon, Jul 25, 2016 at 7:45 AM, Sean Owenwrote: > Are you certain? looks like it was correct in the release: > > > https://github.com/apache/spark/blob/v1.6.2/core/src/main/scala/org/apache/spark/package.scala > > > > On Mon, Jul 25, 2016 at 12:33 AM, Ascot Moss wrote: > > Hi, > > > > I am trying to upgrade spark from 1.6.1 to 1.6.2, from 1.6.2 > spark-shell, I > > found the version is still displayed 1.6.1 > > > > Is this a minor typo/bug? > > > > Regards > > > > > > > > ### > > > > Welcome to > > > > __ > > > > / __/__ ___ _/ /__ > > > > _\ \/ _ \/ _ `/ __/ '_/ > > > >/___/ .__/\_,_/_/ /_/\_\ version 1.6.1 > > > > /_/ > > > > > > > > - > To unsubscribe e-mail: user-unsubscr...@spark.apache.org > >
Re: Spark 1.6.2 version displayed as 1.6.1
Are you certain? looks like it was correct in the release: https://github.com/apache/spark/blob/v1.6.2/core/src/main/scala/org/apache/spark/package.scala On Mon, Jul 25, 2016 at 12:33 AM, Ascot Mosswrote: > Hi, > > I am trying to upgrade spark from 1.6.1 to 1.6.2, from 1.6.2 spark-shell, I > found the version is still displayed 1.6.1 > > Is this a minor typo/bug? > > Regards > > > > ### > > Welcome to > > __ > > / __/__ ___ _/ /__ > > _\ \/ _ \/ _ `/ __/ '_/ > >/___/ .__/\_,_/_/ /_/\_\ version 1.6.1 > > /_/ > > > - To unsubscribe e-mail: user-unsubscr...@spark.apache.org
Spark 1.6.2 version displayed as 1.6.1
Hi, I am trying to upgrade spark from 1.6.1 to 1.6.2, from 1.6.2 spark-shell, I found the version is still displayed 1.6.1 Is this a minor typo/bug? Regards ### Welcome to __ / __/__ ___ _/ /__ _\ \/ _ \/ _ `/ __/ '_/ /___/ .__/\_,_/_/ /_/\_\ version 1.6.1 /_/