This intrigued me as well. - Just for sure, I downloaded the 1.6.2 code and recompiled. - spark-shell and pyspark both show 1.6.2 as expected.
Cheers On Mon, Jul 25, 2016 at 1:45 AM, Daniel Darabos < daniel.dara...@lynxanalytics.com> wrote: > Another possible explanation is that by accident you are still running > Spark 1.6.1. Which download are you using? This is what I see: > > $ ~/spark-1.6.2-bin-hadoop2.6/bin/spark-shell > log4j:WARN No appenders could be found for logger > (org.apache.hadoop.metrics2.lib.MutableMetricsFactory). > log4j:WARN Please initialize the log4j system properly. > log4j:WARN See http://logging.apache.org/log4j/1.2/faq.html#noconfig for > more info. > Using Spark's repl log4j profile: > org/apache/spark/log4j-defaults-repl.properties > To adjust logging level use sc.setLogLevel("INFO") > Welcome to > ____ __ > / __/__ ___ _____/ /__ > _\ \/ _ \/ _ `/ __/ '_/ > /___/ .__/\_,_/_/ /_/\_\ version 1.6.2 > /_/ > > > On Mon, Jul 25, 2016 at 7:45 AM, Sean Owen <so...@cloudera.com> wrote: > >> Are you certain? looks like it was correct in the release: >> >> >> https://github.com/apache/spark/blob/v1.6.2/core/src/main/scala/org/apache/spark/package.scala >> >> >> >> On Mon, Jul 25, 2016 at 12:33 AM, Ascot Moss <ascot.m...@gmail.com> >> wrote: >> > Hi, >> > >> > I am trying to upgrade spark from 1.6.1 to 1.6.2, from 1.6.2 >> spark-shell, I >> > found the version is still displayed 1.6.1 >> > >> > Is this a minor typo/bug? >> > >> > Regards >> > >> > >> > >> > ### >> > >> > Welcome to >> > >> > ____ __ >> > >> > / __/__ ___ _____/ /__ >> > >> > _\ \/ _ \/ _ `/ __/ '_/ >> > >> > /___/ .__/\_,_/_/ /_/\_\ version 1.6.1 >> > >> > /_/ >> > >> > >> > >> >> --------------------------------------------------------------------- >> To unsubscribe e-mail: user-unsubscr...@spark.apache.org >> >> >