[ANNOUNCE] Announcing Apache Spark 2.3.2

2018-09-26 Thread Saisai Shao
We are happy to announce the availability of Spark 2.3.2! Apache Spark 2.3.2 is a maintenance release, based on the branch-2.3 maintenance branch of Spark. We strongly recommend all 2.3.x users to upgrade to this stable release. To download Spark 2.3.2, head over to the download page:

Re: Adding Extension to Load Custom functions into Thriftserver/SqlShell

2018-09-26 Thread Mark Hamstra
You're talking about users starting Thriftserver or SqlShell from the command line, right? It's much easier if you are starting a Thriftserver programmatically so that you can register functions when initializing a SparkContext and then HiveThriftServer2.startWithContext using that context. On

Adding Extension to Load Custom functions into Thriftserver/SqlShell

2018-09-26 Thread Russell Spitzer
I've been looking recently on possible avenues to load new functions into the Thriftserver and SqlShell at launch time. I basically want to preload a set of functions in addition to those already present in the Spark Code. I'm not sure there is at present a way to do this and I was wondering if

RE: Python kubernetes spark 2.4 branch

2018-09-26 Thread Garlapati, Suryanarayana (Nokia - IN/Bangalore)
Hi Ilan/Yinan, My observation is as follows: The dependent files specified with “--py-files http://10.75.145.25:80/Spark/getNN.py” are being downloaded and available in the container at “/var/data/spark-c163f15e-d59d-4975-b9be-91b6be062da9/spark-61094ca2-125b-48de-a154-214304dbe74/”. I guess we

Re: Filtering based on a float value with more than one decimal place not working correctly in Pyspark dataframe

2018-09-26 Thread Sean Owen
Is this not just a case of floating-point literals not being exact? this is expressed in Python, not SQL. On Wed, Sep 26, 2018 at 12:46 AM Meethu Mathew wrote: > Hi all, > > I tried the following code and the output was not as expected. > > schema = StructType([StructField('Id', StringType(),

Unsubscribe

2018-09-26 Thread Iryna Kharaborkina

Re: Filtering based on a float value with more than one decimal place not working correctly in Pyspark dataframe

2018-09-26 Thread Sandeep Katta
I think it is similar to the one SPARK-25452 Regards Sandeep Katta On Wed, 26 Sep 2018 at 11:16 AM, Meethu Mathew wrote: > Hi all, > > I tried the following code and the output was not as expected. > > schema = StructType([StructField('Id', StringType(), False), >>