Re: Koalas show data in IDE or pyspark

2019-04-30 Thread Manu Zhang
Hi, It seems koalas.DataFrame can't be displayed in terminal yet as in https://github.com/databricks/koalas/issues/150 and the work around is to convert it to pandas DataFrame. Thanks, Manu Zhang On Tue, Apr 30, 2019 at 2:46 PM Achilleus 003 wrote: > Hello Everyone, > > I have be

Re: [GraphX] Preserving Partitions when reading from HDFS

2019-04-15 Thread Manu Zhang
You may try `sparkContext.hadoopConfiguration().set("mapred.max.split.size", "33554432")` to tune the partition size when reading from HDFS. Thanks, Manu Zhang On Mon, Apr 15, 2019 at 11:28 PM M Bilal wrote: > Hi, > > I have implemented a custom partitioning

Spark driver crashed with internal error

2019-04-07 Thread Manu Zhang
ad*)+0x108* *C [libpthread.so.0+0x7dd5]* Has anyone seen this kind of error before ? I could provide more information if needed. Thanks, Manu Zhang

Re: mapreduce.input.fileinputformat.split.maxsize not working for spark 2.4.0

2019-02-24 Thread Manu Zhang
Is your application using Spark SQL / DataFrame API ? Is so, please try setting spark.sql.files.maxPartitionBytes to a larger value which is 128MB by default. Thanks, Manu Zhang On Feb 25, 2019, 2:58 AM +0800, Akshay Mendole , wrote: > Hi, >    We have dfs.blocksize configured to be

Re: getting error: value toDF is not a member of Seq[columns]

2018-09-05 Thread Manu Zhang
Have you tried adding Encoder for columns as suggested by Jungtaek Lim ? On Thu, Sep 6, 2018 at 6:24 AM Mich Talebzadeh wrote: > > Dr Mich Talebzadeh > > > > LinkedIn * > https://www.linkedin.com/profile/view?id=AAEWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw >

Re: java.lang.UnsupportedOperationException: No Encoder found for Set[String]

2018-08-16 Thread Manu Zhang
e changes in ExpressionEncoder between 2.3.0 and > 2.2.0 is not much but there might be many other classes underneath > that might have changed. > > On Thu, Aug 16, 2018 at 5:23 AM, Manu Zhang > wrote: > > Hi, > > > > It's added since Spark 2.3.0. > > > https://gi

Re: Unable to see completed application in Spark 2 history web UI

2018-08-16 Thread Manu Zhang
cloudera-scm is a super user and this > is why i was confused how it worked in spark 1.6 and not in spark 2.3 > > > On Thu, Aug 16, 2018 at 5:34 AM Manu Zhang > wrote: > >> If you are able to log onto the node where UI has been launched, then try >> `ps -aux | grep His

Re: java.lang.UnsupportedOperationException: No Encoder found for Set[String]

2018-08-16 Thread Manu Zhang
Hi, It's added since Spark 2.3.0. https://github.com/apache/spark/blob/master/sql/core/src/main/scala/org/apache/spark/sql/SQLImplicits.scala#L180 Regards, Manu Zhang On Thu, Aug 16, 2018 at 9:59 AM V0lleyBallJunki3 wrote: > Hello, > I am using Spark 2.2.2 with Scala 2.11.8. I wrote a

Re: Unable to see completed application in Spark 2 history web UI

2018-08-15 Thread Manu Zhang
se i'm using cloudera manager and i created a user for cloudera > manager and called it spark but this didn't solve me issue and here i'm > trying to find out the user for the spark hisotry UI. > > On Wed, Aug 15, 2018 at 5:11 PM Manu Zhang > wrote: > >> Hi Fawze

Re: Unable to see completed application in Spark 2 history web UI

2018-08-15 Thread Manu Zhang
Hi Fawze, A) The file permission is currently hard coded to 770 ( https://github.com/apache/spark/blob/branch-2.3/core/src/main/scala/org/apache/spark/scheduler/EventLoggingListener.scala#L287 ). B) I think add all users (including UI) to the group like Spark will do. On Wed, Aug 15, 2018 at

Re: Unable to see completed application in Spark 2 history web UI

2018-08-15 Thread Manu Zhang
able to see applications once changing file permissions to 777. Regards, Manu Zhang On Mon, Aug 13, 2018 at 4:53 PM Fawze Abujaber wrote: > Hi Guys, > > Any help here? > > On Wed, Aug 8, 2018 at 7:56 AM Fawze Abujaber wrote: > >> Hello Community, >> >> I'm u

Re: Split a row into multiple rows Java

2018-08-08 Thread Manu Zhang
) ).alias("temp")) val fields = split(col("temp"), ":") ndf.select(col("name"), col("otherName"), fields.getItem(1).alias("time"), fields.getItem(0).alias("value")) Regards, Manu Zhang On Wed, Aug 8, 2018 at 11:41 AM nookala wrote: