Re: Dataset -- Schema for type scala.collection.Set[scala.Int] is not supported

2019-08-09 Thread Mohit Jaggi
switched to immutable.Set and it works. this is weird as the code in ScalaReflection.scala seems to support scala.collection.Set cc: dev list, in case this is a bug On Thu, Aug 8, 2019 at 8:41 PM Mohit Jaggi wrote: > Is this not supported? I found this diff > <https://github.com/apa

Re: Pyspark access to scala/java libraries

2018-07-17 Thread Mohit Jaggi
ala-function-from-a-task >> >> ​Sent with ProtonMail Secure Email.​ >> >> ‐‐‐ Original Message ‐‐‐ >> >> On July 15, 2018 8:01 AM, Mohit Jaggi wrote: >> >> > Trying again…anyone know how to make this work? >> > >> > > On Jul

Re: Pyspark access to scala/java libraries

2018-07-15 Thread Mohit Jaggi
Trying again…anyone know how to make this work? > On Jul 9, 2018, at 3:45 PM, Mohit Jaggi wrote: > > Folks, > I am writing some Scala/Java code and want it to be usable from pyspark. > > For example: > class MyStuff(addend: Int) { > def myMapFunction(x: Int) = x

Fwd: SparkILoop doesn't run

2016-11-17 Thread Mohit Jaggi
ForkJoinWorkerThread.java:107) Process finished with exit code 137 (interrupted by signal 9: SIGKILL) Mohit Jaggi Founder, Data Orchard LLC www.dataorchardllc.com

SparkILoop doesn't run

2016-11-16 Thread Mohit Jaggi
doExec(ForkJoinTask.java:260) at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339) at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107) Process finis

Re: Model abstract class in spark ml

2016-08-30 Thread Mohit Jaggi
new AA(1) } Mohit Jaggi Founder, Data Orchard LLC www.dataorchardllc.com > On Aug 30, 2016, at 9:51 PM, Mohit Jaggi <mohitja...@gmail.com> wrote: > > thanks Sean. I am cross posting on dev to see why the code was written that > way. Perhaps, this.type doesn’t do what i

Re: Model abstract class in spark ml

2016-08-30 Thread Mohit Jaggi
thanks Sean. I am cross posting on dev to see why the code was written that way. Perhaps, this.type doesn’t do what is needed. Mohit Jaggi Founder, Data Orchard LLC www.dataorchardllc.com On Aug 30, 2016, at 2:08 PM, Sean Owen <so...@cloudera.com> wrote: I think it's imitating, for e

Re: Model abstract class in spark ml

2016-08-30 Thread Mohit Jaggi
thanks Sean. I am cross posting on dev to see why the code was written that way. Perhaps, this.type doesn’t do what is needed. Mohit Jaggi Founder, Data Orchard LLC www.dataorchardllc.com > On Aug 30, 2016, at 2:08 PM, Sean Owen <so...@cloudera.com> wrote: > > I think

Re: [spark-csv] how to build with Hadoop 2.6.0?

2015-08-20 Thread Mohit Jaggi
com.databricks.spark.csv.util.TextFile has hadoop imports. I figured out that the answer to my question is just to add libraryDependencies += org.apache.hadoop % hadoop-client % 2.6.0. But i still wonder where is this 2.2.0 default comes from. From:Mohit Jaggi mohitja...@gmail.com

Re: [spark-csv] how to build with Hadoop 2.6.0?

2015-08-19 Thread Mohit Jaggi
spark-csv should not depend on hadoop On Sun, Aug 16, 2015 at 9:05 AM, Gil Vernik g...@il.ibm.com wrote: I would like to build spark-csv with Hadoop 2.6.0 I noticed that when i build it with sbt/sbt ++2.10.4 package it build it with Hadoop 2.2.0 ( at least this is what i saw in the .ivy2

Re: Grouping runs of elements in a RDD

2015-07-02 Thread Mohit Jaggi
be moved to spark-core. not sure if that happened ] - previous posts --- http://spark.apache.org/docs/1.4.0/api/scala/index.html#org.apache.spark.mllib.rdd.RDDFunctions On Fri, Jan 30, 2015 at 12:27 AM, Mohit Jaggi mohitja...@gmail.com wrote: http://mail-archives.apache.org/mod_mbox/spark

Re: RDD.combineBy without intermediate (k,v) pair allocation

2015-01-29 Thread Mohit Jaggi
key and value and then using combine, however. — FG On Tue, Jan 27, 2015 at 10:17 PM, Mohit Jaggi mohitja...@gmail.com mailto:mohitja...@gmail.com wrote: Hi All, I have a use case where I have an RDD (not a k,v pair) where I want to do a combineByKey() operation. I can do

greeting from new member and jira 3489

2014-09-16 Thread Mohit Jaggi
https://issues.apache.org/jira/browse/SPARK-3489 Folks, I am Mohit Jaggi and I work for Ayasdi Inc. After experimenting with Spark for a while and discovering its awesomeness(!) I made an attempt to provide a wrapper API that looks like R and/or pandas dataframe. https://github.com