-dev, +user (this is not a question about development of Spark itself so you’ll get more answers in the user mailing list)
First up let me say that I don’t really know how this could be done - I’m sure it would be possible with enough tinkering but it’s not clear what you are trying to achieve. Spark is a distributed processing system, it has multiple JVMs running on different machines that each run a small part of the overall processing. Unless you have some sort of idea to have multiple C++ processes collocated with the distributed JVMs using named memory mapped files doesn’t make architectural sense. ------------------------------------------------------------------------------- Robin East Spark GraphX in Action Michael Malak and Robin East Manning Publications Co. http://www.manning.com/books/spark-graphx-in-action <http://www.manning.com/books/spark-graphx-in-action> > On 6 Dec 2015, at 20:43, Jia <jacqueline...@gmail.com> wrote: > > Dears, for one project, I need to implement something so Spark can read data > from a C++ process. > To provide high performance, I really hope to implement this through shared > memory between the C++ process and Java JVM process. > It seems it may be possible to use named memory mapped files and JNI to do > this, but I wonder whether there is any existing efforts or more efficient > approach to do this? > Thank you very much! > > Best Regards, > Jia > > > --------------------------------------------------------------------- > To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org > For additional commands, e-mail: dev-h...@spark.apache.org >