Cheers
Date: Wed, 28 Jan 2015 14:18:49 -0800
Subject: Re: unsubscribe
From: yuzhih...@gmail.com
To: 9000r...@gmail.com
CC: user@spark.apache.org
send an email to user-unsubscr...@spark.apache.org
Cheers
On Wed, Jan 28, 2015 at 2:16 PM, Abhi Basu 9000r...@gmail.com wrote:
--
Abhi Basu
Very well stated. Thanks for putting in the effort to formalize your thoughts
of which I agree entirely.
How are these type of decisions made traditionally in the Spark community? Is
there a formal process? What's the next step?
Thanks again
From: nicholas.cham...@gmail.com
Date: Thu, 22 Jan
Maybe I'm misunderstanding something here, but couldn't this be done with
broadcast variables? I there is the following caveat from the docs:
In addition, the object v should not be modified after it is broadcast in
order to ensure that all nodes get the same value of the broadcast variable
I found the following to be a good discussion of the same topic:
http://apache-spark-user-list.1001560.n3.nabble.com/The-concurrent-model-of-spark-job-stage-task-td13083.html
From: so...@cloudera.com
Date: Tue, 20 Jan 2015 10:02:20 +
Subject: Re: Does Spark automatically run different
If using Maven, one simply use whatever version they prefer and at build time
and the artifact using something like:
buildplugins plugin
groupIdorg.apache.maven.plugins/groupId
artifactIdmaven-shade-plugin/artifactIdexecutions
execution
Also, I used the following pattern to extract information from a file path and
add it to the output of a transformation:
https://gist.github.com/btiernay/1ad5e3dea08904fe07d9
You may find it useful as well.
Cheers,
Bob
From: btier...@hotmail.com
To: so...@cloudera.com;
You may also want to keep an eye on SPARK-5182 / SPARK-5302 which may help if
you are using Spark SQL. It should be noted that this is possible with
HiveContext today.
Cheers,
Bob
Date: Sun, 18 Jan 2015 08:47:06 +
Subject: Re: Directory / File Reading Patterns
From: so...@cloudera.com