Hi Team,
I have a code piece as follows.
try{
someDstream.someaction(...) //Step1
}catch{
case ex:Exception =>{
someDstream.someaction(...) //Step2
}
}
When I get an exception for current batch, Step2 executes as
m: Akhil Das [mailto:ak...@sigmoidanalytics.com]
Sent: Wednesday, September 16, 2015 12:24 PM
To: Samya MAITI <samya.ma...@amadeus.com>
Cc: user@spark.apache.org
Subject: Re: Getting parent RDD
How many RDDs are you having in that stream? If its a single RDD then you
could do a .f
Hi Team
I have the below situation.
val ssc =
val msgStream = . //SparkKafkaDirectAPI
val wordCountPair = TransformStream.transform(msgStream)
/wordCountPair.foreachRDD(rdd =>
try{
//Some action that causes exception
}catch {
case ex1 : Exception => {
Hi Team,
I am facing this issue where in I can't figure out why the exception is
handled the first time an exception is thrown in the stream processing
action, but is ignored the second time.
PFB my code base.
object Boot extends App {
//Load the configuration
val config =
Thanks Ameya.
From: ameya [via Apache Spark User List]
[mailto:ml-node+s1001560n24650...@n3.nabble.com]
Sent: Friday, September 11, 2015 4:12 AM
To: Samya MAITI <samya.ma...@amadeus.com>
Subject: Re: Maintaining Kafka Direct API Offsets
So I added something like this:
Runtime.getR
Hi Ameya,
Plz suggest, when you say graceful shut-down, what exactly did you handle?
Thanks.
Thanks,
Sam
--
View this message in context:
http://apache-spark-user-list.1001560.n3.nabble.com/Maintaining-Kafka-Direct-API-Offsets-tp24246p24636.html
Sent from the Apache Spark User List mailing
From: Cody Koeninger [mailto:c...@koeninger.org]
Sent: Thursday, September 10, 2015 1:13 AM
To: Samya MAITI <samya.ma...@amadeus.com>
Cc: user@spark.apache.org
Subject: Re: Spark streaming -> cassandra : Fault Tolerance
It's been a while since I've looked at the cassandra connector, s
Hi Team,
I have an sample spark application which reads from Kafka using direct API &
then does some transformation & stores to cassandra (using
saveToCassandra()).
If Cassandra goes down, then application logs NoHostAvailable exception (as
expected). But in the mean time the new incoming
? Is it
in user control?
Regards,
Sam
From: Jem Tucker [mailto:jem.tuc...@gmail.com]
Sent: Wednesday, August 26, 2015 2:26 PM
To: Samya MAITI samya.ma...@amadeus.com; user@spark.apache.org
Subject: Re: Relation between threads and executor core
Hi Samya,
When submitting an application with spark-submit
Hi All,
Few basic queries :-
1. Is there a way we can control the number of threads per executor core?
2. Does this parameter “executor-cores” also has say in deciding how many
threads to be run?
Regards,
Sam
--
View this message in context:
)
connection object per executor, that is shared between tasks ?
2. If the above answer is YES, is there a way to create a connectionPool for
each executor, so that multiple task can dump data to cassandra in parallel?
Regards,
Samya
--
View this message in context:
http://apache-spark-user-list
Really good list to brush up basics.
Just one input, regarding
* An RDD's processing is scheduled by driver's jobscheduler as a job. At a
given point of time only one job is active. So, if one job is executing the
other jobs are queued.
We can have multiple jobs running in a given
Hi TD,
I want to append my record to a AVRO file which will be later used for querying.
Having a single file is not mandatory for us but then how can we make the
executors append the AVRO data to multiple files.
Thanks,
Sam
On Mar 12, 2015, at 4:09 AM, Tathagata Das
Thanks TD.
On Wed, Dec 31, 2014 at 7:19 AM, Tathagata Das tathagata.das1...@gmail.com
wrote:
1. Of course, a single block / partition has many Kafka messages, and
from different Kafka topics interleaved together. The message count is
not related to the block count. Any message received within
Thank Sean.
That was helpful.
Regards,
Sam
On Dec 30, 2014, at 4:12 PM, Sean Owen so...@cloudera.com wrote:
The DStream model is one RDD of data per interval, yes. foreachRDD
performs an operation on each RDD in the stream, which means it is
executed once* for the one RDD in each interval.
15 matches
Mail list logo