Spark-SQL: SchemaRDD - ClassCastException

2014-10-08 Thread Ranga
approaches that I should be looking at? Thanks for your help. - Ranga

Re: Spark-SQL: SchemaRDD - ClassCastException

2014-10-08 Thread Ranga
) // This query throws the exception when I collect the results I tried adding the cast to the aggRdd query above and that didn't help. - Ranga On Wed, Oct 8, 2014 at 3:52 PM, Michael Armbrust mich...@databricks.com wrote: Using SUM on a string should automatically cast the column. Also you

Re: Spark-SQL: SchemaRDD - ClassCastException

2014-10-08 Thread Ranga
:11 PM, Michael Armbrust mich...@databricks.com wrote: Which version of Spark are you running? On Wed, Oct 8, 2014 at 4:18 PM, Ranga sra...@gmail.com wrote: Thanks Michael. Should the cast be done in the source RDD or while doing the SUM? To give a better picture here is the code sequence

Re: Spark-SQL: SchemaRDD - ClassCastException

2014-10-08 Thread Ranga
as int) ... from table Any other pointers? Thanks for the help. - Ranga On Wed, Oct 8, 2014 at 5:20 PM, Ranga sra...@gmail.com wrote: Sorry. Its 1.1.0. After digging a bit more into this, it seems like the OpenCSV Deseralizer converts all the columns to a String type. This maybe throwing

Re: Spark-SQL: SchemaRDD - ClassCastException

2014-10-09 Thread Ranga
with this for now. - Ranga On Wed, Oct 8, 2014 at 9:18 PM, Ranga sra...@gmail.com wrote: This is a bit strange. When I print the schema for the RDD, it reflects the correct data type for each column. But doing any kind of mathematical calculation seems to result in ClassCastException. Here

S3 Bucket Access

2014-10-13 Thread Ranga
. - Ranga

Re: S3 Bucket Access

2014-10-13 Thread Ranga
Is there a way to specify a request header during the sparkContext.textFile call? - Ranga On Mon, Oct 13, 2014 at 11:03 AM, Ranga sra...@gmail.com wrote: Hi I am trying to access files/buckets in S3 and encountering a permissions issue. The buckets are configured to authenticate using

Re: S3 Bucket Access

2014-10-13 Thread Ranga
Hi Daniil Could you provide some more details on how the cluster should be launched/configured? The EC2 instance that I am dealing with uses the concept of IAMRoles. I don't have any keyfile to specify to the spark-ec2 script. Thanks for your help. - Ranga On Mon, Oct 13, 2014 at 3:04 PM

Re: S3 Bucket Access

2014-10-14 Thread Ranga
. This may or may not be possible in the environment I am in (from a policy perspective) - Ranga On Tue, Oct 14, 2014 at 4:21 AM, Rafal Kwasny m...@entropy.be wrote: Hi, keep in mind that you're going to have a bad time if your secret key contains a / This is due to old and stupid hadoop bug

Re: S3 Bucket Access

2014-10-14 Thread Ranga
based on the IAMRole but they generally expire in an hour - If Spark is not able to use the IAMRole credentials, I may have to generate a static key-id/secret. This may or may not be possible in the environment I am in (from a policy perspective) - Ranga On Tue, Oct 14

Re: S3 Bucket Access

2014-10-14 Thread Ranga
One related question. Could I specify the com.amazonaws.services.s3.AmazonS3Client implementation for the fs.s3.impl parameter? Let me try that and update this thread with my findings. On Tue, Oct 14, 2014 at 10:48 AM, Ranga sra...@gmail.com wrote: Thanks for the input. Yes, I did use

Re: S3 Bucket Access

2014-10-14 Thread Ranga
use AWS SDK in your application to provide AWS credentials? https://github.com/seratch/AWScala On Oct 14, 2014, at 11:10 AM, Ranga sra...@gmail.com wrote: One related question. Could I specify the com.amazonaws.services.s3.AmazonS3Client implementation for the fs.s3.impl parameter? Let

Re: RDDs being cleaned too fast

2014-12-11 Thread Ranga
though. You could also increase the spark.storage.memoryFraction if that is an option. - Ranga On Wed, Dec 10, 2014 at 10:23 PM, Aaron Davidson ilike...@gmail.com wrote: The ContextCleaner uncaches RDDs that have gone out of scope on the driver. So it's possible that the given RDD

Re: Spark Streaming: HiveContext within Custom Actor

2014-12-30 Thread Ranga
. - Ranga -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/Spark-Streaming-HiveContext-within-Custom-Actor-tp20892.html Sent from the Apache Spark User List mailing list archive at Nabble.com

Re: StorageLevel: OFF_HEAP

2015-03-18 Thread Ranga
Thanks for the information. Will rebuild with 0.6.0 till the patch is merged. On Tue, Mar 17, 2015 at 7:24 PM, Ted Yu yuzhih...@gmail.com wrote: Ranga: Take a look at https://github.com/apache/spark/pull/4867 Cheers On Tue, Mar 17, 2015 at 6:08 PM, fightf...@163.com fightf...@163.com

StorageLevel: OFF_HEAP

2015-03-17 Thread Ranga
in a production environment by anybody in this group? Appreciate your help with this. - Ranga

Re: StorageLevel: OFF_HEAP

2015-03-18 Thread Ranga
to create tachyon dir in /tmp_spark_tachyon/spark-e3538a20-5e42-48a4-ad67-4b97aded90e4/driver Thanks for any other pointers. - Ranga On Wed, Mar 18, 2015 at 9:53 AM, Ranga sra...@gmail.com wrote: Thanks for the information. Will rebuild with 0.6.0 till the patch is merged. On Tue, Mar 17

Re: StorageLevel: OFF_HEAP

2015-03-18 Thread Ranga
Thanks Ted. Will do. On Wed, Mar 18, 2015 at 2:27 PM, Ted Yu yuzhih...@gmail.com wrote: Ranga: Please apply the patch from: https://github.com/apache/spark/pull/4867 And rebuild Spark - the build would use Tachyon-0.6.1 Cheers On Wed, Mar 18, 2015 at 2:23 PM, Ranga sra...@gmail.com

Re: StorageLevel: OFF_HEAP

2015-03-18 Thread Ranga
Hi Haoyuan No. I assumed that Spark-1.3.0 was already built with Tachyon-0.6.0. If not, I can rebuild and try. Could you let me know how to rebuild with 0.6.0? Thanks for your help. - Ranga On Wed, Mar 18, 2015 at 12:59 PM, Haoyuan Li haoyuan...@gmail.com wrote: Did you recompile

PySpark tests are failed with the java.util.ServiceConfigurationError: org.apache.spark.sql.sources.DataSourceRegister: Provider org.apache.spark.sql.sources.FakeSourceOne not found

2023-04-12 Thread Ranga Reddy
lang.Thread.run(Thread.java:750) Could someone help me how to proceed further? -- Thanks and Regards *Ranga Reddy* *--* *Bangalore, Karnataka, India* *Mobile : +91-9986183183 | Email: rangareddy.av...@gmail.com *