PySpark tests are failed with the java.util.ServiceConfigurationError: org.apache.spark.sql.sources.DataSourceRegister: Provider org.apache.spark.sql.sources.FakeSourceOne not found

2023-04-12 Thread Ranga Reddy
lang.Thread.run(Thread.java:750) Could someone help me how to proceed further? -- Thanks and Regards *Ranga Reddy* *--* *Bangalore, Karnataka, India* *Mobile : +91-9986183183 | Email: rangareddy.av...@gmail.com *

Re: StorageLevel: OFF_HEAP

2015-03-18 Thread Ranga
Thanks for the information. Will rebuild with 0.6.0 till the patch is merged. On Tue, Mar 17, 2015 at 7:24 PM, Ted Yu yuzhih...@gmail.com wrote: Ranga: Take a look at https://github.com/apache/spark/pull/4867 Cheers On Tue, Mar 17, 2015 at 6:08 PM, fightf...@163.com fightf...@163.com

Re: StorageLevel: OFF_HEAP

2015-03-18 Thread Ranga
to create tachyon dir in /tmp_spark_tachyon/spark-e3538a20-5e42-48a4-ad67-4b97aded90e4/driver Thanks for any other pointers. - Ranga On Wed, Mar 18, 2015 at 9:53 AM, Ranga sra...@gmail.com wrote: Thanks for the information. Will rebuild with 0.6.0 till the patch is merged. On Tue, Mar 17

Re: StorageLevel: OFF_HEAP

2015-03-18 Thread Ranga
Thanks Ted. Will do. On Wed, Mar 18, 2015 at 2:27 PM, Ted Yu yuzhih...@gmail.com wrote: Ranga: Please apply the patch from: https://github.com/apache/spark/pull/4867 And rebuild Spark - the build would use Tachyon-0.6.1 Cheers On Wed, Mar 18, 2015 at 2:23 PM, Ranga sra...@gmail.com

Re: StorageLevel: OFF_HEAP

2015-03-18 Thread Ranga
Hi Haoyuan No. I assumed that Spark-1.3.0 was already built with Tachyon-0.6.0. If not, I can rebuild and try. Could you let me know how to rebuild with 0.6.0? Thanks for your help. - Ranga On Wed, Mar 18, 2015 at 12:59 PM, Haoyuan Li haoyuan...@gmail.com wrote: Did you recompile

StorageLevel: OFF_HEAP

2015-03-17 Thread Ranga
in a production environment by anybody in this group? Appreciate your help with this. - Ranga

Re: Spark Streaming: HiveContext within Custom Actor

2014-12-30 Thread Ranga
. - Ranga -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/Spark-Streaming-HiveContext-within-Custom-Actor-tp20892.html Sent from the Apache Spark User List mailing list archive at Nabble.com

Re: RDDs being cleaned too fast

2014-12-11 Thread Ranga
though. You could also increase the spark.storage.memoryFraction if that is an option. - Ranga On Wed, Dec 10, 2014 at 10:23 PM, Aaron Davidson ilike...@gmail.com wrote: The ContextCleaner uncaches RDDs that have gone out of scope on the driver. So it's possible that the given RDD

Re: S3 Bucket Access

2014-10-14 Thread Ranga
. This may or may not be possible in the environment I am in (from a policy perspective) - Ranga On Tue, Oct 14, 2014 at 4:21 AM, Rafal Kwasny m...@entropy.be wrote: Hi, keep in mind that you're going to have a bad time if your secret key contains a / This is due to old and stupid hadoop bug

Re: S3 Bucket Access

2014-10-14 Thread Ranga
based on the IAMRole but they generally expire in an hour - If Spark is not able to use the IAMRole credentials, I may have to generate a static key-id/secret. This may or may not be possible in the environment I am in (from a policy perspective) - Ranga On Tue, Oct 14

Re: S3 Bucket Access

2014-10-14 Thread Ranga
One related question. Could I specify the com.amazonaws.services.s3.AmazonS3Client implementation for the fs.s3.impl parameter? Let me try that and update this thread with my findings. On Tue, Oct 14, 2014 at 10:48 AM, Ranga sra...@gmail.com wrote: Thanks for the input. Yes, I did use

Re: S3 Bucket Access

2014-10-14 Thread Ranga
use AWS SDK in your application to provide AWS credentials? https://github.com/seratch/AWScala On Oct 14, 2014, at 11:10 AM, Ranga sra...@gmail.com wrote: One related question. Could I specify the com.amazonaws.services.s3.AmazonS3Client implementation for the fs.s3.impl parameter? Let

S3 Bucket Access

2014-10-13 Thread Ranga
. - Ranga

Re: S3 Bucket Access

2014-10-13 Thread Ranga
Is there a way to specify a request header during the sparkContext.textFile call? - Ranga On Mon, Oct 13, 2014 at 11:03 AM, Ranga sra...@gmail.com wrote: Hi I am trying to access files/buckets in S3 and encountering a permissions issue. The buckets are configured to authenticate using

Re: S3 Bucket Access

2014-10-13 Thread Ranga
Hi Daniil Could you provide some more details on how the cluster should be launched/configured? The EC2 instance that I am dealing with uses the concept of IAMRoles. I don't have any keyfile to specify to the spark-ec2 script. Thanks for your help. - Ranga On Mon, Oct 13, 2014 at 3:04 PM

Re: Spark-SQL: SchemaRDD - ClassCastException

2014-10-09 Thread Ranga
with this for now. - Ranga On Wed, Oct 8, 2014 at 9:18 PM, Ranga sra...@gmail.com wrote: This is a bit strange. When I print the schema for the RDD, it reflects the correct data type for each column. But doing any kind of mathematical calculation seems to result in ClassCastException. Here

Spark-SQL: SchemaRDD - ClassCastException

2014-10-08 Thread Ranga
approaches that I should be looking at? Thanks for your help. - Ranga

Re: Spark-SQL: SchemaRDD - ClassCastException

2014-10-08 Thread Ranga
) // This query throws the exception when I collect the results I tried adding the cast to the aggRdd query above and that didn't help. - Ranga On Wed, Oct 8, 2014 at 3:52 PM, Michael Armbrust mich...@databricks.com wrote: Using SUM on a string should automatically cast the column. Also you

Re: Spark-SQL: SchemaRDD - ClassCastException

2014-10-08 Thread Ranga
:11 PM, Michael Armbrust mich...@databricks.com wrote: Which version of Spark are you running? On Wed, Oct 8, 2014 at 4:18 PM, Ranga sra...@gmail.com wrote: Thanks Michael. Should the cast be done in the source RDD or while doing the SUM? To give a better picture here is the code sequence

Re: Spark-SQL: SchemaRDD - ClassCastException

2014-10-08 Thread Ranga
as int) ... from table Any other pointers? Thanks for the help. - Ranga On Wed, Oct 8, 2014 at 5:20 PM, Ranga sra...@gmail.com wrote: Sorry. Its 1.1.0. After digging a bit more into this, it seems like the OpenCSV Deseralizer converts all the columns to a String type. This maybe throwing