Maybe your Spark is based on scala 2.11, but you compile it for 2.10 or the 
other way around?

> On 6. Apr 2017, at 15:54, nayan sharma <nayansharm...@gmail.com> wrote:
> 
> In addition I am using spark version 1.6.2
> Is there any chance of error coming because of Scala version or dependencies 
> are not matching.?I just guessed.
> 
> Thanks,
> Nayan
> 
>  
>> On 06-Apr-2017, at 7:16 PM, nayan sharma <nayansharm...@gmail.com> wrote:
>> 
>> Hi Jorn,
>> Thanks for replying.
>> 
>> jar -tf catalyst-data-prepration-assembly-1.0.jar | grep csv
>> 
>> after doing this I have found a lot of classes under 
>> com/databricks/spark/csv/
>> 
>> do I need to check for any specific class ??
>> 
>> Regards,
>> Nayan
>>> On 06-Apr-2017, at 6:42 PM, Jörn Franke <jornfra...@gmail.com> wrote:
>>> 
>>> Is the library in your assembly jar?
>>> 
>>>> On 6. Apr 2017, at 15:06, nayan sharma <nayansharm...@gmail.com> wrote:
>>>> 
>>>> Hi All,
>>>> I am getting error while loading CSV file.
>>>> 
>>>> val 
>>>> datacsv=sqlContext.read.format("com.databricks.spark.csv").option("header",
>>>>  "true").load("timeline.csv")
>>>> java.lang.NoSuchMethodError: 
>>>> org.apache.commons.csv.CSVFormat.withQuote(Ljava/lang/Character;)Lorg/apache/commons/csv/CSVFormat;
>>>> 
>>>> 
>>>> I have added the dependencies in sbt file 
>>>> // Spark Additional Library - CSV Read as DF
>>>> libraryDependencies += "com.databricks" %% "spark-csv" % “1.5.0"
>>>> and starting the spark-shell with command
>>>> 
>>>> spark-shell --master yarn-client  --jars 
>>>> /opt/packages/xxxx-data-prepration/target/scala-2.10/xxxx-data-prepration-assembly-1.0.jar
>>>>  --name nayan 
>>>> 
>>>> 
>>>> 
>>>> Thanks for any help!!
>>>> 
>>>> 
>>>> Thanks,
>>>> Nayan
>> 
> 

Reply via email to