Maybe your Spark is based on scala 2.11, but you compile it for 2.10 or the other way around?
> On 6. Apr 2017, at 15:54, nayan sharma <nayansharm...@gmail.com> wrote: > > In addition I am using spark version 1.6.2 > Is there any chance of error coming because of Scala version or dependencies > are not matching.?I just guessed. > > Thanks, > Nayan > > >> On 06-Apr-2017, at 7:16 PM, nayan sharma <nayansharm...@gmail.com> wrote: >> >> Hi Jorn, >> Thanks for replying. >> >> jar -tf catalyst-data-prepration-assembly-1.0.jar | grep csv >> >> after doing this I have found a lot of classes under >> com/databricks/spark/csv/ >> >> do I need to check for any specific class ?? >> >> Regards, >> Nayan >>> On 06-Apr-2017, at 6:42 PM, Jörn Franke <jornfra...@gmail.com> wrote: >>> >>> Is the library in your assembly jar? >>> >>>> On 6. Apr 2017, at 15:06, nayan sharma <nayansharm...@gmail.com> wrote: >>>> >>>> Hi All, >>>> I am getting error while loading CSV file. >>>> >>>> val >>>> datacsv=sqlContext.read.format("com.databricks.spark.csv").option("header", >>>> "true").load("timeline.csv") >>>> java.lang.NoSuchMethodError: >>>> org.apache.commons.csv.CSVFormat.withQuote(Ljava/lang/Character;)Lorg/apache/commons/csv/CSVFormat; >>>> >>>> >>>> I have added the dependencies in sbt file >>>> // Spark Additional Library - CSV Read as DF >>>> libraryDependencies += "com.databricks" %% "spark-csv" % “1.5.0" >>>> and starting the spark-shell with command >>>> >>>> spark-shell --master yarn-client --jars >>>> /opt/packages/xxxx-data-prepration/target/scala-2.10/xxxx-data-prepration-assembly-1.0.jar >>>> --name nayan >>>> >>>> >>>> >>>> Thanks for any help!! >>>> >>>> >>>> Thanks, >>>> Nayan >> >