It looks like you have an issue with your classpath, I think it is because you add a jar containing Spark twice: first, you have a dependency on Spark somewhere in your build tool (this allows you to compile and run your application), second you re-add Spark here
> sc.addJar("/home/hadoop/spark-assembly-1.5.2-hadoop2.6.0.jar")//It doesn't work.!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!????? I recommend you remove that line and see if everything works. If you have that line because you need hadoop 2.6, I recommend you build spark against that version and publish locally with maven