It won't affect you if you're not actually running Hadoop. But it's mainly things like Snappy/LZO compression which are implemented as native libraries under the hood. Spark doesn't necessarily use these anyway; it's from the Hadoop libs.
On Tue, Aug 4, 2015 at 8:30 AM, Deepesh Maheshwari <deepesh.maheshwar...@gmail.com> wrote: > Can you elaborate about the things this native library covering. > One you mentioned accelerated compression. > > It would be very helpful if you can give any useful to link to read more > about it. > > On Tue, Aug 4, 2015 at 12:56 PM, Sean Owen <so...@cloudera.com> wrote: >> >> You can ignore it entirely. It just means you haven't installed and >> configured native libraries for things like accelerated compression, >> but it has no negative impact otherwise. >> >> On Tue, Aug 4, 2015 at 8:11 AM, Deepesh Maheshwari >> <deepesh.maheshwar...@gmail.com> wrote: >> > Hi, >> > >> > When i run the spark locally on windows it gives below hadoop library >> > error. >> > I am using below spark version. >> > >> > <dependency> >> > <groupId>org.apache.spark</groupId> >> > <artifactId>spark-core_2.10</artifactId> >> > <version>1.4.1</version> >> > </dependency> >> > >> > >> > 2015-08-04 12:22:23,463 WARN >> > (org.apache.hadoop.util.NativeCodeLoader:62) - >> > Unable to load native-hadoop library for your platform... using >> > builtin-java >> > classes where applicable >> > >> > Tried to find it on internet but not able to find exact root cause. >> > Please let me know what is it, why it is giving warning and how can i >> > resolve it. >> > >> > Thanks, >> > Deepesh > > --------------------------------------------------------------------- To unsubscribe, e-mail: user-unsubscr...@spark.apache.org For additional commands, e-mail: user-h...@spark.apache.org