Amandeep created ZEPPELIN-5914:
----------------------------------

             Summary: spark.sparkContext.textFile(...) is not working 
                 Key: ZEPPELIN-5914
                 URL: https://issues.apache.org/jira/browse/ZEPPELIN-5914
             Project: Zeppelin
          Issue Type: Bug
            Reporter: Amandeep


When using spark-3.1.2 with hadoop 3.2.0 sparkContext.textFile(..) throws a NPE:

 
java.lang.NullPointerException
  at 
org.apache.hadoop.mapred.TextInputFormat.isSplitable(TextInputFormat.java:49)
  at 
org.apache.hadoop.mapred.FileInputFormat.getSplits(FileInputFormat.java:363)
  at org.apache.spark.rdd.HadoopRDD.getPartitions(HadoopRDD.scala:205)
  at org.apache.spark.rdd.RDD.$anonfun$partitions$2(RDD.scala:300)
  at scala.Option.getOrElse(Option.scala:189)
  at org.apache.spark.rdd.RDD.partitions(RDD.scala:296)
  at 
org.apache.spark.rdd.MapPartitionsRDD.getPartitions(MapPartitionsRDD.scala:49)
  at org.apache.spark.rdd.RDD.$anonfun$partitions$2(RDD.scala:300)
  at scala.Option.getOrElse(Option.scala:189)
  at org.apache.spark.rdd.RDD.partitions(RDD.scala:296)
  at org.apache.spark.SparkContext.runJob(SparkContext.scala:2261)
  at org.apache.spark.rdd.RDD.count(RDD.scala:1253)
  ... 114 elided
 

Built zepplein from master branch via
mvn clean package -DskipTests -Dcheckstyle.skip -Pbuild-distr -Pspark-3.1 
-Phadoop3 -Pspark-scala-2.12 -Pscala-2.12 -Pinclude-hadoop -Pweb-dist
 

This works on spark-shell but not on zeppelins spark interpreter. 

 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

Reply via email to