Thanks for the follow-up, Dale.

bq. hdp 2.3.1

Minor correction: should be hdp 2.1.3

Cheers

On Sat, Mar 28, 2015 at 2:28 AM, Johnson, Dale <daljohn...@ebay.com> wrote:

>  Actually I did figure this out eventually.
>
>  I’m running on a Hortonworks cluster hdp 2.3.1 (hadoop 2.4.1).  Spark
> bundles the org/apache/hadoop/hdfs/… classes along with the spark-assembly
> jar.  This turns out to introduce a small incompatibility with hdp 2.3.1.
> I carved these classes out of the jar, and put a distro-provided jar into
> the class path for the hdfs classes, and this fixed the problem.
>
>  Ideally there would be an exclusion in the pom to deal with this.
>
>  Dale.
>
>   From: Zhan Zhang <zzh...@hortonworks.com>
> Date: Friday, March 27, 2015 at 4:28 PM
> To: "Johnson, Dale" <daljohn...@ebay.com>
> Cc: Ted Yu <yuzhih...@gmail.com>, user <user@spark.apache.org>
>
> Subject: Re: Can't access file in spark, but can in hadoop
>
>   Probably guava version conflicts issue. What spark version did you use,
> and which hadoop version it compile against?
>
>  Thanks.
>
>  Zhan Zhang
>
>  On Mar 27, 2015, at 12:13 PM, Johnson, Dale <daljohn...@ebay.com> wrote:
>
>  Yes, I could recompile the hdfs client with more logging, but I don’t
> have the day or two to spare right this week.
>
>  One more thing about this, the cluster is Horton Works 2.1.3 [.0]
>
>  They seem to have a claim of supporting spark on Horton Works 2.2
>
>  Dale.
>
>   From: Ted Yu <yuzhih...@gmail.com>
> Date: Thursday, March 26, 2015 at 4:54 PM
> To: "Johnson, Dale" <daljohn...@ebay.com>
> Cc: user <user@spark.apache.org>
> Subject: Re: Can't access file in spark, but can in hadoop
>
>   Looks like the following assertion failed:
>       Preconditions.checkState(storageIDsCount == locs.size());
>
>  locs is List<DatanodeInfoProto>
> Can you enhance the assertion to log more information ?
>
>  Cheers
>
> On Thu, Mar 26, 2015 at 3:06 PM, Dale Johnson <daljohn...@ebay.com> wrote:
>
>> There seems to be a special kind of "corrupted according to Spark" state
>> of
>> file in HDFS.  I have isolated a set of files (maybe 1% of all files I
>> need
>> to work with) which are producing the following stack dump when I try to
>> sc.textFile() open them.  When I try to open directories, most large
>> directories contain at least one file of this type.  Curiously, the
>> following two lines fail inside of a Spark job, but not inside of a Scoobi
>> job:
>>
>>     val conf = new org.apache.hadoop.conf.Configuration
>>     val fs = org.apache.hadoop.fs.FileSystem.get(conf)
>>
>> The stack trace follows:
>>
>> 15/03/26 14:22:43 INFO yarn.ApplicationMaster: Final app status: FAILED,
>> exitCode: 15, (reason: User class threw exception: null)
>> Exception in thread "Driver" java.lang.IllegalStateException
>>         at
>>
>> org.spark-project.guava.common.base.Preconditions.checkState(Preconditions.java:133)
>>         at
>> org.apache.hadoop.hdfs.protocolPB.PBHelper.convert(PBHelper.java:673)
>>         at
>>
>> org.apache.hadoop.hdfs.protocolPB.PBHelper.convertLocatedBlock(PBHelper.java:1100)
>>         at
>> org.apache.hadoop.hdfs.protocolPB.PBHelper.convert(PBHelper.java:1118)
>>         at
>> org.apache.hadoop.hdfs.protocolPB.PBHelper.convert(PBHelper.java:1251)
>>         at
>> org.apache.hadoop.hdfs.protocolPB.PBHelper.convert(PBHelper.java:1354)
>>         at
>> org.apache.hadoop.hdfs.protocolPB.PBHelper.convert(PBHelper.java:1363)
>>         at
>>
>> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.getListing(ClientNamenodeProtocolTranslatorPB.java:518)
>>         at org.apache.hadoop.hdfs.DFSClient.listPaths(DFSClient.java:1743)
>>         at
>>
>> org.apache.hadoop.hdfs.DistributedFileSystem$15.<init>(DistributedFileSystem.java:738)
>>         at
>>
>> org.apache.hadoop.hdfs.DistributedFileSystem.listLocatedStatus(DistributedFileSystem.java:727)
>>         at
>> org.apache.hadoop.fs.FileSystem.listLocatedStatus(FileSystem.java:1662)
>>         at org.apache.hadoop.fs.FileSystem$5.<init>(FileSystem.java:1724)
>>         at org.apache.hadoop.fs.FileSystem.listFiles(FileSystem.java:1721)
>>         at
>>
>> com.ebay.ss.niffler.miner.speller.SpellQueryLaunch$$anonfun$main$2.apply(SpellQuery.scala:1125)
>>         at
>>
>> com.ebay.ss.niffler.miner.speller.SpellQueryLaunch$$anonfun$main$2.apply(SpellQuery.scala:1123)
>>         at
>>
>> scala.collection.IndexedSeqOptimized$class.foreach(IndexedSeqOptimized.scala:33)
>>         at
>> scala.collection.mutable.ArrayOps$ofRef.foreach(ArrayOps.scala:108)
>>         at
>>
>> com.ebay.ss.niffler.miner.speller.SpellQueryLaunch$.main(SpellQuery.scala:1123)
>>         at
>> com.ebay.ss.niffler.miner.speller.SpellQueryLaunch.main(SpellQuery.scala)
>>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>         at
>>
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>>         at
>>
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>         at java.lang.reflect.Method.invoke(Method.java:606)
>>         at
>>
>> org.apache.spark.deploy.yarn.ApplicationMaster$$anon$2.run(ApplicationMaster.scala:427)
>> 15/03/26 14:22:43 INFO yarn.ApplicationMaster: Invoking sc stop from
>> shutdown hook
>>
>> It appears to have found the three copies of the given HDFS block, but is
>> performing some sort of validation with them before giving them back to
>> spark to schedule the job.  But there is an assert failing.
>>
>> I've tried this with 1.2.0, 1.2.1 and 1.3.0, and I get the exact same
>> error,
>> but I've seen the line numbers change on the HDFS libraries, but not the
>> function names.  I've tried recompiling myself with different hadoop
>> versions, and it's the same.  We're running hadoop 2.4.1 on our cluster.
>>
>> A google search turns up absolutely nothing on this.
>>
>> Any insight at all would be appreciated.
>>
>> Dale Johnson
>> Applied Researcher
>> eBay.com
>>
>>
>>
>>
>> --
>> View this message in context:
>> http://apache-spark-user-list.1001560.n3.nabble.com/Can-t-access-file-in-spark-but-can-in-hadoop-tp22251.html
>> Sent from the Apache Spark User List mailing list archive at Nabble.com.
>>
>> ---------------------------------------------------------------------
>> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
>> For additional commands, e-mail: user-h...@spark.apache.org
>>
>>
>
>

Reply via email to