Hi Imran, Thanks for your reply. I have double-checked the code I ran to generate an nxn matrix and nx1 vector for n = 2^27. There was unfortunately a bug in it, where instead of having typed 134,217,728 for n = 2^27, I included a third '7' by mistake, making the size 10x larger.
However, even after having corrected this, my question about broadcasting is still whether or not a variable >= 2G in size may be transferred? In this case, for n >= 2^28, the broadcast variable crashes, and an array of size MAX_INT cannot be broadcast. Looking at Chowdhury's "Performance and Scalability of Broadcast in Spark" technical report, I realize that the results are reported only for broadcast variables up to 1 GB in physical size. I was hoping, however, that an Array of size MAX_INT would be transferrable via a broadcast (since the previous PR I mentioned seems to have added support for > 2GB variables) such that the matrix-vector multiplication would scale to MAX_INT x MAX_INT matrices with a broadcast variable. Would you or anyone on the dev list be able to comment on whether this is possible? Since the (corrected) overflow I'm seeing is for > 2^31 physical bytes being transferred, I am guessing that there is still a physical limitation on how many bytes may be sent via broadcasting, at least for a primitive Array[Double]? Thanks, Mike 19176&INFO&IndexedRowMatrix&Broadcasting vecArray with size 268435456& 19177&INFO&MemoryStore&ensureFreeSpace(-2147483592) called with curMem=6888, maxMem=92610625536& 19177&INFO&MemoryStore&Block broadcast_2 stored as values in memory (estimated size -2147483592.0 B, free 88.3 GB)& Exception in thread "main" java.lang.IllegalArgumentException: requirement failed: sizeInBytes was negative: -2147483592 On 7/28/15, Imran Rashid <iras...@cloudera.com> wrote: > Hi Mike, > > are you sure there the size isn't off 2x somehow? I just tried to > reproduce with a simple test in BlockManagerSuite: > > test("large block") { > store = makeBlockManager(4e9.toLong) > val arr = new Array[Double](1 << 28) > println(arr.size) > val blockId = BlockId("rdd_3_10") > val result = store.putIterator(blockId, Iterator(arr), > StorageLevel.MEMORY_AND_DISK) > result.foreach{println} > } > > it fails at 1 << 28 with nearly the same message, but its fine for (1 << > 28) - 1 with a reported block size of 2147483680. Not exactly the same as > what you did, but I expect it to be close enough to exhibit the same error. > > > On Tue, Jul 28, 2015 at 12:37 PM, Mike Hynes <91m...@gmail.com> wrote: >> >> Hello Devs, >> >> I am investigating how matrix vector multiplication can scale for an >> IndexedRowMatrix in mllib.linalg.distributed. >> >> Currently, I am broadcasting the vector to be multiplied on the right. >> The IndexedRowMatrix is stored across a cluster with up to 16 nodes, >> each with >200 GB of memory. The spark driver is on an identical node, >> having more than 200 Gb of memory. >> >> In scaling n, the size of the vector to be broadcast, I find that the >> maximum size of n that I can use is 2^26. For 2^27, the broadcast will >> fail. The array being broadcast is of type Array[Double], so the >> contents have size 2^30 bytes, which is approximately 1 (metric) GB. >> >> I have read in PR [SPARK-3721] [PySpark] "broadcast objects larger >> than 2G" that this should be supported (I assume this works for scala, >> as well?). However, when I increase n to 2^27 or above, the program >> invariably crashes at the broadcast. >> >> The problem stems from the size of the result block to be sent in >> BlockInfo.scala; the size is reportedly negative. An example error log >> is shown below. >> >> If anyone has more experience or knowledge of why this broadcast is >> failing, I'd appreciate the input. >> -- >> Thanks, >> Mike >> >> 55584:INFO:MemoryStore:ensureFreeSpace(-2147480008) called with >> curMem=0, maxMem=92610625536: >> 55584:INFO:MemoryStore:Block broadcast-2 stored as values in memory >> (estimated size -2147480008.0 B, free 88.3 GB): >> Exception in thread "main" java.lang.IllegalArgumentException: >> requirement failed: sizeInBytes was negative: -2147480008 >> at scala.Predef$.require(Predef.scala:233) >> at > org.apache.spark.storage.BlockInfo.markReady(BlockInfo.scala:55) >> at > org.apache.spark.storage.BlockManager.doPut(BlockManager.scala:815) >> at > org.apache.spark.storage.BlockManager.putIterator(BlockManager.scala:638) >> at > org.apache.spark.storage.BlockManager.putSingle(BlockManager.scala:996) >> at > org.apache.spark.broadcast.TorrentBroadcast.writeBlocks(TorrentBroadcast.scala:99) >> at > org.apache.spark.broadcast.TorrentBroadcast.<init>(TorrentBroadcast.scala:85) >> at > org.apache.spark.broadcast.TorrentBroadcastFactory.newBroadcast(TorrentBroadcastFactory.scala:34) >> at > org.apache.spark.broadcast.BroadcastManager.newBroadcast(BroadcastManager.scala:63) >> at > org.apache.spark.SparkContext.broadcast(SparkContext.scala:1297) >> at > org.apache.spark.mllib.linalg.distributed.IndexedRowMatrix.multiply(IndexedRowMatrix.scala:184) >> at himrod.linalg.KrylovTests$.main(KrylovTests.scala:172) >> at himrod.linalg.KrylovTests.main(KrylovTests.scala) >> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >> at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) >> at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >> at java.lang.reflect.Method.invoke(Method.java:606) >> at > org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$runMain(SparkSubmit.scala:666) >> at > org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:178) >> at > org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:203) >> at > org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:118) >> at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) >> >> --------------------------------------------------------------------- >> To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org >> For additional commands, e-mail: dev-h...@spark.apache.org >> > -- Thanks, Mike --------------------------------------------------------------------- To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org For additional commands, e-mail: dev-h...@spark.apache.org