Adding in a few guys so they can chime in.
On Mon, Nov 4, 2013 at 4:33 PM, Reynold Xin <r...@apache.org> wrote: > I chatted with Matt Massie about this, and here are some options: > > 1. Use dependency injection in google-guice to make Akka use one version > of protobuf, and YARN use the other version. > > 2. Look into OSGi to accomplish the same goal. > > 3. Rewrite the messaging part of Spark to use a simple, custom RPC library > instead of Akka. We are really only using a very simple subset of Akka > features, and we can probably implement a simple RPC library tailored for > Spark quickly. We should only do this as the last resort. > > 4. Talk to Akka guys and hope they can make a maintenance release of Akka > that supports protobuf 2.5. > > > None of these are ideal, but we'd have to pick one. It would be great if > you have other suggestions. > > > On Sun, Nov 3, 2013 at 11:46 PM, Liu, Raymond <raymond....@intel.com>wrote: > >> Hi >> >> I am working on porting spark onto Hadoop 2.2.0, With some >> renaming and call into new YARN API works done. I can run up the spark >> master. While I encounter the issue that Executor Actor could not >> connecting to Driver actor. >> >> After some investigation, I found the root cause is that the >> akka-remote do not support protobuf 2.5.0 before 2.3. And hadoop move to >> protobuf 2.5.0 from 2.1-beta. >> >> The issue is that if I exclude the akka dependency from hadoop >> and force protobuf dependency to 2.4.1, the compile/packing will fail since >> hadoop common jar require a new interface from protobuf 2.5.0. >> >> So any suggestion on this? >> >> Best Regards, >> Raymond Liu >> > >