Re: Hadoop Java Versions

2011-07-13 Thread Eric Baldeschwieler
We could create an apache hadoop list of product selection discussions. I believe this list is intended to be focused on project governance and similar discussions. Maybe we should simply create a governance list and leave this one to be the free for all? On Jul 2, 2011, at 9:16 PM, Ian

Re: Hadoop Java Versions

2011-07-13 Thread Ted Dunning
On Wed, Jul 13, 2011 at 7:59 AM, Eric Baldeschwieler eri...@hortonworks.com wrote: We could create an apache hadoop list of product selection discussions. I believe this list is intended to be focused on project governance and similar discussions. Maybe we should simply create a governance

Re: Hadoop Java Versions

2011-07-02 Thread Ian Holsman
On Jul 2, 2011, at 7:38 AM, Ted Dunning wrote: On Fri, Jul 1, 2011 at 11:53 AM, Abhishek Mehta abhis...@tresata.comwrote: open and honest conversations on this thread/group, irrespective of whether one represents a company or apache, should be encouraged. Paradoxically, I partially

1Gb vs 10Gb eth (WAS: Re: Hadoop Java Versions)

2011-07-01 Thread Evert Lammerts
Keeping the amount of disks per node low and the amount of nodes high should keep the impact of dead nodes in control. It keeps the impact of dead nodes in control but I don't think thats long-term cost efficient. As prices of 10GbE go down, the keep the node small arguement seems less

Re: 1Gb vs 10Gb eth (WAS: Re: Hadoop Java Versions)

2011-07-01 Thread Ryan Rawson
What's the justification for a management interface? Doesn't that increase complexity? Also you still twice the ports? On Jun 30, 2011 11:54 PM, Evert Lammerts evert.lamme...@sara.nl wrote: Keeping the amount of disks per node low and the amount of nodes high should keep the impact of dead nodes

RE: 1Gb vs 10Gb eth (WAS: Re: Hadoop Java Versions)

2011-07-01 Thread Evert Lammerts
What's the justification for a management interface? Doesn't that increase complexity? Also you still twice the ports? That's true. It's a tradition that I haven't questioned before. The reasoning, whether right or wrong, is that user jobs (our users are external) can get in the way of

Re: 1Gb vs 10Gb eth (WAS: Re: Hadoop Java Versions)

2011-07-01 Thread Steve Loughran
On 01/07/2011 07:41, Evert Lammerts wrote: Keeping the amount of disks per node low and the amount of nodes high should keep the impact of dead nodes in control. It keeps the impact of dead nodes in control but I don't think thats long-term cost efficient. As prices of 10GbE go down, the keep

Re: 1Gb vs 10Gb eth (WAS: Re: Hadoop Java Versions)

2011-07-01 Thread Steve Loughran
On 01/07/2011 08:16, Ryan Rawson wrote: What's the justification for a management interface? Doesn't that increase complexity? Also you still twice the ports? ILO reduces ops complexity. you can push things like BIOS updates out, boot machines into known states, instead of the slowly

Re: Hadoop Java Versions

2011-07-01 Thread Steve Loughran
On 01/07/2011 01:16, Ted Dunning wrote: You have to consider the long-term reliability as well. Losing an entire set of 10 or 12 disks at once makes the overall reliability of a large cluster very suspect. This is because it becomes entirely too likely that two additional drives will fail

Re: 1Gb vs 10Gb eth (WAS: Re: Hadoop Java Versions)

2011-07-01 Thread Ted Dunning
You are twice the ports, but if you know that the management port is not used for serious data, then you can put a SOHO grade switch on those ports at negligible cost. There is a serious conflict of goals here if you have software that can make serious use more than one NIC. On the one hand, it

Re: Hadoop Java Versions

2011-07-01 Thread Scott Carey
Although this thread is wandering a bit, I disagree strongly that it is inappropriate to discuss other vendor specific features (or competing compute platform features) on general@. The topic has become the factors that influence hardware purchase choices, and one of those is how the system deals

Re: Hadoop Java Versions

2011-07-01 Thread Ted Dunning
On Fri, Jul 1, 2011 at 9:12 AM, Steve Loughran ste...@apache.org wrote: On 01/07/2011 01:16, Ted Dunning wrote: You have to consider the long-term reliability as well. Losing an entire set of 10 or 12 disks at once makes the overall reliability of a large cluster very suspect. This is

Re: Hadoop Java Versions

2011-07-01 Thread Abhishek Mehta
i definitely agree with scott. as a user of the hadoop open source stack for building our banking focused big data analytics applications, i speak on behalf of our clients and the emerging hadoop eco-system that open and honest conversations on this thread/group, irrespective of whether one

Re: Hadoop Java Versions

2011-07-01 Thread Ted Dunning
On Fri, Jul 1, 2011 at 11:53 AM, Abhishek Mehta abhis...@tresata.comwrote: open and honest conversations on this thread/group, irrespective of whether one represents a company or apache, should be encouraged. Paradoxically, I partially agree with Ian on this. On the one hand, it is

RE: Hadoop Java Versions

2011-06-30 Thread Evert Lammerts
You can get 12-24 TB in a server today, which means the loss of a server generates a lot of traffic -which argues for 10 Gbe. But -big increase in switch cost, especially if you (CoI warning) go with Cisco -there have been problems with things like BIOS PXE and lights out management

RE: Hadoop Java Versions

2011-06-30 Thread Evert Lammerts
11:38 PM To: Evert Lammerts Subject: Fwd: Hadoop Java Versions what are the other switch options (other than cisco that is?)? cheers Abhishek Mehta (e) abhis...@tresata.commailto:abhis...@tresata.com (v) 980.355.9855 Begin forwarded message: From: Evert Lammerts evert.lamme

Re: Hadoop Java Versions

2011-06-30 Thread Aaron Eng
Keeping the amount of disks per node low and the amount of nodes high should keep the impact of dead nodes in control. It keeps the impact of dead nodes in control but I don't think thats long-term cost efficient. As prices of 10GbE go down, the keep the node small arguement seems less fitting.

Re: Hadoop Java Versions

2011-06-30 Thread Ted Dunning
You have to consider the long-term reliability as well. Losing an entire set of 10 or 12 disks at once makes the overall reliability of a large cluster very suspect. This is because it becomes entirely too likely that two additional drives will fail before the data on the off-line node can be

Re: Hadoop Java Versions

2011-06-30 Thread Todd Lipcon
On Thu, Jun 30, 2011 at 5:16 PM, Ted Dunning tdunn...@maprtech.com wrote: You have to consider the long-term reliability as well. Losing an entire set of 10 or 12 disks at once makes the overall reliability of a large cluster very suspect. This is because it becomes entirely too likely

Re: Hadoop Java Versions

2011-06-30 Thread Ted Dunning
Good point Todd. I was speaking from the experience of people I know who are using 0.20.x On Thu, Jun 30, 2011 at 5:24 PM, Todd Lipcon t...@cloudera.com wrote: On Thu, Jun 30, 2011 at 5:16 PM, Ted Dunning tdunn...@maprtech.com wrote: You have to consider the long-term reliability as well.

Re: Hadoop Java Versions

2011-06-30 Thread M. C. Srivas
On Thu, Jun 30, 2011 at 5:24 PM, Todd Lipcon t...@cloudera.com wrote: I'd advise you to look at stock hadoop again. This used to be true, but was fixed a long while back by HDFS-457 and several followup JIRAs. If MapR does something fancier, I'm sure we'd be interested to hear about it so

Re: Hadoop Java Versions

2011-06-30 Thread Ian Holsman
On Jul 1, 2011, at 2:08 PM, M. C. Srivas wrote: On Thu, Jun 30, 2011 at 5:24 PM, Todd Lipcon t...@cloudera.com wrote: I'd advise you to look at stock hadoop again. This used to be true, but was fixed a long while back by HDFS-457 and several followup JIRAs. If MapR does something

Re: Hadoop Java Versions

2011-06-30 Thread M. C. Srivas
No worries. I read Todd's post as asking for elaboration ... sometimes knowing what another similar system does helps in improving your own. On Thu, Jun 30, 2011 at 9:47 PM, Ian Holsman had...@holsman.net wrote: On Jul 1, 2011, at 2:08 PM, M. C. Srivas wrote: On Thu, Jun 30, 2011 at 5:24

Re: Hadoop Java Versions

2011-06-30 Thread Ted Dunning
Ian, Good point. Srivas was responding to Todd's question, but there might be better fora as you suggest. We have a good one for specific questions about MapR at http://answers.mapr.com That doesn't, however, really provide a useful forum for questions like Todd's which really spans both

Re: Hadoop Java Versions

2011-06-28 Thread Steve Loughran
On 28/06/11 04:49, Segel, Mike wrote: Hmmm. I could have sworn there was a background balancing bandwidth limiter. There is, for the rebalancer, node outages are taken more seriously, though there have been problems in past 0.20.x where there was a risk of a cascade failure on a big

Re: Hadoop Java Versions

2011-06-28 Thread Michel Segel
You're preaching to the choir. :-) With Sandybridge, you're going to start seeing 10 GBe on the motherboard. We built our clusters using 1U boxes where you're stuck w 4 3.5 drives. With larger chassis, You can fit an additional controller card and more drives. More drives reduces the bottleneck

Re: Hadoop Java Versions

2011-06-28 Thread Arun C Murthy
We at Yahoo are about to deploy code to ensure a disk failure on a datanode is just that - a disk failure. Not a node failure. This really helps avoid replication storms. It's in the 0.20.204 branch for the curious. Arun Sent from my iPhone On Jun 28, 2011, at 3:01 AM, Steve Loughran

Re: Hadoop Java Versions

2011-06-27 Thread Steve Loughran
On 26/06/11 20:23, Scott Carey wrote: On 6/23/11 5:49 AM, Steve Loughranste...@apache.org wrote: what's your HW setup? #cores/server, #servers, underlying OS? CentOS 5.6. 4 cores / 8 threads a server (Nehalem generation Intel processor). that should be enough to find problems. I've

Re: Hadoop Java Versions

2011-06-27 Thread Ryan Rawson
On the subject of gige vs 10-gige, I think that we will very shortly be seeing interest in 10gig, since gige is only 120MB/sec - 1 hard drive of streaming data. Nodes with 4+ disks are throttled by the network. On a small cluster (20 nodes), the replication traffic can choke a cluster to death.

Re: Hadoop Java Versions

2011-06-27 Thread Ted Dunning
Come to Srivas talk at the Summit. On Mon, Jun 27, 2011 at 5:10 PM, Ryan Rawson ryano...@gmail.com wrote: On the subject of gige vs 10-gige, I think that we will very shortly be seeing interest in 10gig, since gige is only 120MB/sec - 1 hard drive of streaming data. Nodes with 4+ disks are

Re: Hadoop Java Versions

2011-06-27 Thread Segel, Mike
That doesn't seem right. In one of our test clusters (19 data nodes) we found that under heavy loads we were disk I/O bound and not network bound. Of course YMMV depending on your ToR switch. If we had more than 4 disks per node, we would probably see the network being the bottleneck. What did

Re: Hadoop Java Versions

2011-06-27 Thread Ryan Rawson
There are no bandwidth limitations in 0.20.x. None that I saw at least. It was basically bandwidth-management-by-pwm. You could adjust the frequency of how many files-per-node were copied. In my case, the load was HBase real time serving, so it was servicing more smaller random reads, not a

Re: Hadoop Java Versions

2011-06-27 Thread Scott Carey
For cost reasons, we just bonded two 1G network ports together. A single stream won't go past 1Gbps, but concurrent ones do -- this is with the Linux built-in bonding. The network is only stressed during 'sort-like' jobs or big replication events. We also removed some disk bottlenecks by tuning

Re: Hadoop Java Versions

2011-06-27 Thread Segel, Mike
Hmmm. I could have sworn there was a background balancing bandwidth limiter. Haven't tested random reads... The last test we did ended up hitting the cache, but we didn't push it hard enough to hit network bandwidth limitations... Not to say they don't exist. Like I said in the other post, if

Re: Hadoop Java Versions

2011-06-26 Thread Scott Carey
On 6/23/11 5:49 AM, Steve Loughran ste...@apache.org wrote: On 22/06/2011 21:27, Scott Carey wrote: Problems have been reported with Hadoop, the 64-bit JVM and Compressed Object References (the -XX:+UseCompressedOops option), so use of that option is discouraged. I think the above is

Re: Hadoop Java Versions

2011-06-23 Thread Steve Loughran
On 22/06/2011 21:27, Scott Carey wrote: Problems have been reported with Hadoop, the 64-bit JVM and Compressed Object References (the -XX:+UseCompressedOops option), so use of that option is discouraged. I think the above is dated. It also lacks critical information. What JVM and OS version

Re: Hadoop Java Versions

2011-06-22 Thread Scott Carey
Problems have been reported with Hadoop, the 64-bit JVM and Compressed Object References (the -XX:+UseCompressedOops option), so use of that option is discouraged. I think the above is dated. It also lacks critical information. What JVM and OS version was the problem seen? CompressedOops had

Re: Hadoop Java Versions

2011-06-22 Thread Allen Wittenauer
On Jun 22, 2011, at 1:27 PM, Scott Carey wrote: Problems have been reported with Hadoop, the 64-bit JVM and Compressed Object References (the -XX:+UseCompressedOops option), so use of that option is discouraged. I think the above is dated. It also lacks critical information. What JVM and

Re: Hadoop Java Versions

2011-06-22 Thread Scott Carey
On 6/22/11 1:49 PM, Allen Wittenauer a...@apache.org wrote: On Jun 22, 2011, at 1:27 PM, Scott Carey wrote: Problems have been reported with Hadoop, the 64-bit JVM and Compressed Object References (the -XX:+UseCompressedOops option), so use of that option is discouraged. I think the