Return-Path: X-Original-To: apmail-hadoop-common-dev-archive@www.apache.org Delivered-To: apmail-hadoop-common-dev-archive@www.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id AE88A4B43 for ; Thu, 7 Jul 2011 18:04:07 +0000 (UTC) Received: (qmail 78557 invoked by uid 500); 7 Jul 2011 18:04:06 -0000 Delivered-To: apmail-hadoop-common-dev-archive@hadoop.apache.org Received: (qmail 78393 invoked by uid 500); 7 Jul 2011 18:04:05 -0000 Mailing-List: contact common-dev-help@hadoop.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: common-dev@hadoop.apache.org Delivered-To: mailing list common-dev@hadoop.apache.org Received: (qmail 78385 invoked by uid 99); 7 Jul 2011 18:04:04 -0000 Received: from athena.apache.org (HELO athena.apache.org) (140.211.11.136) by apache.org (qpsmtpd/0.29) with ESMTP; Thu, 07 Jul 2011 18:04:04 +0000 X-ASF-Spam-Status: No, hits=1.5 required=5.0 tests=HTML_MESSAGE,RCVD_IN_DNSWL_LOW,SPF_PASS X-Spam-Check-By: apache.org Received-SPF: pass (athena.apache.org: domain of wt@penguintechs.org designates 209.85.220.176 as permitted sender) Received: from [209.85.220.176] (HELO mail-vx0-f176.google.com) (209.85.220.176) by apache.org (qpsmtpd/0.29) with ESMTP; Thu, 07 Jul 2011 18:03:57 +0000 Received: by vxd2 with SMTP id 2so1402796vxd.35 for ; Thu, 07 Jul 2011 11:03:36 -0700 (PDT) Received: by 10.52.74.98 with SMTP id s2mr1419176vdv.277.1310061816054; Thu, 07 Jul 2011 11:03:36 -0700 (PDT) MIME-Version: 1.0 Received: by 10.52.187.165 with HTTP; Thu, 7 Jul 2011 11:03:16 -0700 (PDT) X-Originating-IP: [64.71.7.198] In-Reply-To: References: <31981952.post@talk.nabble.com> <4E119987.2050501@apache.org> <4E12DC2B.9040903@apache.org> From: Warren Turkal Date: Thu, 7 Jul 2011 11:03:16 -0700 Message-ID: Subject: Re: Hadoop Master and Slave Discovery To: common-dev@hadoop.apache.org Content-Type: multipart/alternative; boundary=bcaec50166972e022804a77e893d --bcaec50166972e022804a77e893d Content-Type: text/plain; charset=ISO-8859-1 Stable infrastructures require deterministic behavior to be understood. I believe that mdns limits the determinism of a system by requiring that I accept that a machine will be picking a random address. When I setup my DCs I want machines to have a single constant ip address so that I don't have to do a bunch of work to figure out which machine I am trying to talk to when things go wrong. As such, I find it hard to believe that mdns belongs in large DCs. wt On Thu, Jul 7, 2011 at 10:17 AM, Eric Yang wrote: > Internet Assigned Number Authority has allocated 169.254.1.0 to > 169.254.254.255 for the propose of communicate between nodes. This is > 65024 IP address designed for local area network only. They are not > allow to be routed. Zeroconf is randomly selecting one address out of > the 65024 available address, and broadcasts ARP message. If no one is > using this address, then the machine will use the selected ip address > and communicate with zeroconf service for name resolution. If the > address is already in use, the system restart from scratch to pick > another address. Hence, the actual limit is not bound to 1000, but > 65024. In real life, it is unlikely to use all 65024 for name > resolution due to chance of loss packet on modern ethernet (10^-12) or > delay from repetitive selection of the the same ip address from > different hosts. It can easily push the limit to 10,000-20,000 nodes > without losing reliability in server farm settings. > > It would be nice to support both dynamic discovery of master and > slaves, and preserve the exist configuration style management for EC2 > like deployment. This is one innovation worth having. > > regards, > Eric > > On Wed, Jul 6, 2011 at 5:49 PM, Allen Wittenauer wrote: > > > > On Jul 6, 2011, at 5:05 PM, Eric Yang wrote: > > > >> Did you know that almost all linux desktop system comes with avahi > >> pre-installed and turn on by default? > > > > ... which is why most admins turn those services off by default. > :) > > > >> What is more interesting is > >> that there are thousands of those machines broadcast in large > >> cooperation without anyone noticing them? > > > > That's because many network teams turn off multicast past the > subnet boundary and many corporate desktops are in class C subnets. This > automatically limits the host count down to 200-ish per network. Usually > just the unicast traffic is bad enough. Throwing multicast into the mix > just makes it worse. > > > >> I have recently built a > >> multicast dns browser and look into the number of machines running in > >> a large company environment. The number of desktop, laptop and > >> printer machines running multicast dns is far exceeding 1000 machines > >> in the local subnet. > > > > From my understanding of Y!'s network, the few /22's they have > (which would get you 1022 potential hosts on a subnet) have multicast > traffic dropped at the router and switch levels. Additionally, DNS-SD (the > service discovery portion of mDNS) offers unicast support as well. So there > is a very good chance that the traffic you are seeing is from unicast, not > multicast. > > > > The 1000 number, BTW, comes from Apple. I'm sure they'd be > interested in your findings given their role in ZC. > > > > BTW, I'd much rather hear that you set up a /22 with many many > machines running VMs trying to actually use mDNS for something useful. A > service browser really isn't that interesting. > > > >> They are all happily working fine without causing any issues. > > > > ... that you know of. Again, I'm 99% certain that Y! is dropping > multicast packets into the bit bucket at the switch boundaries. [I remember > having this conversation with them when we setup the new data centers.] > > > >> Printer works fine, > > > > Most admins turn SLP and other broadcast services on printers off. > For large networks, one usually sees print services enabled via AD or > master print servers broadcasting the information on the local subnet. This > allows a central point of control rather than randomness. Snow Leopard (I > don't think Leopard did this) actually tells you where the printer is coming > from now, so that's handy to see if they are ZC or AD or whatever. > > > >> itune sharing from someone > >> else works fine. > > > > iTunes specifically limits its reach so that it can't extend > beyond the local subnet and definitely does unicast in addition to ZC, so > that doesn't really say much of anything, other than potentially > invalidating your results. > > > >> For some reason, things tend to work better on my > >> side of universe. :) > > > > I'm sure it does, but not for the reasons you think they do. > > > >> Allen, if you want to get stuck on stone age > >> tools, I won't stop you. > >> > > > > Multicast has a time and place (mainly for small, non-busy > networks). Using it without understanding the network impact is never a > good idea. > > > > FWIW, I've seen multicast traffic bring down an entire campus of > tens of thousands of machines due to routers and switches having bugs where > they didn't subtract from the packet's TTL. I'm not the only one with these > types of experiences. Anything multicast is going to have a very large > uphill battle for adoption because of these widespread problems. Many > network vendors really don't get this one right, for some reason. > --bcaec50166972e022804a77e893d--