Follow Slashdot blog updates by subscribing to our blog RSS feed

 



Forgot your password?
typodupeerror
×
The Internet Hardware

How Many Google Machines, Really? 476

BoneThugND writes "I found this article on TNL.NET. It takes information from the S-1 Filing to reverse engineer how many machines Google has (hint: a lot more than 10,000). 'According to calculations by the IEE, in a paper about the Google cluster, a rack with 88 dual-CPU machines used to cost about $278,000. If you divide the $250 million figure from the S-1 filing by $278,000, you end up with a bit over 899 racks. Assuming that each rack holds 88 machines, you end up with 79,000 machines.'" An anonymous source claims over 100,000.
This discussion has been archived. No new comments can be posted.

How Many Google Machines, Really?

Comments Filter:
  • by Anonymous Coward on Sunday May 02, 2004 @01:17PM (#9034371)
    working at abovenet google has pulled there machines in and out of our data centers many a times. its incredible the way they have there shit is setup.

    they fit about 100 or so 1u's on each side of the rack, there double sided cabinets that look like refrigerators. there seperated in the center by noname brand switches and they have castor wheels on the bottoms of them. google can at the drop of a dime roll there machines out of a datacenter onto there 16 wheeler, move, unload and plug into a new data center in less than a days time.
  • by Anonymous Coward on Sunday May 02, 2004 @01:19PM (#9034389)
    In your standard 42U cabinet, you're talking a half-U per server. Umm.. not happening.

    IBM has a blade center that can hold 84 2-way blades in a 42U cabinet.
  • Re:$278k ?? (Score:4, Informative)

    by toddler99 ( 626625 ) on Sunday May 02, 2004 @01:20PM (#9034404)
    google doesn't buy pre-built machines they have been building costum machines from the very beginning. Although, with fab'n their own memory, i'm sure today they do a lot more. Google runs the cheapest most unreliable hardware you can find. It's in the software that they make up for the unreliable hardware. Though unreliable hardware is ok so long as you have staff to get the broken systems out and replaced with a new unreliable cheap ass system. When google started they used lego's to hold their costum built servers together
  • by PenguinOpus ( 556138 ) on Sunday May 02, 2004 @01:22PM (#9034413)
    Racksaver was selling dual-machine 1U racks for several years and I owned a few of them. Think deep, not tall. Racksaver seems to have renamed itself Verari and only has dual-Opteron in a 1U now. Most dense configs seem to be blade-based these days. Verari advertises 132 processors in a single rack, but I suspect they are not king in this area.

    If Google is innovating in this area, it could either be on price or in density.
  • Heat (Score:5, Informative)

    by gspr ( 602968 ) on Sunday May 02, 2004 @01:22PM (#9034415)
    A Pentium 4 dissipates around 85 W of heat. I don't know what the Xeon does, but let's be kind and say 50 W (wild guess). Using the article's "low end" estimate, that brings us to 4.7 MW!
    I hope they have good ventilation...
  • by SporkLand ( 225979 ) on Sunday May 02, 2004 @01:29PM (#9034469)
    When you can just open "Computer Architecture: A Quantitavie Approach, 3rd Edition" by Hennessy and Patterson to page 855 and find out that in summary:
    Google has 3 sites (two west coast, one east)
    Each site connected with 1 OC48
    Each OC48 hooks up to 2 Foundry BigIron 8000 ...
    80 Pc's per rack * 40 racks(at an example site)
    = 3200 PC's.
    A google site is not a homogenous set of PC's instead there are different types of PC's that are being upgraded on different cycles based on the price/performance ratio.

    If you want more info get the patterson hennessy book that I mentioned. Not the other version they sell. This one rocks way harder. You get to learn fun things like Tomosulo's algorithm.

    If I am violating any copy rights feel free to remove this post.
  • Re:Can you imagine (Score:1, Informative)

    by Anonymous Coward on Sunday May 02, 2004 @01:31PM (#9034475)
    Three things:

    First, funny moderations don't affect karma so making a joke like that isn't karma-whoring as the user is not gaining karma.

    Second, you can turn off the meta-moderation messages by simply deselecting the check box that says "willing to moderate" in your user preferences - which is what I am assuming you want based on your somewhat cryptic sentence in all caps.

    Third, decaf is your friend.

  • Re:What a waste (Score:5, Informative)

    by Waffle Iron ( 339739 ) on Sunday May 02, 2004 @01:31PM (#9034476)
    I'm sure a single IBM mainframe could do the same amount of work in half the amount of time and cost a fraction of what that Linux cluster cost.

    Mainframes are optimized for batch processing. Interactive queries do not take full advantage of their vaunted I/O capacity.

    Moreover, while a mainframe may be a good way to host a single copy of a database that must remain internally consistent, that's not the problem Google is solving. It's trivial for them to run their search service off of thousands of replicated copies of the Internet index. Even the largest mainframe's storage I/O would be orders of magnitude smaller than the massively parallel I/O operations done by these thousands of PCs. Google has no reason to funnel all of the independent search queries into a single machine, so they shouln't buy a system architecture designed to do that.

  • by nbahi15 ( 163501 ) on Sunday May 02, 2004 @01:34PM (#9034508) Homepage
    The CIO and Head Brainsurgeon (he really is a medical doctor) was at SVLUG last year he said there were about 11500 Linux boxes at Google.
  • Re:Heat (Score:5, Informative)

    by gammelby ( 457638 ) on Sunday May 02, 2004 @02:34PM (#9034813) Homepage
    I once attended a talk by google fellow Urs Hölzle on the google architecture, and he mentioned how they handle the cooling issue: They do not depend on each individual unit to be cooled separately - instead they have an enormous flow of air between the racks (sitting back to back), generated by some large fan in the roof.

    Ulrik

  • Re:At $699 per CPU (Score:3, Informative)

    by damiam ( 409504 ) on Sunday May 02, 2004 @02:53PM (#9034918)
    No. They run Linux, with their own proprietary software over it.
  • Re:Acquisition (Score:5, Informative)

    by Anonymous Coward on Sunday May 02, 2004 @03:03PM (#9034983)
    >>Disks are going to fail at a rate of several hundred or thousand PER DAY

    that's a little over the top big guy. i've worked at a 10,000 node corp doing desktop support. We lost ONE disk perhaps a week....if that much. We often went several weeks with no disks lost.

    even if you factor in multiple drives per server, say TWO (because they are servers not desktops)

    Interpolate for 100,000, that's a max of 20 disks per week...on the high end.

  • Re:$278k ?? (Score:2, Informative)

    by Anonymous Coward on Sunday May 02, 2004 @03:06PM (#9034999)
    Google Filesystem replicates same data on three nodes (by default, can be configured to more), so the probability of data loss is rather small. Source here [rochester.edu].
  • by gammelby ( 457638 ) on Sunday May 02, 2004 @03:12PM (#9035040) Homepage
    In the talk mentioned in a previous posting, mr. Hölzle also talked about disk failures: They have so many disks (obviously of low quality, according to you) and read so much data, that they cannot rely on standard CRC-32 checks. They use their own checksumming in a higher layer to circumvent the fact that CRC-32 gives false positive results in one out of some-large-number.

    Ulrik

  • Re:$278k ?? (Score:5, Informative)

    by Anonymous Coward on Sunday May 02, 2004 @03:15PM (#9035057)
    The high-end Sun machines are designed for high availability. Not only will a CPU failure not crash the machine, the CPUs are hot swappable so you can replace a failed CPU without so much as a reboot.

  • Re:$278k ?? (Score:5, Informative)

    by sql*kitten ( 1359 ) * on Sunday May 02, 2004 @03:16PM (#9035066)
    Any of those 64 CPUs fails, and your system will crash.

    Doesn't work like that, kid. A CPU on a high-end Sun fails, and the system will keep on running. You can swap the CPU out and replace it with a new one, the system will simply pick it up, assign threads to it, and keep on running. Had a couple of CPUs fail a little while ago... the first we users noticed of it was that the application slowed down slightly. Sysadmin just said yeah, I know, I'll replace 'em when the parts arrive this afternoon. Cool, we said. No data lost, no need to shut down or even restart our app. 'Course you gotta architect your app to deal with that - like don't have just one thread that does a crucial task, 'cos there's a chance that might be on the CPU that fails. But still, it's no big deal.
  • Re:$278k ?? (Score:5, Informative)

    by Anonymous Coward on Sunday May 02, 2004 @03:44PM (#9035241)
    Dude, big iron is not comparable in the slightest to that dinky little dual PPro Linux 'server' you keep in your closet. A CPU can fail, on a live running system, and the machine and Solaris or AIX won't even hiccup. Your application will notice, because suddenly a couple of its threads will quit, but that's ok, software like Oracle already knows how to deal with failed transations. And if you can schedule a CPU board removal/swap, then there won't be ANY problems at all, as the OS will migrate threads to other CPUs and allow the removal or hardware.

    And hey, if you want to mix and match CPU types (uSparc 2 and 3, etc), speeds, etc, no problem either. So if you wanna upgrade your server's CPUs, there will be zero downtime, you just do it a board at a time (board = 2 or 4 CPUs).
  • Re:$278k ?? (Score:2, Informative)

    by mikis ( 53466 ) on Sunday May 02, 2004 @04:03PM (#9035352) Homepage

    google doesn't buy pre-built machines

    Yes, they do. [rackable.com]

  • Server pricing (Score:5, Informative)

    by JWSmythe ( 446288 ) <jwsmythe@nospam.jwsmythe.com> on Sunday May 02, 2004 @04:15PM (#9035423) Homepage Journal
    His pricing in the summary may be a bit off.

    Every article I've read about Google's servers says they use "commodity" parts, which means they buy pretty much the same stuff we buy. They also indicate that they use as much memory as possible, and don't use hard drives, or use the drives as little as possible. From my interview with Google, they asked quite a few questions about RAID0, RAID1 (and combinations of those), I'd believe they stick in two drives to ensure data doesn't get lost due to power outages.

    We get good name brand parts wholesale, which I'd expect is what they do too. So, assuming 1u Asus, Tyan, or SuperMicro machines stuffed full of memory, with hard drives big enough to hold the OS plus an image of whatever they store in memory (ramdrives?), they'd require at most 3Gb (OS) + 4Gb (ramdrive backup). I don't recall seeing dual CPU's, but we'll go with that assumption.

    The nice base machine we had settled on for quite a while was the Asus 1400r, which consisted of dual 1.4Ghz PIII's, 2Gb RAM, and 20Gb and 200Gb hard drives. Our cost was roughly $1500. They'd lower the drive cost, but incrase the memory cost, so they'd probably cost about $1700, but I'm sure Google got better pricing, buying the quantity they were getting.

    The count of 88 machines per rack is a bit high. You get 80u's per standard rack, but you can't stuff it full of machines, unless you get very creative. I'd suspect they have 2 switches, and a few power management units per rack. The APC's we use take 8 machines per unit, and are 1u tall. There are other power management units, that don't take up rack space, which they may be using, but only the folks at Google really know.

    Assuming the maximum density, and equipment that was available as "commodity" equipment at the time, they'd have 2 Cisco 2948's and 78 servers per rack.

    $1700 * 78 (servers)
    +
    $3000 * 2 (switches)
    +
    $1000 (power management)
    --------
    $139,600 per rack (78 servers)

    Lets not forget core networking equipment. That's worth a few bucks. :)

    Each set of 39 servers would probably be connected to their routers via GigE fiber (I couldn't imageine them using 100baseT for this) Right now we're guestimating 1700 racks. They have locations in 3 cities, so we'll assume they have at least 9 routers. They'd probably use Cisco 12000's, or something along that line. Checking eBay, you can get a nice Cisco 12008 for just $27,000, but that's the smaller one. I've toured a few places who had them, and pointed at them citing them to be just over $1,000,000.

    So....

    $250,000,000 (ttl expenses)
    - $ 9,000,000 (routers)
    ------
    $241,000,000
    / $ 139,600
    ------
    1726 racks
    * 78 (machines per rack)
    ------
    134,682 machines

    Google has a couple thousand employees, but we've found that our servers make *VERY* nice workstations too. :) Well, not the Asus 1400r, those are built into a 1u case, but other machines we've built for servers are very easy to build into midtowers instead. Those machines don't get gobs of memory, but do get extras like nice sound cards and CD/DVD players. The price would be the same, as they'd probably still be attaching them to the same networking equipment. 132,000 servers, and 2,682 workstations and dev machines is probably fairly close to what they have.

    I believe this to be a more fair estimate, than the story gave. They're quoting pricing for a nice fast *CURRENT* machine, but Google has said before that they buy commodity machines. They do like we do. We buy cheap (relatively) and lots of them, just like Google does. We didn't pattern ourselves after Google, we made this decision long before Google even existed.

    When *WE* decided to go this router, we looked at many options. The "provider" we had, before we went on our own, leasing space and bandwidth directly from Tier 1 providers, opted for the monolythic sy
  • Re:Acquisition (Score:2, Informative)

    by onepoint ( 301486 ) on Sunday May 02, 2004 @04:52PM (#9035618) Homepage Journal
    Well it really depends on what your willing to spend for a drive and the quality. I will agree, 10000 drives should give you about 2 to 6 failures per week. But I have seen that sometimes in a web server situation 10000 drives have a failure rate of about 15 per week. In one case ( very very bad case ) we had a batch of bad drives come in, the first 70 had a complete failure within 4 weeks then the rest of the order failed within 6 months.... we nevered ordered that model of drive again.

    Now we have had some great luck also, where we found a brand that almost never failed for 12 to 18 months at a time, so we set up a specific policy that we used those drives as back-up redundancy drives for every main drive ( about 2500 drives ), to this day I have yet to see more than 1 failure per every 3 weeks with those drives.

    Now I have a pc at home that has been abused daily and have never had drive failure, it's been turned on every day since 1999, so it cycles completely from hot/cold and sleep/aware. maybe I'm lucky but I've abused that drive consistantly ( and back up weekly ) so maybe I'm due.

    Drive spin has become a huge factor in relation to drive failure in a web server farm. You want the fastest spin rate and at the same time you need the fast read times, but the faster spin rates give you higher failures, so you really have to learn to blend cache's, hardware and software and the dreaded mix drive raid.

    best of luck to all

    Onepoint
  • No (Score:5, Informative)

    by metalhed77 ( 250273 ) <`andrewvc' `at' `gmail.com'> on Sunday May 02, 2004 @05:43PM (#9035900) Homepage
    It would not be a very distributed DDOS and that would stop any attack quite quickly. Quite simply google's bandwidth providers (or the providers above them) would just unplug them. They may be global, but they probably have less than 40 datacenters. It would not be distributed enough to sufficiently attack. If you could take over the same number of machines with the same amount of bandwidth, but distributed globally on various subnets (say a massive virus), *then* you'd have a DDOS machine. As is, google's DDOS would be shut down quite quickly.
  • Re:$278k ?? (Score:0, Informative)

    by Anonymous Coward on Sunday May 02, 2004 @06:05PM (#9036056)
    You are dreaming. FWIW, I am responsible for 5 15k's and lots of 6800's. Sun expects that a 15k domain with 60 CPU's and 480 DIMMs will experience 7 - 8 hard failures per year.

    When a Sun server loses a CPU or DIMM it will crash. The benefit is that it will generally come back up with those parts blacklisted so you won't see another crash from the same parts. When the parts are replaced, you can bring the new ones back without taking an outage. If you happen to see that you are getting soft errors (e.g. persistent ECC corrections), you can take it offline preemptively without causing an outage.

  • Re:$278k ?? (Score:2, Informative)

    by Anonymous Coward on Sunday May 02, 2004 @06:49PM (#9036303)
    DIMMs is one thing, CPU's are another.
    If its a soft error - "cosmic ray" or what ever then it will log it and keep going. For CPU's if you haven't affinitied any processes/threads to it you should be able to do :
    psradm -v -f cpuid
    To take the processor offline yourself, obviously failing is not exactly the same but I thought I have had some fail without crashing. - Unless, of course you are trolling and if so, you got me.
  • by geniusj ( 140174 ) on Sunday May 02, 2004 @07:13PM (#9036487) Homepage
    I'm not saying that it's impossible. I'm sure any dedicated individual could do it. However, tours in datacenters are typically guided (especially at equinix). As far as getting in via unlocked doors, I'd say definitely would not happen here. You have to go through about 4 doors and 4 hand scanners to get in. There are no other entrances.

    Of course, most of it is more for show than practicality. I mean, they have hand scanners on every single cage. Definitely a little bit excessive :). However, I'm sure it impresses many decision makers.

    -JD-
  • Re:$278k ?? (Score:3, Informative)

    by tekunokurato ( 531385 ) <jackphelps@gmail.com> on Sunday May 02, 2004 @10:53PM (#9037593) Homepage
    Funny--I've spent LOADS of time on the consumer side of Brigham & Women's/Dana Farber in Boston getting various operations, chemotherapies, catscans, etc, and seeing many different doctors and nurses (don't worry I'm fine now ^_^). I'm consistently impressed with both the ubiquity and the reliability of their information systems. They're extremely universal and always seem to display quite simply exactly what Medical Care Personnel X needs to access. Perhaps a model to suggest to your clinic.
  • by B4RSK ( 626870 ) on Sunday May 02, 2004 @10:53PM (#9037594)
    The high-end Sun machines are designed for high availability. Not only will a CPU failure not crash the machine, the CPUs are hot swappable so you can replace a failed CPU without so much as a reboot.

    Yes, 10 years ago this was a important thing to have... As were many other "big iron" features. And it still sounds very cool in a geeky kinda way.

    But with redundant relatively cheap clusters available, these types of things aren't worth the $$$ they used to be.

    Except at the extreme high end of the computing world hardware is steadily progressing to commodity level.
  • Re:$278k ?? (Score:2, Informative)

    by Craigy ( 763167 ) on Monday May 03, 2004 @07:53AM (#9039016)
    As it happens on high end (regatta-class) pSeries Kit from IBM, you don't need to worry about how your app works either as the firmware reassigns your (single, very important) thread to a still-working CPU. Craigy

He has not acquired a fortune; the fortune has acquired him. -- Bion

Working...