Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

I have been following this math /very closely/ - and I am fairly certain that if someone wanted to build a board like this in quantity, with more ram, of course, and if I could get my hands on a suitable power supply so I would have one psu for every twenty of these servers, it would likely come out pretty close to the cost of a big box.

The problem is that all boards I can get for a reasonable price in the quantities I can afford have a bunch of crap on them I don't need; video and the like, which pushes the cost up higher than I'd like. I've been talking with a friend of mine who does ARM stuff about maybe seeing if we could get venture funding, as I think we could get into reasonable quantities in the low-six figure range.

But, like I said, as far as I can tell, right now the cost difference is 'close' - even if I could get a custom board with lots of ram and no video crap, for most things it wouldn't be dramatically better than virtualizing a massive supermicro box, and considering that design time would mean that the thing would be kinda old by the time we were done, there probably isn't a lot of profit to be had. But, if trends continue, this may change; and I can tell you right now, customers are willing to pay a premium for dedicated hardware, and for good reason.

(Yes, you have the management problem of thousands of servers, but most of the things that don't have PXE have some way to boot from serial, and I can deal with that. Having managed thousands of virtuals and tens of thousands of hardware nodes, I'm not convinced that virtuals are all that much easier to manage, once you have a good system in place.)



What your not taking into account is that thousands of small servers will likely fail faster than dozens of big servers. I think you'll find that this and the fact that you'd have to maintain much more hardware (the servers themselves and switches) tips the balance.


I've managed clusters of 60K plus large servers. I mean, I didn't manage them by myself; my primary job was automating hardware failure detection, but I took my turn on the pager. My personal focus was to see to it that bad hardware didn't go back into production; as someone who shared pager, that was... offensive. But really, hardware failure is less of an issue than people seem to say. I mean, sure, it happens. But not nearly as often as software issues that bring down the box, unless you count bad hard drives; and those happen all the time.

Even with 60K servers, I quickly learned that if I found a lot of bad servers that were bad in similar ways, I needed to be careful before having them sent back to the vendor. More than once I embarrassed myself blaming software problems on the hardware. For the first 3-5 years of it's life, if the computer was properly assembled, very likely the only things that will need replacement are the disks.

My thought would be to treat the things as hard drives. two options: For the lowest tier of support, if your system fails, I leave it and provision a new one. Leave the dead ones in the rack until it's time to upgrade. The second option is to put the boards on some sort of hot-swap power backplane, and when the thing fails, pop it out and replace it like a hard drive.

The question would be "how expensive is it to make the hot-swap system, and does the manufacturer give me a warranty that makes it worth my time to swap 'em out and send 'em back."

I suspect that if I can get 'em to give me a discount to not have a warranty, my best bet would be to set it up so I could hot swap whatever i used for storage, but if anything but storage failed, I'd just leave the system in place.

sure, maybe I'd want to offer 'premium' support with mirrored storage where I'd swap your storage if it fails, and yeah, in that case, expenses would be higher than a giant server with fewer disks. (by fewer disks, I mean fewer disks per 256mb guest. Obviously, it has more disks total) I/O performance is probably also going to be worse on the giant server with fewer disks, so I can charge more for the 'premium' dedicated hardware with dedicated I/O.

I mean, you get into the 'all troubleshooting must be automated' range much faster with those than with physical hardware, but somewhere between a couple hundred and a couple thousand hardware nodes, you are going to get there either way.


OK, well I'm certainly more convinced now by your idea, but I feel that there are too many unknowns to be able to say confidently that it would actually be a viable alternative.


oh yeah, it's definitely still in the "maybe" zone; and right now, things still lean towards VPSs being enough more efficient that it doesn't make sense to try little ARM servers seriously. I'm just saying, things are moving towards the point, as far as I can tell, where tiny arm servers could be competitive, maybe, with VPSs.

The big problem that is stopping me from betting my money on little ARM servers is that nobody makes commodity-priced ARM boards that use the latest ARM CPUs and have reasonable amounts of ram and don't come with a bunch of stuff like video that I don't want to pay for.

I mean, even if I had that, it's still pretty borderline and quite possible that I'm wrong and it still wouldn't make sense; but it's close enough that I'd be willing to bet a large (by my standards) chunk of my money on it.


A few thoughts:

* It seems likely that there exists an ARM IC that would suit your purposes, you'd just need to commission a PCB and if you'd planning on having 10s of thousands of them, then that would be no great cost.

* Having more memory will increase the board area and cost significantly.

* You might be better off PXE booting and using a SAN rather than having local disks.

* If you wanted to test the waters, you could try stuffing dozens of raspberry pis into a 4U case with a switch and co-locating. If this took off, you could develop your own custom hardware.


Well, a SAN has other problems; the primary advantage of a dedicated over a VPS is I/O, and SANs, well, you know with software, you say "good, fast, cheap; pick two" - with a SAN, really, the hard part is software. In this case, "good" means "reliable" and you have to be really careful to get two; usually you only get one.

I have a lot of admiration for Amazon's engineering; but even they can't come up with a reliable, performant, and cost-effective SAN solution; and they are able to charge a lot more per gigabyte and per iop than I can.

Now, it's a tradeoff, I mean, local hard drives will /massively/ increase the failure rates, and moderately increase the cost of the individual servers.

I've got someone close to me who is an ARM person who has done PCB layouts and has participated in making custom SBCs; according to this person, the nice, fast dual-core chips you see in cellphones that we'd like to use are often unavailable in quantities of around 1000, which is where I want to start out (even this, 1000 at a target cost of $100 each, is going to require investment. I can drop maybe 1/5th to 1/10th of that of my money, but there is zero chance of custom hardware at the volume I can pay for being economical. My understanding is that you start to see reasonable pricing at around 1000 units.)

Memory, yeah, will be a big cost. But it's also really, really important; Until you have enough ram, nothing else matters.


Ok then, it seems unlikely that you'll be able build custom hardware without a an investment and it also seems unlikely you'll be able to get investment without a proof of concept.

How about, as a proof of concept, raspberry pis, each with a 16GB SD card. I know there are obvious problems: ethernet speed, non-volatile storage speed and only 256MB of memory, but do you think that it would work as a proof of concept?


Depends on what you are testing. The big problem is that the 256MiB ram server market is selling to different people than the 2048MiB ram server market. Different people buy them for different reasons. I mean, there is some overlap, but the 256 people? Mostly hobbyists, some really cheap people doing development or test. The 2048? people who are well off doing test and development, or cheap people doing production. 2048 is the top end of my people and the bottom end of 'serious people' - I wanted to sell dedicated servers as something higher end than my VPSs, and you'd need at least 2048MiB ram for that. Preferably double that or more, but we take what we can get.

On the other hand, if I just wanted to demonstrate I could build a system to manage those small servers, then yeah, your suggestion is great. The management would be about the same. But I'd be selling to a different market, so it wouldn't work as a market test, but I could show the investors that I could make the things go.

I am interested in the 256MiB market; I mean, the hobbyists are my core customers, the people I know how to sell to and serve best. But really, having a stand-alone server is much less compelling, technically speaking, for those people than for the 2048 customers. (and having a sd card is less compelling than a pair of 2.5" drives) really, I'm pretty uneasy about anything with unmirrored storage; the hobbyists, while data loss has a lower dollar value than commercial customers, are usually less prepared with backups and the like to deal with hardware failures. Amazon can sell disks that go away; I can't.

Still, a bunch of 256 rasberry pi and two SD cards, if priced in the half-sawbuck range, would probably sell pretty well to the demographic I know. Of course, that's good profit before you count the storage, and assumes the things can handle Power over ethernet or some other cheap psu. In hosting you usually expect the hardware to earn itself out after 4-5 months if you don't count other costs.

For those people, the ethernet speed (or really the sd card speed) won't be a huge problem. for ten bucks a month, what do you want?

now, the other thing to think about is that VPS prices are about to drop again. (I mean, resources for the same money will increase) I mean, I probably won't drop mine until signups slow or Linode makes a move, but you can get 8GiB reg. ecc ddr3 ram modules for under a hundred bucks. this is silly cheap. You are now paying for your servers in four months paying frigging dell prices (and nobody at my end of the market pays anything like the dell tax.) If my $8 plan (which right now has 256M ram) doubles to 512m? suddenly you go from paying two bucks more for your own hardware to paying two bucks more and getting half the ram. I mean, I'd still get takers, but not nearly as many.

And really, I don't know. I know 'my people' would be willing to pay me money for ARM servers; I really have no idea how that would translate to the behavior of 'serious people' so maybe it's best for me to play around the 256 level first?

Man, I sure wish these things had ECC. that's important.

drat, and they also don't have PoE; that's going to add to the cost and management hassle (I mean, PoE is nice for this 'cause the switch is also a rebooter if the thing gets really hung. down the port.) without PoE, I will have to wire up a power supply /and/ a way to hard reboot the servers, which I think I could do in one go with PoE.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: