Compelling Alternatives to RAID Setups? 113
jabbadabbadoo asks: "Our software shop has about 30 Linux servers and 15 NT servers running enterprise applications for our customers. Since we have service level agreements with most of them, uptime is crucial. One of the things we've done is to use RAID setups extensively, using products from well renowned disk- and controller vendors. However, we have discovered the paradox that introducing RAID controllers actually reduces overall uptime! Not only does more 'steel' increase the probability of failure, but what fails first is usually the RAID controllers. What is your experience? Have we been having bad luck?"
"A related problem, especially on Linux, is that setting up RAIDs is actually a quite costly process. There seems to be endless problems with library versions, and upgrading existing servers simply takes too many hours. To keep the customers happy, we routinely have to create a 'shadow' server while upgrading which in turn means we, at some point, have to synchronize data to the new server, which in turns means a bit of a downtime. Ouch. Does anyone have a good solution to these problems? Of course, cost is a major issue, but so is uptime (which also means cost if we don't provide the uptime dictated in the SLA). What setup gives the best cost/uptime ratio? Thank for any thoughts!"
RAID is good (Score:2, Informative)
Must be bad luck (Score:2)
A few tips (Score:5, Insightful)
I'll agree that setting it up is a nightmare. I'm currently helping test two 4TB arrays for use on a Linux box (16 SATA drives presented as a single SCSI device). Benchmarks under linux are slower than under windows. It's a mess figuring out why. Meanwhile, vendors (who I will not name [dell.com] ship crappy software, and take months to act on bug reports.
As for transitioning servers, I've been there too. And yes, copying a terabyte of disk in single is a very long process. It'd have taken several days, which is of course unacceptable. This is where the magic of rsync comes in handy. Copy the data over several days in advance, sync it just before the scheduled downtime, and you'll have a fairly short downtime.
Re:A few tips (Score:1, Interesting)
SATA is designed for desktops. SATA drives don't meet MTBF criteria of the equiv. SCSI drive, nor the performance.
If you've chosen it because it's the cheaper of the solutions, ok... if you chose it for performance... well, make sure you have a good backup solution.
Re:A few tips (Score:2)
Yes, they have a lower MTBF, but it's in a raid array, so who cares? It'll automagically rebuild on our hot spare, and when we wake up to the email, we just go in and replace the downed drive with another.
And yes, they have slightly lower performance. But the real performance reason for using SCSI instead of IDE is that it offloads the work to the SCSI controller/disk instead of wasting CPU time
Re:A few tips (Score:2)
Missed point - Rebuild times (Score:2, Interesting)
Have you tested how long it takes? Probably better than 24 hours if your system is moderately loaded.
Guess what you have now? The marvelous opportunity for a CASCADING FAILURE!
That's right kids! Because you just had a drive fail, and all the other drives are doing double the work to rebuild from pari
Re:Missed point - Rebuild times (Score:2)
I've had a 250gb SATA drive fail on a 1TB array on a 3ware card. It was about 4 hours for it to rebuild. The system was slower, but we didn't have 'cascading failures.'
In fact, the only time I've experienced 'Cascading Failures' was on an expensive Mylex SCSI raid controller. There is nothing like saying "Shit, we just lost two drives on that raid 5."
Banks can continue to use SCSI, but I'm going to use SATA everywhere. It'll save me over half the cost for the same
Re:Missed point - Rebuild times (Score:2)
If you spread your RAID 5 over sixteen volumes (as someone upthread said they did), it seems to me that any individual drive failing wouldn't incur a ton of work on the rest of the drives, because the amount of data any one of them would have to c
Re:A few tips (Score:1)
It's because even workstations do simultaneous read requests. SCSI has this great feature that basically when you request data from 3 different sections of the drive, it reprioritizes on the fly, picking up everything you requested along the way to the furthest request.
Re:A few tips (Score:2)
Re:A few tips (Score:3, Informative)
Life is made a litte annoying by the 2TB limit in the 2.4 kernel. But we're willing to live with that, for now. I'm told there ar
Re:A few tips (Score:2)
I wouldn't have a problem with the 2TB limitation, I've been thinking that I'd make each array no bigger than 1TB anyway (or 5 drives as RAID5, whichever
Re:A few tips (Score:1)
But we are still confused about many benchmark tests that we get, in windows 2003 and RHEL3-AS. I'd love to speak with someone who has a lot of experience with running benchmarks on RAID arrays, especially if they use SATA-SCSI enclousers, from any manfacturer.
Re:A few tips (Score:2)
The Gateway 840 would be $6,549 (if you bought the disks separately) 12x250GB ($545/disk) but that's with only 12 bays and 256MB cache. It uses StorView Storage Management from nStor [nstor.com] (the 840 is probably a re-branded version of nStor's NexStor 4700S [nstor.com]). Does anyone have any experience with StorView? It only lists RedHat as a supported Linux distro but again I'm wondering if that really matters.
The App
Re:A few tips (Score:2)
RAID 10? (Score:4, Insightful)
Brands? (Score:5, Informative)
I'm a firm believer that you get what you pay for with enterprise-class servers. You shouldn't expect Tier-1 reliability from servers that are built with commodity hardware. There is a reason that Compaq/Dell/IBM servers are more expensive.
We also haven't had any issues installing other than the default Debian boot disks not supporting the SmartArray controller. A custom set of disks took care of that though.
Jason
Re:Brands? (Score:1)
We are using Novell Netware as well, so this is more of a comment about hardware reliability rather than software woes.
Software RAID? (Score:4, Interesting)
In the past two years, none of the "downtime" that I've experenced has been attributed to the disk array or controller.
The biggies have been: power outage that exceeded the capacity of the UPS (3 hours), planned upgrades and an anonymous gremlin who bumped the reset button - since detached.
Re:Software RAID? (Score:2)
I had to compile a version of the kernel with the auto-detect and activate RAID feature turned on. That way I could make the disk array my root file system.
To your question, The disk array is abstracted to appear as a single device. The disk array has a mission to defend against problems below it, not above it. So if the filesystem code has bugs and corrupts itself, the array doesn'
So would XSan help? (Score:5, Interesting)
Re:So would XSan help? (Score:2, Informative)
From what I read, the XSan software is first and foremost a distributed file system for shared volumes from the Xserve RAID.
If you look at the applications, it's about multiple servers or workstations with concurrent access to a single volume - distributed file locking.
Great stuff for the stated purpose, can't wait to get my hands on it!
Hiding the complexity of RAID is the domain of storage 'virtualization' solutions. The ones that let you mix and match raid types across any
Major problems with Promise RAID controllers. (Score:5, Interesting)
This is on a lower level than the RAID you are using, but we are having major problems with 10 Promise Technology TX2000 mirroring RAID controllers that we bought. The mirrors go critical for no detectable reason. Promise Technology technical support is unable to find the problem, and the company is unwilling to escalate the issue. The Promise Technology technicians escalate the issue, but 2nd level technical support never calls back.
Promise mirroring controllers on ECS (EliteGroup) L7VTA v 1.0 motherboards have the same problem. When we call ECS tech support, there is a recorded message saying they are busy and to call back later.
We've been supplying computers with Promise mirroring RAID controllers since the company began doing business, and we've had very few problems until now.
Possibly the problems are associated with newer, faster motherboards, or with AMD VIA chipset motherboards. We've never had problems with RAID controllers on Intel chipset motherboards.
Another possibility is that the RAID controllers are incompatible with DVD burner drivers that are installed with Roxio or Nero DVD burning software.
Re:Major problems with Promise RAID controllers. (Score:1)
PC Chips motherboards? How so? (Score:2)
What do you recommend?
Re:PC Chips motherboards? How so? (Score:1)
Most business people will see that hardware is dirt cheap compared to downtime.
Re:Major problems with Promise RAID controllers. (Score:4, Informative)
Those are not really true hardware RAID controllers. They are regular hacked up IDE controllers with a bit of BIOS firmware on them that handles software RAID via INT13 until the OS loads and the software RAID in the "driver" can take over.
They offer nothing that a legitimate hardware raid setup should give you such as cache RAM or CPU offloading. Mirrored setups on these types of pseudo-hardware RAID controllers HURTS PERFORMANCE. Don't believe me? Benchmark it yourself versus software raid and hardware raid on a real controller such as Adaptec AAA or 3ware...
Do you have a link? (Score:2)
The performance we get with Promise controllers (when they work) has been satisfactory. The application is a cash register; the computer is always faster than the operator. We only need a mirror copy of our data.
3Ware told me they cannot boot from one drive, after one fails. A 3ware formatted drive cannot boot from the IDE controller on the motherboard. Promise can do both. We need features, not performance, in this case.
Do you have a link to an Adaptec IDE mirroring RAID controller you would recom
Adaptec ATA RAID 1200A = HighPoint RocketRaid 133? (Score:2)
Question: Are Adaptec ATA RAID 1200A cards the same as HighPoint RocketRaid 133 cards? I notice the BIOS setup screens look identical.
Re:Do you have a link? (Score:2)
Even then, why can't you just plug into a 3ware again and get the data off the drive?
The Adaptec you linked to is another one of those software-driven RAID cards, and offers no real value.
You're going to have to spend $200-$300+ to get a decent RAID hardware card, and then make sure you have
Re:Do you have a link? (Score:2)
It's a nice backstop that Promise "arrays" are still accessable with conventional hardware.
This author is talking about cash registers - where they're likely out and about all day from place to place doing a service run with
Re:Do you have a link? (Score:2)
That is my primary reason for using software raid. The other reasons being that a raid card is much more expensive than an ordinary IDE controller, and I have read more than once, that it is really still software raid. My setup with three 120GB disks and identical partitioning of all three disks goes like this. One 31MB
Re:Do you have a link? (Score:1)
We're using Windows 2000 & 2003, and trust me, it's simpler to use a Promise card to mirror the boot volume than to use 2k's software mirroring. Recently moved to SATA based drives.
There are advantages to the promise cards still - with the enclosures, you get hot-swap and you get status LED's and what not. (We're hoping to be able to say to person on the phone, "which one has the orange light?")
3Ware said no, in a telephone conversation. (Score:2)
What you said is what I would expect. However, I called and talked to someone in 3Ware technical support, and he said it would not boot with only one drive; it would be necessary to rebuild the array to boot the computer. Maybe you are using a different controller.
In addition, here are questions and answers from a session on the 3Ware chat system:
Request:- 10th January 2004 at 8:23
[Irrelevant questions removed here.]
We've been using Promise RAID controllers with our cash register software, an
Re:Major problems with Promise RAID controllers. (Score:2)
Actually, they do provide one particularly useful feature and that is to present the RAIDed disks to the OS (and the BIOS) as a single device.
Mirrored setups on these types of pseudo-hardware RAID controllers HURTS PERFORMANCE.
The software overhead for RAID1 should be, for all intents and purposes, insignificant. It just doesn't *do* anything that requires much CPU work.
Re:Major problems with Promise RAID controllers. (Score:2)
What it does is blocking. A good hardware raid mirror will have a battery backed up cache so it can acknowledge the write as successful either immediatly or after the data is on one disk, which a software raid setup can't do reliably. What you end up with is the additive rotational latency for two disks, which can signifigantly hurt performance for small random
Re:Major problems with Promise RAID controllers. (Score:2)
The int13 firmware does this for the code before the OS loads and the driver is responsible for doing this job afterwords. Don't let this software trickery fool you. Behind the veil of the driver, the system software is reading and writing to the two disks individually. The int13 stuff is a nice trick, but it's only necessary due to the inability to replace the OS's
Re:Major problems with Promise RAID controllers. (Score:2)
I'm well aware of how the "trickery" work
Re:Major problems with Promise RAID controllers. (Score:3, Informative)
On a real hardware raid controller this overhead exists only on the controller CPU (normally an i960 or somesuch) and is further alleviated by the cache ram on the card.
Compared to what, though ? OS-level software RAID is going to have to do precisely the same thing and IMHO the processing involved, taken in the context of modern, fast CPUs, is insignificant.
Well, I wasn't trying to compare promise/hpt/et a
Re:Major problems with Promise RAID controllers. (Score:2)
No commonly used software RAID does this.
Re:Major problems with Promise RAID controllers. (Score:2)
Of course they can! Data access to disk is never necessarily serialized at any point in the process. Even the drive firmware itself tries to read requested blocks in the most optimized order possible.
On a SCSI, this feature is called command queueing. It's present in the SATA specification also, but is not a part of IDE or ATA. It's kind of a 'fire and forget' thi
We are using good supplies. Rebooting continuously (Score:2)
We are using power supplies that seem like the best (KingWin), although not expensive.
We have tested these units by putting a re-boot program in the Win XP startup folder. This causes continous reboots. We have run several computers more than 12 hours continously rebooting. This should show problems with the power supplies.
We do NOT see problems, usually, with the Promise Tx2000 controllers when continously rebooting. The problems come after the units are delivered to the customer, a terrible situat
Sounds like a design problem on your end. (Score:5, Informative)
The only time RAID should ever be down, is during initial setup. Thereafter you should replace bad drives while it's running and you should never have cause to shut it down due to a RAID issue.
If you are experiencing RAID hardware problems then take a good look into these areas:
RAID Hardware --> Are you using cheap stuff? It honestly isn't worth it. Perhaps you're just discovering the 'real' value of 'cheap' hardware.
RAID Software --> If you're using unsupported drivers (ie, vendor doesn't supply or support them) then ditch the hardware and get hardware with supported drivers - make sure they support them on your configuration. You've already proven that you can't support them yourself.
System Hardware --> If the system is generally cheap (cheap power, bad airflow, cheap components, etc) then you simply can't expect the RAID card to work 24/7.
Server Room --> Make certian your server room can handle the power and ventilation needs of the servers. This should go without saying, but all too often it is the problem.
The reason people go with cheap components is the lower initial cost. They only work for a few thousand hours of heavy operation. You must get server rated components if you want them to operate for more than a year or two. There really is a difference.
Lastly, I use 20+ Promise FastTrack ATA RAID cards in 20+ Novell networks. I use cheap components, and they work in harsh conditions. They are not set up for hot-swap, as that's not a need in this situation. I have to replace the cheap hardware every 2-4 years, powersupplies every year, hard drives every 2-3 years. The only time the RAID cards have gone bad is when a power supply failure (usually due to a power outage/surge/brownout) fries the motherboard and usually most of the components in the case.
I have never had a failure where both HDs completely failed simultaneously, though usually when the rest of the computer goes I replace the whole thing and get the data off one of the old hard drives. This is not an advertisement for Promise. They simply are the only one's with supported Novell 3.12 drivers.
I'd be surprised if you've covered all these bases and are still having problems.
-Adam
Re:Sounds like a design problem on your end. (Score:1)
Re:Sounds like a design problem on your end. (Score:2)
The cards I use store array info on the hard drives themselves. I can move the hard drives from one card to another without reconfiguring anything.
Of course, I would suggest staying away from RAID cards that use batteries...
-Adam
Re:Sounds like a design problem on your end. (Score:1)
Better know how to replace your batteries (Score:2)
What Promise FastTrak RAID controllers? (Score:2)
What Promise FastTrak RAID controllers are you using? As I said in the comment just above yours, we are having major problems with FastTrak Tx2000 mirroring controllers.
Re:What Promise FastTrak RAID controllers? (Score:2)
-Adam
The FastTrak Tx100 always worked for us, too. (Score:2)
The Promise FastTrak Tx100 cards always worked for us, too. The only Promise cards that fail for us are the Tx2000 cards. Since we have been unable to get help from Promise for this problem, I presume they know there is a problem, and are unable to fix it.
Multi-engine aircraft (Score:4, Insightful)
However, which would you rather be in, a twin engine aircraft that just lost one engine, or a single engine aircraft that just lost an engine?
Yes, RAID cards die - I've been shocked at how often that happens. And 5 disk RAID will have more failures than a 4 disk JBOD (just a bunch of disks) array.
But the question is, are you seeing a reduction in UPTIME, or just in mean time to failure? Maybe the RAID system throws an error once a month and the JBOD system throws an error every two months, but if you can recover in 5 minutes by swaping cards or drives rather than 5 hours for restoring the JBOD from backup, you are better off.
Perhaps what you might look at would be using RAID software on the server's processor, coupled with Firewire drive bays, disks, and multiple Firewire cards. If you have a card die, move the disks to another card until you can schedule downtime. A disk dies, hot-swap and rebuild in background.
Re:Multi-engine aircraft (Score:1)
That depends, can the twin engine plane successfully land with only one engine running?
Hardware, Configs, Backups (Score:5, Informative)
1. Buy quality hardware.
IDE RAID for critical servers is a bad idea.
In my experience, RAID hardware tends to be very picky and suffers from subtle and often bizarre hardware conflicts. In general, using a RAID solution that is packaged with the hardware is the best idea.
If you cannot afford good RAID hardware, stick to conventional JBOD configurations.
2. Configuration
Design your the configuration of your systems around consistency first, performance second.
You need to document your procedures for building servers, allocating storage, etc. Create scripts whenever possible.
If you are not confident that you could not talk a marginally qualified technician through a server rebuild over the phone, your docs aren't good enough. If you don't have the time to write docs, make the time or work late.
3. Backups
You need documented, tested backup AND restore procedures. All of your oncall staff need to be able to restore a server.
With 50 servers, disk controller or disk failures should not be a common event. We work with approximately 400 datacenter and 200 field servers (varying in age from 1-9 years), and replaced 3 controllers and 19 disks last year.
Look for electrical issues, you may have crappy electrical service.
What RAID controllers would you recommend? (Score:2)
What RAID controllers would you recommend?
What hardware is "quality"?
Re:What RAID controllers would you recommend? (Score:2)
In the past I've worked with Compaq hardware, which I believed shipped with Symbios controllers. (Been awhile)
Sun storage was usually good, except that we tended to get alot of flaky gbics and cabling from them.
"Certified" hardware really is important, especially in larger environments, where wasted time is more expensive than buying the vendor's recommened hardware. A good example is when due to a supply shortage, we ordered a differe
Re:What RAID controllers would you recommend? (Score:2)
Re:Hardware, Configs, Backups (Score:1)
Granted, I'm fully in agreement with you--SCSI is more reliable and better for processing servers etc. But when it comes down to a cost-effective way to get a hell of a lot of disk, I can heartily recommend the 3Ware Escalade stuff.
The most trouble I ever had with raid... (Score:5, Interesting)
Maybe you were the first. (Score:2)
So, did you "post" that raid-5 + raidtools + kernel 2.6 locks up a computer, to save somebody else going through what you had to ?
Storage Cluster (Score:2, Interesting)
Look at Google (Score:3, Interesting)
You can still do RAID with this setup but you'd have the added security of 2 or more systems making up your entire functional system so if one is down the other can continue normally. Then it's trivial to repair the dead machine and bring it back into the cluster.
Re:Look at Google (Score:2)
The compaq proliant 7000 (xeon) I've got came with a smart array 3100es, which does raid on the 3 hot swap scsi cages, and there's a special slot on the i/o board for a second 3100es, so that if one dies, i can just hotswap in a spare. (the two slots pci-x plus 3 channels of scsi that go to scsi cages, and a fourth scsi channel for controller to controller communication)
Re:Look at Google (Score:2)
Well you obviously don't know anything about proper RAID then do you? All enterprise storage costing 25k+ at least has this option.
The normal configuration is an array, which has 2 controllers in it. You create LUN's, and assign them to the primary + secondary controller. The primary + secondary controllers have a heart beat, which ensures one takes over the others configuration if it fails. You dual attach your host to each controller. Set up IO
Re:Look at Google (Score:2)
Hey...be nice.
Dell sells some cheapie dual/redundant controllers for well under $10K -- I know that they're available in their tower servers for sure.
Re:Look at Google (Score:3, Informative)
For their production service, I understand that they keep it all in memory. But it's hardly temporary.
Hardly anyone is like Google.
For now. Google was one of the first companies to take advantage of the fact that RAM and procesing power have become ridiculously cheap. SQL databases arose in an era when 32k was a fair bit of RAM, and where a business computer was one or more refrigerato
Re:Look at Google (Score:2)
There are whole classes of applications where that can't possibly work. If you'r
just unlucky (Score:2)
You say you've had more raid controller failures than disk failures. Did any of the raid controller failures require a restore from backup? A non-redundant-disk failure would have.
Add up the total time you were down due to raid controller failures and the total time you would have been down for disk failures if you didn't have raid. That's a better measure than instances of failure.
RAID Alone != good design (Score:5, Informative)
You can't slap a buzzword like RAID onto whatever you were doing before and expect results. Reliable systems have to be carefully engineered correctly.
From the sound of your posting, I'm assuming when you say you're using RAID, you mean internal RAID cards inside a server with internal disks attached, and relatively small amounts of it. In these types of scenarios, the highest performing, most reliable, and most cost effective option is to put two seperate scsi controllers in your boxes, buy twice as much storage as you need, and mirror between the controllers using the OS's software mirroring capabilities. You are now indepedant of controller failure, the controllers themselves are less likely to "fail" (which doesn't always mean hardware frying) than a complex raid controller by their simpler nature, and you're getting the performance benefit of full mirroring instead of that clunky raid5 business. If you have enough storage to warrant four or more internal disks of some size, use mirror+striping. Always mirror at the lowest level, and then stripe on top of that (in a 4 disk design actually it doesn't matter which way you layer them, but in 6+ disk designs it gives higher data availability in the unlikely event of multiple disk failures). Or in other words - raid5 and hardware cards = bad, mirroring/striping + software raid = good.
Your goal is not to be buzzword compliant by slapping in a raid controller, your goal is to carefully analyze your systems, your options, your requirements, and your budget, and eliminate single points of failure everywhere that it's feasible and desirable to do so, starting with the lowest MTBF items in the system and working your way up. There are no magic bullet answers of course - change the situation and the "right" answer can change dramatically.
Re:RAID Alone != good design (Score:2)
You're right of course, I was just on a run there typing quicker than I was thinking. Striping first means you can only tolerate a single disk loss, period. Mirroring first means that after the first disk loss, there's only a 33% chance the 2nd disk lost will cause data loss. So yeah, even in a 4-disk setup, the same rules apply - mirror at the lowest level for better availability
There is a reason EMC and others charge $$$ (Score:1)
Shadow copies? Look at SnapView and SANCopy in EMCs CLARiiON line - no downtime to create a copy. I would expect Hitachi and others to have similar features. There are a lot of used EMC disk arrays on Ebay and other places - just make sure yo
Fibre Channel and the Xserve RAID (Score:5, Informative)
Bang for the buck, you can't beat the Apple Xserve RAID. They are IDE, but almost as fast as the fastest scsi arrays, and seem to be very reliable. The array can be easily partitioned into a variety of raid types with hot spares. The unit can then connect to Windows or Linux via standard fibre channel interface and look like simple scsi drives. The RAID is administered via an ethernet connection using a nice java gui tool.
We set our Xserve RAIDs up such that each array (each Xserve RAID box has 2 arrays with separate controller logic for each) is RAID 5 plus a hot spare, and then the array is mirrored with the other one. This gives is
Better deal than Xserve RAID (Score:2)
Gateway 840 Serial-ATA RAID Enclosure [gateway.com] is cheaper per GB than Xserve RAID. It has 12 bays and uses U320 SCSI instead of Fiber Channel for the connection to the system. Currently the cheapest config you can do is $4,749. That's with 4 250GB SATA drives and their cheapest 3yr warranty (another nice thing is you can increase the warrany to 4 or even 5 years and they have a variety of response times you can choose). Gateway gives you all 12 carriers no matter how many
Re:Fibre Channel and the Xserve RAID (Score:2)
Yes you can. Easily. Shop around even a little, you'd have to work pretty hard to find an ATA-based solution as expensive as theirs.
Netblock device RAID? (Score:2)
RAID (Score:1)
Re:RAID (Score:2)
RAID 10 is when you take 2n raw drives, building n mirrors (The RAID 1 portion of RAID 10). You then take the n mirrors and put them in a RAID 0 stripe.
RAID 0+1 is less preferrable, but is sometimes all you can do. Take 2n drives, now build two RAID 0 stripes in n devices in them. Now, take the 2 stripes and mirror them
Re:RAID (Score:1)
Re:RAID (Score:2)
http://www1.us.dell.com/content/topics/global.asp x /power/en/ps1q02_long?c=us&cs=555&l=en&s=b iz [dell.com]
That's a link to Dell documentation discussing the in's and out's of RAID configuration and reliability. Any chance you've got a link that shows where a mirrored RAID 5 configuration is referred to as RAID 10. I'm
Re:RAID (Score:1)
Re:RAID (Score:2)
Mylex 960 (Score:1)
Compaq's newer SmartArray have seemed to be more stable.. however only time will tell.
-un1xloser
If RAID cards are failing, that's important! (Score:2)
If Mylex cards are failing, that's important! If RAID cards fail, then the company, and all its employees, are out of business. And that's what apparently happened to Mylex [lsilogic.com]. It's now owned by LSI Logic.
At the low end of the scale, we seem to be having the same kind of problem. We are having a high failure rate with Promise Technology FastTrak Tx2000 controllers. Promise Technology seems to have lost the will, or maybe ability, to deal with problems.
When I read through the comments to this story, th
Re:If RAID cards are failing, that's important! (Score:2)
It seems that most (all?) of those stories relate to controllers doing IDE RAID. I suspect the answer to the question of why so many are failing is that it's still a relatively new technology, only really widely available in the last 18 months. SCSI RAID controllers on the other hand don't seem to be plagued with the same issues.
This is a new problem. Did Microsoft do something to break m
Buy Windows 2003 server? (Score:2)
Microsoft's solution is that everyone should buy Windows 2003 server and use software RAID, available only on that Windows OS.
That's all we need is software RAID mirroring, but it doesn't make sense, for this application, to support a much more complex system and much more expensive system to get it.
Re:Buy Windows 2003 server? (Score:2)
But Microsoft only recommend software RAID for small environments. They don't even use it themselves - they use massive HP EVA SAN's.
Better Hardware? (Score:1)
I have been working with compaq proliant servers for several years (support for RedHat Linux is good) with nary a hardware problem.
http://h18004.www1.hp.com/products/servers/prol i an tml530/index.html
http://h18004.www1.hp.com/produ cts/servers/prolian tstorage/arraycontrollers/index.html
http://h1800 4.www1.hp.com/products/servers/prolian tstorage/drives-enclosures
Don't mirror HD - mirror the server! (Score:2, Interesting)
Re:Don't mirror HD - mirror the server! (Score:1)
I am having a hard time coming up with a scenario where "yesterdays data" wouldn't get me fired.
Change your design (Score:2)