With $45000 in hardware, over 100 hours spent pushing it to it’s limit, speeds of 2.7GB/s and over 461,800 IOPS, we are fairly certain that this RAID Controller and SSD review will be one of our most exciting reports to date, and not matched anytime soon.
Today we are going to go head-to-head with the LSI 9265-8i and the Adaptec 6805 RAID cards, these being two of the most powerful RAID devices available worldwide. Grab a seat and buckle up as the hunt for the most powerful 6Gb/s RAID card solution is on! Who will emerge the victor?
Our goal is to showcase the raw power of both solutions under some of the most demanding situations. In the world of servers and datacenters, the most power in the smallest package is ‘King’ and both of these devices aim to deliver!
We have amassed several sponsors to partake in this shootout which assures us that we have top performing storage devices available today, both commercially and in the enterprise sector. The purpose is simply to explore the absolute maximum of 6Gb/s RAID performance. In order to saturate these controllers, we had to bring in the heavyweights of SSD and EFD performance as the only way to illustrate the true capabilities of these controllers is to have the absolute fastest devices available today.
The competition for rack space between these companies is stiff as both look to dominate tomorrows data centers. With the explosion of data centers around the world, and high speed devices such as SSDs and EFDs coming into play, the onus is on these companies to capitalize on the game changing advances that have been made in the past few years in storage technology.
New emerging caching capabilities are coming to the forefront giving customers the option of upgrading-in-place. Only by finding ways to utilize these new technologies and the performance of the most robust solutions can these companies hope to gain rack space, and thus, market share. With some surveys indicating that 83 percent of data centers are planning significant expansions within the next 12 to 24 months, the stage is definitely set for a showdown with billions of dollars hanging in the balance.
The performance of these solutions can be broken into several categories and, by no means, do we intend to test cover all in this review. There are simply too many variables that these types of devices are designed to operate under.
We have broken the review into results from several standpoints. Performance and then of course, Price versus Performance. Performance will be broken into two segments, Throughput and IOPS.
- Performance: IOPS and Throughput – In these scenarios, we will blatantly attempt to push these controllers to the very limit of their power. IOPS will consist of the small file random performance that is so demanding and throughput will be massive sequential power.
- Price Vs. Performance – These two solutions both represent the fastest devices from the two largest RAID card manufacturers. Both utilize a Dual Core ROC, which is a dual core processor for the RAID computations. The price of both devices is somewhat different so we will strive to answer, which is better bang for the buck?
The features and performance of both of these flagship controllers are just a few clicks away, so join us as we first take a look at the contenders…
I unfortunately have the Adaptec 6805 controller and… LOL… where is my garbage can…
I’m running a setup with 4x Kingston 240GB HyperX in R5, the ATTO results arent that shabby… I get 1.5GB/s write and 2.1GB/s read (seq) but…. the short random reads are just terrible… in Winsat it’s only 170 MB/s(!?!).
Funny thing is that when I turn off controller cache I get 9000 MB/s in sequential 64 read… 🙂
I dont however get more than 5ms max response time vs your 55ms value.
Winsat with all cache OFF (Swedish W7):
> Disk Sequential 64.0 Read 8922.99 MB/s 7.9
> Disk Random 16.0 Read 183.16 MB/s 7.6
> Responsiveness: Average IO Rate 1.01 ms/IO 7.9
> Responsiveness: Grouped IOs 7.94 units 7.5
> Responsiveness: Long IOs 1.98 units 7.9
> Responsiveness: Overall 15.69 units 7.9
> Responsiveness: PenaltyFactor 0.0
> Disk Sequential 64.0 Write 576.71 MB/s 7.9
> Genomsnittlig l„stid med sekventiella skriv†tg„rder0.494 ms 7.9
> Svarstid: 95:e percentilen 0.936 ms 7.9
> Svarstid: maximal 2.820 ms 7.9
> Genomsnittlig l„stid med slumpm„ssiga skrivningar0.542 ms 7.9
Winsat all cache ON:
> Disk Sequential 64.0 Read 1207.40 MB/s 7.9
> Disk Random 16.0 Read 170.22 MB/s 7.6
> Responsiveness: Average IO Rate 0.26 ms/IO 7.9
> Responsiveness: Grouped IOs 4.58 units 7.9
> Responsiveness: Long IOs 0.45 units 7.9
> Responsiveness: Overall 2.04 units 7.9
> Responsiveness: PenaltyFactor 0.0
> Disk Sequential 64.0 Write 882.76 MB/s 7.9
> Genomsnittlig l„stid med sekventiella skriv†tg„rder0.239 ms 7.9
> Svarstid: 95:e percentilen 1.013 ms 7.9
> Svarstid: maximal 5.359 ms 7.9
> Genomsnittlig l„stid med slumpm„ssiga skrivningar0.326 ms 7.9
Next steps to do:
PC –> Adaptec –> Garbage bin
LSI –> PC
Hahaaa… my 6805 beats the LSI in the AS SSD copy benchmark 2 of 3:
ISO 880 / 1.22
Program 377 / 3.73
Game 701 / 1.97
As an existing 6805 owner running four (4) Corsair 256GB (SATA-II) SSD’s (just have not upgraded to the SATA-III’s yet as my 2 year-old SSD’s are working so well); I was always interested in the LSI 9265-8i ever since you first published that article of the card with the eight (8) Patriot Wildfires. Wow! Did that ever open my eyes. I had just not been a huge fan of LSI products because we have that controller running (embedded) in all of our Dell PowerEdge Servers and I was just not over-impressed with that chip – and besides, I was an “old-school Adaptec fan-boy†as they’ve just been around so much longer.
However, that has all changed with this article and the prior one – Thanks Paul for an awesome review – as usual.
Once – and if – my existing Corsair SATA-II SSD’s ever die or crash, then I’ll be looking at upgrading my current workstation with the LSI 9265-8i (w/ FastPath); along with eight (8) OCZ Vertex 3 120-GB SATA-III SSD’s and I might even add in that ARC-4036! The entire cost of everything is nearly identical to a single OCZ 960GB PCI-e RevoDrive 3 x2 card and yet, I should be maxing out the LSI RAID cards full capability.
I’ve never “wished†for an existing SSD to die just so that I could “upgrade†– but if ever there was a time to crash and burn, it’s now baby! Problem is, these dang Corsairs are just rock-solid for over two years now . . . thus, I could be waiting for a while longer {: = (
Thanks guys for your feedback! It is nice to know that you guys are reading.
@bamer8906- I hope to see some of your results when you upgrade. I most certainly do not hope that you lose your devices, unless of course you have exhibited due diligence in regards to your data backup scheme 🙂
@Eero it would be helpful if you could post any of your testing results in our forum (as linked on the last page of the review).
This would allow us to discuss your testing and methodology in depth, so that we can get an accurate representation of your accomplishments.
Many things, such as motherboard, cpu, drives, stripe sizes, card settings, can come into play here, so we would really need a “big picture” of these types of parameters.
I think that the comments section of the article will not provide us enough room and space for this
Thanks guys for your feedback! It is nice to know that you guys are reading.
@bamer8906- I hope to see some of your results when you upgrade. I most certainly do not hope that you lose your devices, unless of course you have exhibited due diligence in regards to your data backup scheme 🙂
@Eero it would be helpful if you could post any of your testing results in our forum (as linked on the last page of the review).
This would allow us to discuss your testing and methodology in depth, so that we can get an accurate representation of your accomplishments.
Many things, such as motherboard, cpu, drives, stripe sizes, card settings, can come into play here, so we would really need a “big picture” of these types of parameters.
I think that the comments section of the article will not provide us enough room and space for this.
What about the 9260 Lsi controller? Thats a more apples to apples comparison?
these are comparisons of the top-tier offerings from each manufacturer at the time of publication.
Thanks guys for your feedback! It is nice to know that you guys are reading.
@bamer8906- I hope to see some of your results when you upgrade. I most certainly do not hope that you lose your devices, unless of course you have exhibited due diligence in regards to your data backup scheme 🙂
@Eero it would be helpful if you could post any of your testing results in our forum (as linked on the last page of the review).
This would allow us to discuss your testing and methodology in depth, so that we can get an accurate representation of your accomplishments.
Many things, such as motherboard, cpu, drives, stripe sizes, card settings, can come into play here, so we would really need a “big picture” of these types of parameters.
I think that the comments section of the article will not provide us enough room and space for this.
Your information is either out-dated, or flat out biased. Why should anyone believe anything on your site when you have a huge LSI banner on every page? Have some intelligence. The 4K stripe is the worst Adaptec setting…yet, you focus on that (proving my point). Even basic info like packaging and pricing of the SSD caching solution from Adaptec (which btw, kicks LSI’s butt in all “real world” testing I’ve done)…its not a SW + SSD solution, its SW only, you can use any SSD vendor you want, and its approx a $250 uplift in price over the card…not $1000+ as you say. Anyway, be smart…test in your own environment and decide.
We particularly enjoy posts like this because it offers us the opportunity to invite the comments author to submit any test results that he has done or knows of. The simple fact is that advertising cannot ‘bend’ actual tests in any way shape or form so the comment is invalid. As the Editor, I particularly pay close attention to such and can also clearly point out that the author is an ‘Independent Evaluator’ and not staff of the site. He derives no income whatsoever through this. Looking forward to your return!
word of clarification- All tests were conducted with a 64k strip size. This is noted several times in the review.
Adjustments can be made to the strip size to favor sequential (larger) or random (smaller) access. This will also only further optimize the other controller in the testing scenario. Any gains will be felt equally (or close to) by both controllers. Also, for the purposes of the review, this would lead to many many more testing results. As stated in the review, we are not testing any and all scenarios, just what we felt would be an honest ‘baseline’ of overall performance.
We used 64k as a middle ground. This is a common size for arrays that lends itself to both random and sequential speed.
Any and all information regarding the MaxIQ caching of the Adaptec controller would be appreciated. The research that I conducted, and the information readily available on Adaptecs’ site, all pointed towards branded SSDs being used in conjunction with their SSD caching. If there is in fact an error with the information provided, I will amend the article to state the correct information. I am not trying to post inaccurate information, and apologize if that has occurred. I have sent an email personally to “Fbomb” asking for further verification of the issue, and if there is such provided, by any source, the information will be corrected.
Note that never were any implications of Adaptec SSD caching performance in comparison to LSIs caching performance, merely mention of the competing solutions. We have in fact reached out to Adaptec several times to conduct testing of their caching solution, but as of this time we have not received a reply. I was intending to do a head-to-head test of LSI and Adaptecs caching solutions as part of our CacheCade Pro 2.0 review. The offer still stands for us to test the Adaptec solution.
Finally, I receive no financial gain from this website. It is owned wholly by Les Tokar, and I receive no financial compensation. The results speak for themselves, and I stand by them. Feel free to post your results in our forum, the link is on the last page of the review. Any requests for specific testing will be accommodated, as long as we have the hardware to provide them.
I would be interested in seeing some of your MaxIQ results!
I do not think that it is fair to change the default stripe size of the Adaptec to match the one from LSI. As you can see in the review from Tom’s Hardware, the Adaptec performance close to the LSI when run with thei default stripe size. Also I have benchmarked both with 16 SAS HDDs (I am not yet using SSDs in RAIDs because of reliability issues and high price) and found that the Adpatec beats LSIs cards quite often in application benchmarks. No doubt would I use the LSI card if I would want to squeeze performance out of SSDs but I would not want to pay the premium for HDD based storage nor would I want to abstain of the flash based cache protection (too many problems with batteries in the past).
This review was configured with the absolute best possible stripe selected for the total hardware configuration. Simply put, a different stripe would not have shown more favorable for either card. Trying to base comparison between two RAID card reviews, one which utilizes hard drives and the other SSDs, is truly like comparing a circle to a square. The two reviews cannot be compared simply because the stripe is completely different for an SSD than it is for a hard drive. It has nothing to do with squeezing the performance, but rather, enabling the performance of both cards to their fullest extent.
I think this review is a bit unfair. Ok they were tested at default settings each, but if Adaptec was tested with all cache enabled (and obviously the random works is affected specially read ahead), why didn’t test LSI too with all cache enable?
Word of Clarification-
Toms conducted their tests at 64QD as well. If you look to the page of that review that shows the configuration screens, you can see their configuration in their own screenshot. They did not disclose their testing parameters, but they did provide a screenshot. 64K is a very common stripe size used in testing and evaluation of RAID controllers.
A larger stripe size on the Adaptec would paint a prettier picture for sequential throughput, but it would reduce the random performance by much larger magnitudes. In other types of testing, the controller would look even worse.
So, you have tradeoffs. For both of them.
In order for Fastpath to be enabled, one must use the write through feature. If Write Through were used with the Adaptec, the divide would have been much greater.
The settings that were used were by far the best for the Adaptec.
We conducted a long series of tests to determine the best settings, every single variation was tested.
Each controller needed to be tested to its best potential over a large range of tests.
In simple terms: there is no ‘magic’ setting for the Adaptec controller that is suddenly going to make it a racehorse.
I spent a large amount of time testing this controller with HDDs, SSDs, and EFDs. I also tested all configurations that were possible in order to try to paint the Adaptec in the best light possible.
If one were to read other results in other reviews with the Adaptec, they would see similar results.
Unfortunately, those other reviews are actually unfair to the LSI controller. Sounds weird, right? Not really. They simply do not have the devices to push the LSI to its full potential. Being a SSD specific review site, we are lucky enough to have the gear that others simply do not.
If they were to have the gear to push this hardware to its true potential, you would see the results just as plainly there as you do here.
I understood that there was going to be discord and disagreement when i posted this review, and I do appreciate the discussion.
Whenever there is such a big difference in performance there will always be questions that will be asked, but for clarification of the results, all one has to do is simply look at the other reviews of that controller and their publicly posted specifications.
The true performance of the LSI is the real story here, in other reviews no one else (that i am aware of) outside of professional validation companies (Oakgate, etc) have pushed these to their full potential and posted the results.
I encourage discussion of the testing, but would recommend that we take it to our forums. There is a link on the final page of the review.
Very well stated. Thanks Boss!
An interesting read, n comments, thank you all .
Greetings Paul and friends,
I’m new to this forum, so a short honeymoon will be most appreciated.
I for one would like to see a cost break-out for each of the discrete components you discussed in this excellent review e.g. 5 x Sandisk Enterprise SSDs @ $6,800 = $34,000 + ?? The article begins by stating that the total was $45,000, so a breakout of the remaining components would be nice to add here i.e. controllers, additional cables, external enclosure etc.
FYI: my peculiar interest in solid-state storage is more inclined to explore reasonable alternatives for small office/home office (“SOHO”) users and singular enthusiasts who crave the “most bang for the buck”. Clearly, $45,000 is far beyond the individual budgets of these often overlapping market sectors, as are the Sandisk “Enterprise” drives you reviewed.
Also, a mere 5% improvement in the performance of workstations can accumulate significant benefits for small- to medium-sized organizations that employ expensive professionals.
I honestly believe that the IT industry can do a much better job of accelerating storage across the board for the SOHO sector, as opposed to “holding back” and requiring that sector to accept the “trickle-down” effect:
https://www.supremelaw.org/systems/superspeed/RamDiskPlus.Review.htm
Several questions which I have asked of vendors like Intel and such, still remain unanswered e.g.:
(1) does TRIM work with an OS “software” RAID (i.e. where partitions are initialized as “dynamic” and either XP or Windows 7 manages RAID 0 logic)? Even though a Windows OS cannot be installed on such a RAID array, it offers a very cost-effective way to increase the performance of dedicated data partitions;
(2) is TRIM being enabled for native RAID arrays on future Intel chipsets?
(3) is it true that each SSD in a 2-member RAID 0 array will experience 50% as many WRITEs over time and thus the write endurance of each should double, and each SSD in a 4-member RAID 0 array will experience 25% as many WRITEs over time and thus the write endurance of each should quadruple?
When I tried to encourage comments concerning the effect of RAID 0 on write endurance of Nand Flash SSDs at another Internet Forum, my comments were promptly censored without any explanation(s) whatsoever. Did I touch a nerve, and if so, why?
I believe question (3) above can and should be answered definitively with proper scientific experimentation e.g. by charting Media Wear Indicators at one end of the cables, and I/O requests at the other end of the cables (cf. Windows Task Manager and the extra “I/O” columns that can be added under the “Processes” tab).
Similarly, if you read the list of “Known Issues” in the documentation for Intel’s latest Solid-State Drive Toolbox (October 2011), there you will find a rather daunting list of obstacles for any SOHO system builder or administrator (Pages 4 thru 6).
I think the SSD industry would do well to strive for the kind of Plug-and-Play compatibility that is now available with most modern rotating platter drives: I can plug a modern WDC SATA HDD into just about any motherboard or add-on controller, and it’s off and running.
Thank you very much and … KEEP UP THE GOOD WORK!
MRFS
Good day MRFS… Its good to see someone so dedicated in their thoughts and questions and, in order to facilitate the discussion, can I ask that we carry it to the forums? Start a forum thread and I believe that this will be an excellent topic of conversation.
Thanks ahead
Les
Thanks for the testing. Very imformative.
Were are in the process of testing both cards ourselves but it appears the Adaptec 6805 doesn’t support the ARC-8026 SAS expander (based on the LSI SAS2x36/28 expander) – so it’s out of the running.
You also mentioned the 9265-8i can support 240 devices. I think you may have that confused with the 9285-8e at 240 devices while the 9265-8i supports only 128 (atleast after talking directly to LSI for confirmation). That is the only difference between the models other than the 9285-8e have sff8080 external and 9265-8i having sff8097 internal.
There is also some confusing documentation that says CacheCade 2.0 Pro feature (not out yet) will only be available on the 9265-8i but one would suspect it would be available for the 9285-8e. If you have any contacts at LSI that could clear that up, that would be good to know.
thanks for noticing the inaccuracy, we have amended the article to make the correction. I will see what controllers the next version of cachecade will support. We have been assured of a testing opportunity with the next version as soon as it is ready, so stay tuned!
Costly, but now even more glad I invested in the 9265-8i in the summer of 2011. I’m now getting great speeds with 7 1TB 7200rpm drives in a RAID5, speeds of roughly 2000 MB/sec writes and 1850 MB/sec reads, posted here at:
https://tinkertry.com/webbios
I’m looking forward to playing with CacheCade 2.0 next near using a SATA3 SSD for my virtualization/backup server, complete parts list at:
https://tinkertry.com/vzilla
and method to reliably get into WebBIOS with Ctrl+H on an ASRock Z68 motherboard:
https://tinkertry.com/webbios
So glad thessdreview.com site exists, very useful info, I love to learn!
I might believe your results if you actually took the trouble to get your b****dy apostrophes correct.
I also want to chime in and state that we’ve put these cards back to back and found that the Adaptec performs far better than this review would suggest. Of course, we were using tens of 2TB drives, rather than SSDs (as well as cachecade 2.0 read/write vs MaxCache 2.0 read/write), and a 6445Q card that an Adaptec rep custom made for us (but is not really any different than the 6805Q shipping now. We’ve been exclusively LSI, with several hundred cards employed, but are designing new systems and considering a switch.
Granted, we run Linux, and driver differences can be impressive. It’s actually a shame that the tests don’t try Linux as well, since I’d wager a bet that most of these cards will end up in enterprise systems. It could be that Adaptec focuses on Linux drivers (or their Linux driver devs are better), and the performance is much better there. We won’t know though until someone does a back to back. I can understand not testing every possible config, but not testing a product on the platform it’s very likely to be used on (50/50 at least) seems like a missed opportunity.
As for our tests, as can be seen in the sequential tests in this benchmark, we actually found that the LSI card had problems sustaining heavy sequential writes (notice the huge spikes). In our tests, we exported several Fibre Channel luns from identical servers, one with an 9265, one with the 6445Q, and then on the FC client we did a RAID 1 between them and watched the IO to ensure it was the same on both sides. The server with the LSI card suffered from a dirty memory backup, to the point that the system became unresponsive in IO wait, while the one with the Adaptec card streamed the writes effortlessly. We then swapped cards between systems and the problem followed the LSI card. It just didn’t seem to be able to sustain continued heavy writes. It was so bad that I was forced to reduce the benchmark time from 5 minutes of sustained sequential to 3 minutes, because the LSI system would die before the test completed.
We otherwise didn’t see much performance variance at all between the 9265 and 6445Q, each with 20-2TB disks and two 240GB SSDs as cache. I will give you that you’re not going to notice that you have 400,000 iops of performance on 20 platters, but from other comments I gather that people are seeing better performance than the review indicates. I’m glad I found the review, but in the end it just raises more questions for me.
One thing I do like about the Adaptec MaxCache 2.0 is that it allows you to adjust the algorithm used to determine ‘hot’ data, as well as adjust read/write percentage of SSD. The LSI CacheCade 2.0 was just ‘on’ or ‘off’.
I do not see any mention of TRIM support in either of these devices, which is a very important concern with SSD usage over time. Testing shows that without TRIM support, SSDs can actually lose enough performance to dip below raid 10 arrays of spinning drives after the SSD has been fully populated and starts needing to use ERASE to clear out space for new data. Do either of these devices support trim?
There is no trim support on raid controllers. You need to have a physical connection to the OS for that.
The 9265 is great but be warned: it’s picky about motherboards. I had massive problems trying to get it to work on an Asus Rampage III Formula motherboard with an i7-990x micro. I finally had to switch over to a Supermicro X8SAX. The Supermicro was one of the few approved motherboards that supports an i7 micro. LSI doesn’t even want to talk to you if you attempt to use their card with a non-approved motherboard.
DrLechter, I’ve had a good working relationship with LSI support, with phone and email options all listed here:
https://www.lsi.com/about/contact/pages/support.aspx
I’m using a Z68 motherboard that isn’t on LSI’s list, but we got it working just fine, and LSI even points to the way I figured out to reliably get into MegaRAID WebBIOS:
https://tinkertry.com/lsi-knowledgebase-article-points-tinkertry-method-configure-lsi-raid-z68-motherboard/
Wish I had figured that out before returning my MSI motherboard, since the similar WebBIOS technique probably would have worked there too.
That said, I’m not saying LSI support could do anything about some fundamental hardware incompatibility. I’m just saying that LSI did very much talk to me, giving me reasonable support on minor issues I had, as an individual end-user of a personally purchased, 9260-8i.
Later on, I also purchased a 9265-8i and became a blogger mid 2011, and wrote about how to install drivers and get basic hardware monitoring working in the new ESXi 5.0 release, which had 9260 drivers built in, but things were more complex on the 9265, discussed here (and since fixed with ESXi Update Rollup 1):
https://tinkertry.com/lsi92658iesxi5
Most recently, recorded some video of me “kicking the tires” of an LSI based RAID5 array here:
https://tinkertry.com/raidabuse
And eagerly awaiting CacheCade Pro 2.0 for some SSD caching fun!
I make no claims to be anywhere near as technical as this great site, with far more benchmarking patience than I’ll ever have!
Very nice reply and I have been to your site reading of your testing prior. Personally, I think you might get significantly more attention here eheh
do you keep on adding SSDs all they way up to 256 or you hit a performance bottleneck along the way. Can you do a benchmark with 2,4,8,16 SSDs to check if the performance improvement is linear or not.
so it’s no way close to saturating the x8 bus.
The “x8 bus” speed varies, depending on which PCI-E version has been implemented. PCI-E 2.0 oscillates at 5 GHz and uses the legacy 8b/10b frame (10 bits per byte), for a raw bandwidth of 500 MBps per x1 lane. Thus, x8 PCI-E 2.0 lanes @ 500 MBps = 4.0 GBps raw bandwidth. The PCI-E 3.0 spec increases the clock to 8 GHz and upgrades the frame to 128b/130b “jumbo frame”, for a raw bandwidth of 1.0 GBps per x1 lane. Thus, x8 PCI-E 3.0 lanes @ 1.0 GBps = 8.0 GBps. Assuming a rough average of 500 MBps per 6G SSD, then 16 such SSDs would be needed to come close to saturating x8 lanes of the PCI-E 3.0 bus. 32 such SSDs would be needed to saturate x16 lanes e.g. RAID controller with an x16 edge connector.
can you add highpoint’s high end card to the mix as well
Why did you show RAID5 read tests? There’s no penalty for reads in RAID5. Show us RAID5 write tests. That will show the true speed of the controller. Also, why no RAID10 comparisons?
How come the 6805 performs better at Raid 5 with traditional harddrives than the 9265? Shouldn’t the fastest card be fastest at raid 5? Is it the firmware?
Thanks.
I was considering using an adaptec 6805 with 4 intel ssd 330 180 gb in raid 10. I keep wondering what will happen over the long run since there is no trim. How would the drives ever optimize.
I can’t believe there’s no mention of TRIM. This article is a waste of time for anyone planning to use SSDs in RAID long term.
Isn’t the speed of the Adaptec a bit slow? Makes me wonder why that is? perhaps because of the cables? If it were 10% slower of even 20% slow, OK, but I cannot deny that half the speed, something must be wrong with it…
I know this article is old but I’m in the market for a good RAID card si I though I’d share the following info for whoever else is (I’m surprised at the lack of recent RAID controller reviews … I can only find old ones).
Although the credibility of this article is hindered by the abundance of LSI banners on the website and by the over the top and obvious ‘LSI-fanboy’ tone of the author (more professionalism would have helped), the numbers are actually not unexpected and I’m surprised nobody mentioned this: the Adaptec 6805 specs themselves mention max 50K IOPS. That’s about it … it’s a limitation of the controller and slow RAM. So it’s not surprising at all that it can’t reach the impressive numbers that the LSI 9265 did.
However, the Adaptec 6805 is not in the same class as the LSI 9265. A better comparison nowadays would be against between the LSI 9271 (same RoC, same RAM, but PCI-e 3.0) and the Adaptec 7805 (much improved RoC, DDR3 RAM and PCI-e 3.0, just like the LSI 9271). Adaptec advertises the 7805 for more than 530K IOPS. Prices are comparable.
I’d be very interested to see a review of the two.