Multiple flash locations in the server-storage stack are upsetting balanced I/O conventions and making overall system design much more difficult. Designing and implementing server-to-storage systems is going to become very much harder, because existing assumptions about server and storage I/Os per second (IOPS) handling are …
grmpf. seems like it makes sense to put some intelligent decision making in that. oh no wait. sun's kit already does that. Damn. what was the point of this article again?
Seeing problems where none exist...
The fact that caching can be implemented anywhere -- from L1 processor cache to a DRAM write buffer on an HDD is NOT a problem.
Caching (Flash or otherwise) needs to go where the cost/benefit is greatest, and this is a function of the application workload profile and overall system architecture.
In general, the idea that Flash should go behind an ANSI T-10 Block Device interface (as in the "solid-state disk" contradiction-in-terms) is just goofy. There is simply not enough application-related metadata at the block-device level to make any kind of intelligent cache replacement decisions. This is why (for example) approaches like Fusion-IO and Sun's FMOD approach will win in the end (because they are closer to the file-system and application metadata), while STEC and all the other "HDD form-factor" applications of Flash will die (and not soon enough).
But in any case, the idea that there should be a "standard" or "agreed upon" place to put Flash is ludicrous -- it all depoends on the application.
- NASA boffin: RIDDLE of odd BULGE FOUND on MOON is SOLVED
- SOULLESS machine-intelligence ROBOT cars to hit Blighty in 2015
- BuzzGasm! Thirteen Astonishing True Facts You Never Knew About SCREWS
- Worstall on Wednesday YES, iPhones ARE getting slower with each new release of iOS
- Microsoft's Euro cloud darkens: Redmond must let feds into foreign servers