Multiple flash locations in the server-storage stack are upsetting balanced I/O conventions and making overall system design much more difficult. Designing and implementing server-to-storage systems is going to become very much harder, because existing assumptions about server and storage I/Os per second (IOPS) handling are …
grmpf. seems like it makes sense to put some intelligent decision making in that. oh no wait. sun's kit already does that. Damn. what was the point of this article again?
Seeing problems where none exist...
The fact that caching can be implemented anywhere -- from L1 processor cache to a DRAM write buffer on an HDD is NOT a problem.
Caching (Flash or otherwise) needs to go where the cost/benefit is greatest, and this is a function of the application workload profile and overall system architecture.
In general, the idea that Flash should go behind an ANSI T-10 Block Device interface (as in the "solid-state disk" contradiction-in-terms) is just goofy. There is simply not enough application-related metadata at the block-device level to make any kind of intelligent cache replacement decisions. This is why (for example) approaches like Fusion-IO and Sun's FMOD approach will win in the end (because they are closer to the file-system and application metadata), while STEC and all the other "HDD form-factor" applications of Flash will die (and not soon enough).
But in any case, the idea that there should be a "standard" or "agreed upon" place to put Flash is ludicrous -- it all depoends on the application.
- It's true, the START MENU is coming BACK to Windows 8, hiss sources
- iSPY: Apple Stores switch on iBeacon phone sniff spy system
- Pic NASA Mars tank Curiosity rolls on old WET PATCH, sighs, sniffs for life signs
- How UK air traffic control system was caught asleep on the job
- Google embiggens its fat vid pipe Chromecast with TEN new supported apps