I spent 15+ years supporting and managing enterprise storage at major Storage vendors and I can only blame myself for it .... In my experience few outages are truly the result of bugs inherent to the storage system or hardware fault.
Most often customers drive the system into the ground and the reasons can be:
- no performance and capacity monitoring in place and no idea what the original specifications where
- the person that sized and architected the system has long buggered off and is now doing something with cloud, devops or IoT - or is trying to get a job with a vendor
- the storage admin hasn't been afforded any training on the product
- the budget is tight and the required upgrade is on hold
- we've bought the upgrade and we want to "configure it ourselves"
Long story short - the "enterprise, solution or storage - architect" are teflon covered and won't accept responsibility.
For the slightest glitch the storage admin will log a ticket with the vendor as a matter of arse covering.
With any luck the vendor will hang themselves.
The vendor SE and Sales rep have moved on or won't accept responsibility. Customer management refuse to tae ownership and "lean on" storage admin to sort it out. Storage admin "leans on" vendor support to sort it out.
It eventually gets all to hard. Customer writes a cheque for a system that's fit for purpose.
Problems stop for 24 months and the process starts over. Maybe this time we go to the cloud ? :)
Vendor sales rep takes Its manager to overseas executive briefing centre. Next they buy another system.