Am I naive?
Or is it silly that software is still being written in a way that makes buffer overflows possible? It seems like something that should be guarded against at a low level. No doubt there would be a performance cost, but wouldn't it be worth it to stamp out an entire class of severe security holes?