I'm just a lay-astrophotographer. I have a degree in psychology and mathematics, for whatever that's worth (not much if you ask my paycheck). I'd just like to point out that there are some pretty huge margins of error in Astronomy. Margins unacceptable in even the world of "soft science". For example, if we attempt to use parallax to measure the distance to Betelgeuse, the distance is 643 ± 146 ly. That's ± 25% (approximately but who cares)!!! Perhaps I should have gotten a degree in astronomy, but, I am at a loss for how we can, in any reasonable or scientific way, measure the distance to stars and galaxies 100's of millions of ly away, when we can't even reasonably measure the distance to ones in our own back yard... let alone understand the beginning of the universe. Blah, blah, red shift, blah blah, Hubble was a genius, blah blah, but what is the base measurement used to derive distance based on red-shift? For me, it takes as much faith to believe much of the astronomy jargon as it does to simply believe a cosmic Jewish zombie's father spoke and, POOF... we're all here a couple thousand years ago.