Best Picture is more than just an award: it's a reflection of where Hollywood — and America — wants to be.