Jenny McCarthy says Hollywood 'gets everything wrong' about real America after moving to Midwest

Jenny McCarthy slams Hollywood as out of touch with real America, explaining why she moved back to the Midwest to raise her son away from the industry.