Hi. There's been a lot of talk about Hollywood going woke and thus going broke. But is this really true? Is Hollywood, perhaps, pretty un-woke?