Women have quietly taken their places at the top of the Hollywood movie business. “Though men still figure most prominently in the corporate echelons of the media companies that own the studios, and talent agencies like William Morris and Creative Artists Agency are still male dominated, these women, who over the years have fought and fostered one another as part of a loose sisterhood, have finally buried the notion that Hollywood is a man’s world. So striking is the change that some now see Hollywood as a gender-balanced model for the rest of corporate America.”