"It's just lip service. Hollywood sells violence against women as entertainment on practically every drama on network television," writes Kristine Marsh, who covers cultural issues for the watchdog group. "News outlets criticize the Right for alleged attacks on women. In reality, liberal Hollywood is home for incredibly violent and shocking depictions of violence against women."
"The results were especially alarming, given Hollywood's attempts to curtail violence against women," says Ms. Marsh.