Women are raised to know the truth about men
When a girl is growing up, everyone in their life tells them how men are, they tell them about how men use women for sex, r*pe, abuse "they all say anything to get in your pants", "they only care about looks", blah blah we all know the story. Men on the other hand dont get this message, we're told that women are perfect angels, that we should respect them no matter what. We never get the talk about female sexual behavior, the games they play etc, we have to learn this by ourselves and get burned multiple times in the process. A lot of our anger comes from being absolutely blindsided by true female nature. Teach boys that women are cunts, the same way you teach girls that men are shit.