I hate what feminism has done to this country
I hate being called misogynistic because I don’t conform to the notion of a “rape culture”
I hate when men are always seen as aggressors in encounters between men and women
I hate when it is assumed that I am being nice to a girl just because I want her body.
I hate it when women can claim all of this crap about “male privilege” when they have much more actual privileges than men
I hate that women can opt out of parenthood, but a man can not.
I hate that women are always seen as victims in almost every situation.
I hate that actual issues that almost exclusively affect men and boys such as higher suicide rates, higher work place death rates, higher chance of developing a learning disability, lower chance of getting into college and graduate school, false rape allegations, custody disputes, child support, alimony and many other things are overlooked and anyone who brings this up is automatically labeled as misogynist,
I hate what feminism has done to our society.
The hubris, stubbornness, and stupidity of most young woman today is the sole reason I will never ever ever have kids.
I grew up with two sisters (both are smart, good looking and healthy); and the way they absolve themselves of any sort of responsibility or fault without self-reflection has fundamentally changed the way I view human nature. Feminism needs to be dismantled and society needs to address female nature by raising daughters to be HUMBLE and THOUGHTFUL.
I hear plenty of stories about older woman learning to be humble and thoughtful after they've "had their fun", but it's often too late at that age as they've already pumped toxic femininity into the world.