Confession: I love watching the west fall apart, the suicide rate rise and normie males getting WRECKED by life
I know what most men are going through these days and how you're basically untouchable scum if you're male. Or how the economy is fucked and most men can barely make enough money to survive.
But I also hate that most men deny the blackpill, worship women, and refuse to admit how bad things are.
In that case, I hope society burns to the fucking ground with normie males to be used as fuel for the fire. Why should anyone give a fuck about men who just get in your way and make everything exponentially worse by giving women more power?