Apparently women think that girls are taught to "smile and hide their feelings/troubles". What the fuck kind of reality do they live in?
I've noticed this kind of comment many times in all kinds of communities. It's honestly quite puzzling to me. How the fuck do their brains do this kind of mental gymnastics, and moreso, how the fuck do most normies believe this shit?
Men are undoubtedly told to shut the fuck up and man up. Toughen up, take it like a man. You show emotion as a man and you've got a target on your back. Everybody will mock you or even bully you.
Women are coddled, pampered, helped, listened to, assisted in every imaginable way. Girls can cry and complain and show any kind of emotion and they'll be praised and supported.
What the fuck, how are people so brainwashed?