The working world, when it was male dominant, was 100x more fun to work in. Guys could "shoot-the-shit" with each other, say whatever they wanted and competitiveness was seen as fun not as "toxic". The work environment when it was male dominant made people better because the men would compete with one another and strive to improve. Work was actually a lot more "fun" and in some ways it was a "safe-space" for men to get away from all the drama and toxicity that existed in their personal lives.
Women have utterly destroyed the workplace. Men's lives are inarguably much worse with women in the workplace. Nothing good has come of this.
Women wanted freedom from being a slave in the home, looking after children and husband. Instead they wanted to be a slave in the workplace.
They got what they wanted and yet they still complain.
They don’t even know what they want, much less what is good for them. It’s better not to listen to them, and just do what is right.
I've seen this posted a lot on 4chan, and it definitely cuts to the heart of the bullshit of the feminist message of women in the workforce.
Women working for her husband and children: slavery and patriarchy Women working for a company: empowered and free