The working world, when it was male dominant, was 100x more fun to work in. Guys could "shoot-the-shit" with each other, say whatever they wanted and competitiveness was seen as fun not as "toxic". The work environment when it was male dominant made people better because the men would compete with one another and strive to improve. Work was actually a lot more "fun" and in some ways it was a "safe-space" for men to get away from all the drama and toxicity that existed in their personal lives.
Women have utterly destroyed the workplace. Men's lives are inarguably much worse with women in the workplace. Nothing good has come of this.
I was in a meeting about how to be promoted. The white 40+yr old women presenting told us "we're looking to promote people... gestures to room of white men not like you". The black women they promoted spend all say on slack complaining about how theyre oppressed. Only a small group of white men and east asians keep the company running.
Why even bother? Do all the work and not even get matching compensation.
Most at this point do the minimum required or move to Indian companies that hire them into high management positions as white monkeys.
Pretty much every industry has been coopted and they're all held together by the few remaining core employees who do the vast majority of the work.
We're set for another global bubble burst as soon as people in these companies realize that they can just leave.