The working world, when it was male dominant, was 100x more fun to work in. Guys could "shoot-the-shit" with each other, say whatever they wanted and competitiveness was seen as fun not as "toxic". The work environment when it was male dominant made people better because the men would compete with one another and strive to improve. Work was actually a lot more "fun" and in some ways it was a "safe-space" for men to get away from all the drama and toxicity that existed in their personal lives.
Women have utterly destroyed the workplace. Men's lives are inarguably much worse with women in the workplace. Nothing good has come of this.
Anything that was male dominated... now becomes mediocre or trash when women get involved. There was a kotaku article recently shitting on nintendo japan for having only 5% women in management positions lol.
And guess what?? Nintendo is profitable and respectable. Then you look at microsoft and sony (especially sony usa/cali) it becomes more and more shit filled with woke shit. Any company that promotes diversity of its staff or product, instead of just the good quality is most likely dog shit.