The working world, when it was male dominant, was 100x more fun to work in. Guys could "shoot-the-shit" with each other, say whatever they wanted and competitiveness was seen as fun not as "toxic". The work environment when it was male dominant made people better because the men would compete with one another and strive to improve. Work was actually a lot more "fun" and in some ways it was a "safe-space" for men to get away from all the drama and toxicity that existed in their personal lives.
Women have utterly destroyed the workplace. Men's lives are inarguably much worse with women in the workplace. Nothing good has come of this.
Every self reported study done on women, be it from college, hs, work etc. shows that women believe they work harder than men to achieve the same results. They argue it is because of bias, which is laughable. Society has conditioned women to believe they are equal in all capacities that men are, when reality hits they claim bias.
I don't give a shit what any self-reported study shows. The only information you can ever get from a self-reported study is the bias of the people doing the reporting. All that study shows me is that women think their jobs are supposed to be easier and don't want to admit that men have always had to work harder than they realized and are just used to it.
I thought this is actually basically what the other poster is saying.
I'm agreeing with him