The working world, when it was male dominant, was 100x more fun to work in. Guys could "shoot-the-shit" with each other, say whatever they wanted and competitiveness was seen as fun not as "toxic". The work environment when it was male dominant made people better because the men would compete with one another and strive to improve. Work was actually a lot more "fun" and in some ways it was a "safe-space" for men to get away from all the drama and toxicity that existed in their personal lives.
Women have utterly destroyed the workplace. Men's lives are inarguably much worse with women in the workplace. Nothing good has come of this.
Facts. I used to be a mortgage broker. We had an all-male staff for the first two years I worked there and we had a fucking blast. Then they hired two women in the same month and the vibe changed completely.
The first week that the first women was there, I was in my office speaking with a client and I said, "Great, I'll have my secretary prepare the forms and send them your way before end of business."
The next morning I had an email in my inbox from our new female co-worker. It was almost two pages worth of text. She was livid lol and she scolded me for referring to her as, "my secretary." She told me she isn't my secretary and that it was rude, inconsiderate, blah blah blah.
I told her I wasn't referring to her. I had been using that line (which was a lie) for two years before we had any female staff. I used that line to make client's think we were a bigger shop than we really were.
She refused to believe me and she filed a complaint with the owner haha. The owner knew I'd been using that line all along and sided with me, but the cunt wouldn't let it go and she was always a huge bitch to me until she quit a few months later.
Jfc...