The working world, when it was male dominant, was 100x more fun to work in. Guys could "shoot-the-shit" with each other, say whatever they wanted and competitiveness was seen as fun not as "toxic". The work environment when it was male dominant made people better because the men would compete with one another and strive to improve. Work was actually a lot more "fun" and in some ways it was a "safe-space" for men to get away from all the drama and toxicity that existed in their personal lives.
Women have utterly destroyed the workplace. Men's lives are inarguably much worse with women in the workplace. Nothing good has come of this.
Your post is right but,
Us men today didn't have a part with this but this is fully the fault of the men in our fathers', grandfathers' and great-grandfathers' generations. Like Adam did with Eve they completely stood by and either allowed or encouraged it to happen. Women are in the workplace today because men allowed it.
They arguably had no idea of the extent and scale of unforeseen consequences that resulted from this, but they still had a part to play in it. To be fair, I didn't realize this until the last few years, myself.
This is, of course, impotent whinging. Men really let things devolve in a bad way and I'm don't think things could get ever get back on the right track besides a collapse or something else horrible causing a societal reset.
I don’t see it as whining. We do have to recognize what happened in order to fix it. I see men blaming women for the problems between the sexes so that shows we have a long way to go in terms of understanding the cause.
All we can do is be honest and try our best to provide correction where we see error. Society will get better or worse as it will. It’s not up to us. I can only do my best to improve my own life by adhering to God’s will as much as possible.
Spot on. You get what you tolerate.