The working world, when it was male dominant, was 100x more fun to work in. Guys could "shoot-the-shit" with each other, say whatever they wanted and competitiveness was seen as fun not as "toxic". The work environment when it was male dominant made people better because the men would compete with one another and strive to improve. Work was actually a lot more "fun" and in some ways it was a "safe-space" for men to get away from all the drama and toxicity that existed in their personal lives.
Women have utterly destroyed the workplace. Men's lives are inarguably much worse with women in the workplace. Nothing good has come of this.
I think it's being done artificially. Your average woman doesn't care, and I would guess would want men to have their own spaces. It's mostly globalists and one of their tentacles (feminism) that's pushing for the destruction of male only spaces. Why? It weakens men, the gender that's most likely to see through the globalist lies and fight back against them. That's why men are denigrated in every area the globalists control.
No, women have done this shit for centuries.
Only when a society has become uber successful, like ours has. Success breeds abundance, safety, and technological progress which puts up barriers between people and nature. Nature culls stupid individuals and stupid practices. As stupidity propagates, the society collapses, hardship reigns again, the barriers to nature disappear, and the hard lessons are relearned. Feminism, and other stupid practices, cease to exist in hardship, and in a SHTF scenario women cease to be strong independent women immediately, and want men to protect and provide for them.
However, this modern push for feminism is absolutely unnatural. Some indoctrined bitter useful idiot women are involved, yes, but it's mostly pushed by malicious forces that use feminism to weaken us and gain power.