I have been seeing alot of grumbling about Christians and complaining that Christian moral activism is returning.
Name a single part of social justice/woke culture that did not result from what Christians were warning yall about.
I guess some lessons will never be learned though. Society has just gone too far down the path of liberalism 🤷♂️
The problem wasn't usually with the idea of what they were preaching, it was the nonsensical method they went about doing so. Satanists and other degenerates were everywhere, but accusing literally everything you don't like of being so just makes nobody take your seriously.
The Right is filled with "right about it, but failing to convince anyone" like that, which is why its always losing. Just look at most people here who will screaming about Jews or women or niggers and be completely right but are so unlikable in doing so that you still don't want to deal with them.
The Christian moral activism specifically is stupid because its being propagated by people who are so dirty and inundated with sin they have no room to even be talking about other people. The tatted up whore that keeps getting spammed around here is a great example of someone whose own house isn't in order enough to lecture anyone yet.
You can still reject bad actors with retarded takes, without rejecting the point they are making.