I have noticed white women, women in general, but white women seem to believe themselves to be in a position of righteousness and moral authority. They believe themselves to be the ones everyone should turn to because they know it better. Black women tend to do the same, but are more violent about it.
I have noticed white women, women in general, but white women seem to believe themselves to be in a position of righteousness and moral authority. They believe themselves to be the ones everyone should turn to because they know it better. Black women tend to do the same, but are more violent about it.