What do people mean when they talk about the "WESTERNIZATION OF WOMEN"? And is it a good thing or a bad thing?I've only heard people bring this subject up on college campuses though, so maybe it's not that big of a deal.
Western thought and civilization is the purest, highest, and most correct way of thinking possible. Racist freaks who hate White people, love wearing tiny hats, and have a strange sexual desire for children have been trying for decades to change the meaning of western culture to instead mean gay buttsex and skin the color of feces. You would need to consider context, but it's probably the perverted wrong meaning they are trying to express here.