i know that America is a big huge place with lots of people living everywhere spread out in different environments, locations, areas and situations etc.but are white Americans somewhat aware of white decline and becoming a minority in the future? does it bother them? or aren't they that bothered by it and believe the whole "whites stole America from the natives and made the natives a minority so it doesn't matter if we become a minority ourselves" guilt narrative pushed on them?
The purpose of World War II and its resulting world order was to destroy white nations and establish Pax Judaica.