White Man: Why Are You Giving Away Your Country?

From amren.com:

A Hispanic student wants to know.
As a child in public schools and now as a graduate student in history, I have learned one thing to be true about the United States: It is a white country. The founders were white, white men established its core principles and political system, and white men and women built the nation into what it is today.

Even before I became aware of the importance of race, I never thought this was “racist” or unfair; it was simply a fact. America has always been a white country and always should be. Why, then, are white people giving away their country?

http://www.amren.com/features/2015/04/white-man-why-are-you-giving-away-your-country/#