A Hispanic student wants to know.
As a child in public schools and now as a graduate student in history, I have learned one thing to be true about the United States: It is a white country. The founders were white, white men established its core principles and political system, and white men and women built the nation into what it is today.
Even before I became aware of the importance of race, I never thought this was “racist” or unfair; it was simply a fact. America has always been a white country and always should be. Why, then, are white people giving away their country?