is it still considered a white country?
Is it still considered a white country?
As a country, no
it never was one
Post white americans
If you remove NYC, Louisiana, Florida, Texas, and California it is white
Outside the big states pretty much yeah.
9/10 white people who live here would say so
>It's so sad what's happened to Britain recently. Sharia no go zones, ugh, when will the Brits get their balls back?
ABSOLUTELY NOT
>remove this and that and it becomes wh*te I swear!!
The absolute state of Muttland.
Most of the non-whites are mexican
The white race is a meme
>tfw exclusively date brown girls
Essentially if you remove all the states with a lot of cotton fields, and remove the states bordering Mexico, America is 95% white
i'm actually white and no, it's not white at all. everyone here is some sort of brown subhuman mutt. i need to leave but having a hard time bringing myself to do so
Don't let the door hit you on the way out.
At least we are still culturally white
the white race is a meme, you are an anglo
No
Europr invaded by nafris and is le 20% the whites are leftist anti fascists faggots
USA is 200 million whites all stonecold racists that a national socialist society can literally rule.
You tell me.
There is nowhere to flee to goyim, Europe is under their rule.
they flooded it with migrants for that reason, so there would be nowhere left to run to
25% Jewish
25% Sicilian
25% Negro
25% Mexican
Mohammad.
At least our people our whites will recognize our enemy.
They are exterminating Europe, genociding it and flooding it with arabs and africans.
France is completely colonized and conquered, the black pill.
Kek America is more like 20% wh*Te