Of course; generations of black Americans have been taught that the black is more important than the American. If you were bombarded daily by people telling you your skin color was your defining feature, you'd react the same way. Personally, if I were a black person in America, especially a male, I'd be POed as all get out at the left for ruining my culture and diminishing my person with all their racial politics.
I hear ya. I agree.