Why must there be “diversity and inclusion” in everything done by White people? There are plenty of “Black” organizations that don’t have “diversity and inclusion”. It sounds to me like a feel-good way of weakening White America. It also amazes me that any CEO or business exec falls in line with this nonsense without a peep of objection.
Think about it:
Govt bureaucrat: “You can’t run your business or build your widgets unless you hire some blacks and trannys.”
Businessman: “I can’t waste time searching for qualified employees that share those characteristics, I need qualified skilled people.”
Govt bureaucrat: “We will tell you what qualifies as a skilled and qualified employee.”
Why must there be “diversity and inclusion” in everything done by White people?