I am going to state an outrageous statement but it is controversial but relevant: The woman’s right to vote killed more Americans through war and abortion and pretty much systematically ruined the country culturally. This is historically proven. I apologize to true conservative women. This does not mean the right to vote was to be denied to anyone.
As a woman, I dont disagree.
it also gave us Prohibition
You'll get no argument from me.
Why Women DESTROY NATIONS*/CIVILIZATIONS - and other UNCOMFORTABLE TRUTHS