I joined my local org today to take away women's rights. I also protested about it with hundreds of other men of action infront of our all-American city.
I want to bring the full power of America back into the hand's of men where it belongs. I'm convinced that women are voting for crappy presidents, doubling traffic, and taking all our jobs, too. They should all have their rights taken away! What gives women the right to have real jobs and get hired over men? They are pulling us down. Women should be in the kitchen where they belong, pregnant, and taking care of the house for their superior men. It's a sad state of affairs when the women of America, The Land of the Free, the America that the White Man built from a wilderness, are getting jobs over men. Men do all the work in this great country, and women have no respect for their men, or for our precious America, God's country. The solution? Have men dominate the face of the earth. If men dominated political office and ruled this nation and the earth everything would be perfect and advancing.
I'll definitely be complaining to George W. Bush about this. I'm also going to run for USA president in the future. It's true.