The history taught at U.S. universities must have changed drastically since I attended school in the 1960s. I was never taught this "PC" history. For instance, it was rarely if ever mentioned that many of our great and glorious leaders were slaveowners--that they held other human beings as their property, like farm animals. (If we had not been slaveowners ourselves, how would our history books have described this uncivilzed behavior in other societies?) Our history never spoke of Columbus's genocide, never questioned later U.S. interference in South America, Woodrow Wilson's racism, or anything remotely negative about the U.S. As I stated in a previous post, my senior class could not think of any immoral act which the U.S. had committed in its entire history. What mistakes were we supposed to learn from? Paul