How War Changes America
Such bullishness in the immediate aftermath of war is far more characteristic of American history than the Vietnam syndrome was, mainly because the United States usually won its wars. Though Americans like to think of war as an aberration from the peaceful flow of their history, it has in fact played an integral role in the economic and political development of the nation. Military conflicts have created “emergency” conditions in which social and economic arrangements have been reshaped or scrapped....