After World War II, the story goes, the United States parted ways with its isolationist past and asserted itself as a political and military power.[1] Recently, though, historians and political scientists have begun to question this narrative, concluding that the United States sought to avoid political and military commitments to Europe for much longer than…