In Hollywood Left and Right, Steven J. Ross tells a story that has escaped public attention: the emergence of Hollywood as a vital center of political life and the important role that movie stars have played in shaping the course of American politics. Ever since the film industry relocated to Hollywood early in the twentieth century, it has had an outsized influence on American politics. Through compelling larger-than-life figures in American...