Survey data and anecdotal evidence agree that Christianity is losing its hold on American life. The Roman Catholic hierarchy struggles to regain its credibility following the pedophilia scandals, mainline Protestantism wrestles with issues of sexual identity and attrition, and white evangelicalism has merged with the far-right precincts of the Republican Party. Moral authority, it seems, is hard to come by, with all three of the major Christian...