A majority of Americans believe that the United States was founded as a Christian nation, and many American leaders have promoted this idea to argue that the federal government should privilege Christianity over other religions. Together, we'll examine the historical evidence for this claim and assess its social, cultural, and political implications for our society today.