The US Constitution doesn’t mention Christianity or any specific religion. The Declaration of Independence famously proclaims that people’s rights come from a “Creator” and “Nature’s God”—but doesn’t specify who that is. Yet large numbers of Americans believe the founders intended the US to be a Christian nation, and many believe…