
Over the past few years, something about our culture has changed. We all sense it, though we can’t always put it into words. Some say we’ve moved from a world that’s neutral toward Christianity to a more “negative world.” Others say we’re living through the slow death of Enlightenment liberalism. Still others assert that Western society is now fully post-Christian. Whatever you call it, what we...