What's happening in our culture? Why is society turning against Christians and their beliefs? Where is this headed? And what should we do about it?