Christianity is almost the only one of the great religions which thoroughly approves of the body—which believes that matter is good, that God himself once took on a human body, that some kind of body is going to be given to us even in heaven and is going to be an essential part of our happiness.