Do College Degrees Mean Less Disease?

We know that a college degree can often help ensure employment, creating pathways to better opportunities and resources in someone’s career and even one’s personal health. A recent article in The Washington Post shows that the health benefits of higher education are more nuanced than scholars originally believed. Drawing from the work of sociologists Andrew … Continue reading Do College Degrees Mean Less Disease?