Wellness Versus Diet Culture: How Do I Know What’s What?
Diet culture has long ruled over the fitness and health industry. So, how do you know which companies care about self-love, body positivity and acceptance and which ones are just trying to profit by making you feel unhealthy? Look for these telltale signs to determine if a company embraces wellness or if it’s rooted in diet culture.