Body Image

Healing the wounds diet culture has inflicted on your body image and learning to see yourself in a new way.