Why does western society frown on self harm? In many cultures around the world self harm and cutting are celebrated and encouraged, yet in western society if you cut yourself you are treated like a freak. I don't understand what the big deal is. I personally enjoy cutting as I have a self-destructive nature but I am ashamed to do it because of how people will react if they see my scars... Cutting myself doesn't make me feel ashamed, just other people's reactions. Also I love skin carvings and how they can be combined with ink to make some fantastic body art. I fully intend to get a skin carving but it's illegal (I guess because of health reasons) Still, I'm gonna get one anyway and people will probably look at me with disgust when they see it but whatever, I don't care what they think.