Highsnobiety
Double Tap to Zoom

In the past decade alone, social media has ruled our world. The vast majority of us spend every morning scrolling through Instagram, reading a couple of Tweets, and if you're old, checking Facebook, and it goes without saying that it has had a major impact on our mental health and general wellbeing.

As Facebook has faced some difficulties with its rebrand to Meta, it isn't hard to understand why. As a platform, Facebook let misinformation and harmful content spread like wildfire, and as a result, it gave birth to some pretty horrible groups. When Facebook then purchased Instagram, its reluctance to address mental health and other issues has resulted in people leaving the platform. Now, it is trying to address sexual harassment in the Metaverse – but how was that even able to exist in the first place?

In the past few years, Instagram has gone from a photo-sharing social media app, to becoming a money-making platform for influencers and talent all over the world. All of a sudden, we're seeing overly retouched images, unrealistic body standards, and careless promotions of weight loss supplements, vitamins, face-altering filters, and so on, without any warning or regulation.

Recently, Instagram made an effort in controlling the content we see on the app by labeling when something has been altered or posted with a filter, as well as making tagging branded content a rule. But, it was a little too little, a little too late.

Now, TikTok has announced updates to its community guidelines, which aim to protect users from harmful content.

The new policy updates include: "Strengthening our dangerous acts and challenges policy, Broadening our approach to eating disorders, Adding clarity on the types of hateful ideologies prohibited on our platform, Expanding our policy to protect the security, integrity, availability, and reliability of our platform."

Unlike other social media platforms, TikTok has been holding itself accountable and aims to be transparent when it comes to the work it is doing surrounding community guidelines. Quarterly, the platform releases a "Community Guidelines Enforcement Report," with its most recent report showcasing that 91 million videos that violated its policy were removed in Q3 of 2021 – which is 1% of the platform's videos. "Of those videos, 95.1% were removed before a user reported it, 88.8% before the video received any views, and 93.9% within 24 hours of being posted."

TikTok is far from perfect, but at least it is taking important steps to becoming less harmful for users and doesn't shy away from tackling issues head-on – something that plenty of other platforms struggle with.

At the end of the day, social media will never be good for you, but at least it can improve and make us feel less impacted as a result. You can read TikTok's full outline here, and learn more about the actions taken by the platform.

We Recommend
  • How to Make an Iconic Nike Air Max Better? Go Premium
  • Trust a Skate Brand To Make a Workwear Hoodie This Fucking Awesome
  • Can You Make a Parisian Souvenir Shop Cool? Pigalle Says, Yes
  • Is King James’ Richard Mille Really Fit for a King?
  • Does a Brand Make Culture, or Does the Culture Make a Brand?
What To Read Next
  • Botanical Ballet Flat or Nike's Tabi-Toed Trail Shoe?
  • Retuning a Trail Shoe Supercar (EXCLUSIVE)
  • Footwear so Beyond Flat That It's Basically Barefoot
  • This Is Not the Same Stylish adidas Samba, But It Is Stylish
  • Action Bronson Crowns the King of All Desert Watches
  • Kith’s Stacked Runway Show Literally Stopped Traffic (EXCLUSIVE)