For the first time, internal TikTok communications have been made public that show a company unconcerned with the harms the app poses for American teenagers. This is despite its own research validating many child safety concerns.

The confidential material was part of a more than two-year investigation into TikTok by 14 attorneys general that led to state officials suing the company on Tuesday. The lawsuit alleges that TikTok was designed with the express intention of addicting young people to the app. The states argue the multi-billion-dollar company deceived the public about the risks.

In each of the separate lawsuits state regulators filed, dozens of internal communications, documents and research data were redacted — blacked-out from public view — since authorities entered into confidentiality agreements with TikTok.

But in one of the lawsuits, filed by the Kentucky Attorney General’s Office, the redactions were faulty. This was revealed when Kentucky Public Radio copied-and-pasted excerpts of the redacted material, bringing to light some 30 pages of documents that had been kept secret.

[…]

TikTok’s own research states that “compulsive usage correlates with a slew of negative mental health effects like loss of analytical skills, memory formation, contextual thinking, conversational depth, empathy, and increased anxiety,” according to the suit.

In addition, the documents show that TikTok was aware that “compulsive usage also interferes with essential personal responsibilities like sufficient sleep, work/school responsibilities, and connecting with loved ones.”

TikTok: Time-limit tool aimed at ‘improving public trust,’ not limiting app use

The unredacted documents show that TikTok employees were aware that too much time spent by teens on social media can be harmful to their mental health. The consensus among academics is that they recommend one hour or less of social media usage per day.

The app lets parents place time limits on their kids’ usage that range from 40 minutes to two hours per day. TikTok created a tool that set the default time prompt at 60 minutes per day.

[…]

    • averyminya@beehaw.org
      link
      fedilink
      arrow-up
      15
      ·
      1 month ago

      Or we could, you know, force social media companies to not use psychologists to make their apps more addictive by design. Something called ethics.

      It’s extremely telling that you can look for a job as a psychologist for Meta and all the opportunities that are available are UX researchers.

    • Kissaki@beehaw.org
      link
      fedilink
      English
      arrow-up
      4
      ·
      1 month ago

      Can you explain what you mean by free speech?

      Is the free choice of content selection algorithm free speech? Isn’t the speech, the content, there either way, and could be selected through other alternative algorithms?

      Is using deliberately engaging or addicting design free speech? Isn’t the speech, the content, there either way?

    • Steve
      link
      fedilink
      English
      arrow-up
      4
      ·
      edit-2
      1 month ago

      You’re conflating free speech of individuals, with engagement driven black box recommendation algorithms of corporations. It’s a common mistake. I think most people make it.

      A company can allow people to post things, and for people to see them if they like, without algorithmically pushing it in endless scrolling interfaces.

      For example Lemmy and Mastodon. You only see what you choose to subscribe to. The sites don’t chose to push any content into your feed because an algorithm thinks you’ll like it.

      There is a big difference between the two.
      And removing the algorithms isn’t a hindrance to free speech, only profits.