The Chinese-owned app which is wildly popular with young smartphone users, also said it took down nearly 62 million videos in the first quarter for violating community standards — including for “hateful” content, nudity, harassment, or safety for minors.
In its first disclosure on underage users, TikTok said it uses a variety of methods, including a safety moderation team, that monitors accounts where users are suspected of being untruthful about their age.
Those age 12 or younger are directed to “TikTok for Younger Users” in the United States.
TikTok, owned by China-based ByteDance, is believed to have one billion users worldwide including more than 100 million in the United States.
Last month, the Biden administration reversed orders from former president Donald Trump which would have banned TikTok or forced its sale to American investors.
The report comes with social media operators facing increased pressure to remove abusive and hateful content while remaining open to a variety of viewpoints.
TikTok’s transparency report said that in addition to the suspected underage users, accounts from nearly four million users additional were deleted for violating the app’s guidelines.
“Our TikTok team of policy, operations, safety, and security experts work together to develop equitable policies that can be consistently enforced,” the report said.
“Our policies do take into account a diverse range of feedback we gather from external experts in digital safety and human rights, and we are mindful of the local cultures in the markets we serve.”
TikTok said its automated systems detect and remove the vast majority of offending content: “We identified and removed 91.3 percent before a user reported them, 81.8 percent before they received any views, and 93.1 percent within 24 hours of being posted.”
Overall, fewer than one percent of the videos uploaded on TikTok was taken down for violations.