“Addictive” social media feeds that keep children online targeted by New York lawmakers
New York would restrict the way online platforms like Instagram and YouTube can collect and share children’s personal information and let parents keep their kids from being bombarded by “addictive” feeds from accounts they don’t follow, under legislation proposed Wednesday.
The bills offered by state leaders are aimed at protecting young people from features designed to keep them endlessly scrolling, endangering their mental health and development, Attorney General Letitia James said.
“Young New Yorkers are struggling with record levels of anxiety and depression, and social media companies that use addictive features to keep minors on their platforms longer are largely to blame,” James said. “This legislation will help tackle the risks of social media affecting our children and protect their privacy.”
The regulations sought by James and Gov. Kathy Hochul, both Democrats, are similar to rules already in place in Europe, where violations could incur fines worth a percentage of revenue, which could run into the billions of dollars for wealthy tech companies.
Neither Meta, which owns Facebook and Instagram, nor YouTube immediately responded to requests for comment from those two California-based companies.
One of the bills, the Stop Addictive Feeds Exploitation (SAFE) for Kids Act, would would allow parents to opt their kids out of getting feeds curated by an algorithm. Instead it would have them get a chronological feed of content from users they already follow. Algorithms are the automated systems that social media platforms use to keep users engaged by suggesting content based on the groups, friends, topics and headlines a user has clicked on in the past.
The legislation also would let users block access to social media platforms from midnight to 6 a.m. and limit the hours a child spends on a site.
The second bill, the New York Child Data Protection Act, would prohibit all online sites from collecting, using, sharing, or selling personal data of anyone under 18 years old, unless they receive informed consent or it’s otherwise necessary.
Companies could be fined $5,000 per violation of either law.
Under new digital rules that came into force this year across the 27-nation European Union, platforms have to give users an alternative to automated systems that recommend videos and posts based on their profiles. Thus Meta, for example, now also allows European users to see chronological Facebook and Instagram posts only from people they follow.
The rules, known as the Digital Services Act, also prohibit platforms from using children’s data and online activity to target them with personalized ads.
Another set of rules, the General Data Protection Regulation, or GDPR, provide beefed-up data safeguards and rights for EU residents. Regulators slapped TikTok with a $366 million fine last month for breaching GDPR by failing to protect children’s privacy.
The legislation in New York also follows actions taken by other U.S. states this year to curb social media use among children. In March, Utah became the first state to pass laws that require minors to get parental consent before using social media. The laws also compel companies to verify the ages of all their Utah users, impose a digital curfew for people under 18 and ban ads from being promoted to minors. But experts have noted the new rules, which are set to take effect next year, could be difficult to enforce.
Meanwhile, another state law in Arkansas that would have also required parental consent for children to create social media accounts was put on hold by a federal judge in August.
Aside from issuing new laws, some states also have been taking social media companies to court over a host of issues, including their algorithms and data collection practices. This week, Utah filed a lawsuit against TikTok, alleging the app’s addictive algorithm is damaging minors. Arkansas is also suing TikTok and Meta, which owns Facebook and Instagram. Indiana sued TikTok last year claiming the Chinese-owned app misleads users about the level of inappropriate content on the app and the security of their information, but the challenge doesn’t appear to be going in the state’s favor.
The U.S. Supreme Court is preparing to decide whether state attempts to regulate social media platforms violate the Constitution. The justices will review two laws from Florida and Texas that mostly aim to prevent social platforms from censoring users based on their viewpoints.
Disclaimer: The copyright of this article belongs to the original author. Reposting this article is solely for the purpose of information dissemination and does not constitute any investment advice. If there is any infringement, please contact us immediately. We will make corrections or deletions as necessary. Thank you.