Kids Online Safety Act: Difference between revisions

From Bibliotheca Anonoma
No edit summary
Line 1: Line 1:
In 1997, the Supreme Court of the United States unanimously struck down two provisions of the Communications Decency Act, which made "the knowing transmission of obscene or indecent messages to any recipient under 18 years of age" and "knowing[ly] sending or displaying of patently offensive messages in a manner that is available to a person under 18 years of age" felonies<ref>[https://www.supremecourt.gov/opinions/boundvolumes/521bv.pdf 521 U.S. 844, 859]</ref>. "The level of discourse reaching a mailbox simply cannot be limited to that which would be suitable for a sandbox", quoted the Court; "the Government may not 'reduce the adult population . . . to . . . only what is fit for children.'"<ref>[https://www.supremecourt.gov/opinions/boundvolumes/521bv.pdf 521 U.S. 844, 875]</ref>
In 1997, the Supreme Court of the United States unanimously struck down two provisions of the Communications Decency Act, which made "the knowing transmission of obscene or indecent messages to any recipient under 18 years of age" and "knowing[ly] sending or displaying of patently offensive messages in a manner that is available to a person under 18 years of age" felonies<ref>[https://www.supremecourt.gov/opinions/boundvolumes/521bv.pdf 521 U.S. 844, 859]</ref>. "The level of discourse reaching a mailbox simply cannot be limited to that which would be suitable for a sandbox", quoted the Court; "the Government may not 'reduce the adult population . . . to . . . only what is fit for children.'"<ref>[https://www.supremecourt.gov/opinions/boundvolumes/521bv.pdf 521 U.S. 844, 875]</ref>


It's been 25 years and a bill pending in the Senate hasn't learned this lesson. The bill is entitled the "Kids Online Safety Act"<ref>https://www.congress.gov/bill/117th-congress/senate-bill/3663/text</ref>. The following summarizes and analyzes this Act; numbers and letters in parentheses indicate particular clauses of the bill.
It's been 25 years and a bill pending in the Senate hasn't learned this lesson. The bill is entitled the "Kids Online Safety Act"<ref>https://www.congress.gov/bill/117th-congress/senate-bill/3663/text</ref>. This bill wasn't passed in the 117th Congress, but now another bill has been reported from committee<ref>https://www.congress.gov/bill/118th-congress/senate-bill/1409/text</ref>. The following summarizes and analyzes this Act; numbers and letters in parentheses indicate particular clauses of the bill.
==Current Status==
==Current Status==
A bill called the "Protecting Kids on Social Media Act" has been introduced in the 118th Congress. Due to the likelihood that it may be amended in committee, it will not be analyzed until and unless it has been reported from committee.
A bill called the "Protecting Kids on Social Media Act" has been introduced in the 118th Congress. Due to the likelihood that it may be amended in committee, it will not be analyzed until and unless it has been reported from committee.


The "Kids Online Safety Act" has also been reintroduced.
The "Kids Online Safety Act" has also been reintroduced and reported from committee.


==Definitions==
==Definitions==
Anyone 16 or under is a "minor". (2(5))
Anyone under 17 is a "minor" (2(8)), and anyone under 13 is a "child" (2(1)).


A "covered platform" is an "online platform" that "is reasonably likely to be used, by a minor." (2(3))
A "covered platform" is an "online platform", "online video game", messaging, or streaming service that "is reasonably likely to be used, by a minor." Broadband internet, email, teleconferencing, and text messaging services, schools, nonprofits, libraries, news apps, and VPNs are not covered platforms. (2(3))


==Duty of Care==
==Duty of Care==
This Act requires a "covered platform...to act in the best interests of a minor that uses [its] products or services." (3(a))
A covered platform must "take reasonable measures...to prevent and mitigate the following harms to minors:"
* "anxiety, depression, eating disorders, substance use disorders, and suicidal behaviors" (3(a)(1))
* "addiction-like behaviors" (3(a)(2))
* bullying and harassment (3(a)(3))
* "sexual exploitation" (3(a)(4))
* advertising of narcotics, gambling, tobacco, and alcohol (3(a)(5))
* "financial harms" (3(a)(6))


It must "take reasonable measures...to prevent and mitigate:"
It doesn't have to stop minors from "deliberately and independently searching for...content" (3(b)(1)).
* "mental health disorders" (3(b)(1))
* "addiction-like behaviors" (3(b)(2))
* bullying and harassment (3(b)(3))
* "sexual exploitation" (3(b)(4))
* advertising of illegal drugs, gambling, tobacco, and alcohol (3(b)(5))
* "financial harms" (3(b)(6))


==Restrictions==
==Restrictions==
===Safeguards for Minors===
===Safeguards for Minors===
A covered platform must have "readily accessible and easy-to-use safeguards" for minors and their parents to:
A covered platform must have "readily accessible and easy-to-use safeguards" for minors to:
* stop other people, especially "adults with no relationship to" them, from finding them (4(a)(1)(A))
* stop other people from finding them (4(a)(1)(A))
* "prevent other users, registered or not from viewing" and "restricting public access to" their personal data (4(a)(1)(B))
* "prevent other users, whether registered or not from viewing" and "restricting public access to" their personal data (4(a)(1)(B))
* limit features that make them want to use the covered platform more (4(a)(1)(C))
* limit features that make them want to use the covered platform more (4(a)(1)(C))
* opt out of, or limit recommendations from, algorithms (4(a)(1)(D))
* opt out of, or limit recommendations from, algorithms (4(a)(1)(D))
* delete their account and all its data (4(a)(1)(E))
* restrict sharing of their geolocation, and tell them when it's tracked (4(a)(1)(E))
* restrict sharing of their geolocation, and tell them when it's tracked (4(a)(1)(F))
* limit the time they spend there (4(a)(1)(G))


If a platform "knows or reasonably believes" a user to be a minor, these settings must be "the strongest option available" (4(a)(2)).
A minor must be able to delete their account and all its data, and limit the amount of time they spend there (4(a)(2)).
 
If a platform knows a user is a minor, these settings must be "the strongest option available" (4(a)(3)).


===Parental Tools===
===Parental Tools===
A covered platform must provide "readily accessible and easy-to-use" tools for parents of minors to:
A covered platform must provide "readily accessible and easy-to-use" tools for parents of minors to:
* "control" their "privacy and account settings" including the above "[[#Safeguards for Minors]]" (4(b)(2)(A))
* view their "privacy and account settings" including the above "[[#Safeguards for Minors]]", and change the settings for any child (4(b)(2)(A))
* restrict their purchases (4(b)(2)(B))
* restrict their purchases (4(b)(2)(B))
* track the time they spend (4(b)(2)(C))
* track the time they spend (4(b)(2)(C))
* turn off parental controls (4(b)(2)(D))
* "allow [them] to address the harms" covered in "[[#Duty of Care]]" (4(b)(2)(E))


A minor will be told when those tools are in effect (4(b)(3)).
A minor will be told when those tools are in effect (4(b)(3)).
Line 49: Line 47:


===Reporting Mechanisms===
===Reporting Mechanisms===
A covered platform has to have "a readily accessible and easy-to-use" way to report "harms to minors" (4(c)(1)(A)), an "electronic point of contact" specific to them (4(c)(1)(B)), and a way to confirm and track reports (4(c)(1)(C)). It has to respond "in a reasonable and timely manner", within 14 days of getting them (4(c)(2)).
A covered platform has to have "a readily accessible and easy-to-use" way to report "harms to minors" (4(c)(1)(A)), an "electronic point of contact" specific to them (4(c)(1)(B)), and a way to confirm and track reports (4(c)(1)(C)). It must respond "in a reasonable and timely manner", within 10 days of getting them if it has at least 10 million users in the US (4(c)(2)(A)), and 21 days if it doesn't (4(c)(2)(B)).
===Exceptions===
===Exceptions===
This Act doesn't:
This Act doesn't:

Revision as of 06:56, 1 February 2024

In 1997, the Supreme Court of the United States unanimously struck down two provisions of the Communications Decency Act, which made "the knowing transmission of obscene or indecent messages to any recipient under 18 years of age" and "knowing[ly] sending or displaying of patently offensive messages in a manner that is available to a person under 18 years of age" felonies[1]. "The level of discourse reaching a mailbox simply cannot be limited to that which would be suitable for a sandbox", quoted the Court; "the Government may not 'reduce the adult population . . . to . . . only what is fit for children.'"[2]

It's been 25 years and a bill pending in the Senate hasn't learned this lesson. The bill is entitled the "Kids Online Safety Act"[3]. This bill wasn't passed in the 117th Congress, but now another bill has been reported from committee[4]. The following summarizes and analyzes this Act; numbers and letters in parentheses indicate particular clauses of the bill.

Current Status

A bill called the "Protecting Kids on Social Media Act" has been introduced in the 118th Congress. Due to the likelihood that it may be amended in committee, it will not be analyzed until and unless it has been reported from committee.

The "Kids Online Safety Act" has also been reintroduced and reported from committee.

Definitions

Anyone under 17 is a "minor" (2(8)), and anyone under 13 is a "child" (2(1)).

A "covered platform" is an "online platform", "online video game", messaging, or streaming service that "is reasonably likely to be used, by a minor." Broadband internet, email, teleconferencing, and text messaging services, schools, nonprofits, libraries, news apps, and VPNs are not covered platforms. (2(3))

Duty of Care

A covered platform must "take reasonable measures...to prevent and mitigate the following harms to minors:"

  • "anxiety, depression, eating disorders, substance use disorders, and suicidal behaviors" (3(a)(1))
  • "addiction-like behaviors" (3(a)(2))
  • bullying and harassment (3(a)(3))
  • "sexual exploitation" (3(a)(4))
  • advertising of narcotics, gambling, tobacco, and alcohol (3(a)(5))
  • "financial harms" (3(a)(6))

It doesn't have to stop minors from "deliberately and independently searching for...content" (3(b)(1)).

Restrictions

Safeguards for Minors

A covered platform must have "readily accessible and easy-to-use safeguards" for minors to:

  • stop other people from finding them (4(a)(1)(A))
  • "prevent other users, whether registered or not from viewing" and "restricting public access to" their personal data (4(a)(1)(B))
  • limit features that make them want to use the covered platform more (4(a)(1)(C))
  • opt out of, or limit recommendations from, algorithms (4(a)(1)(D))
  • restrict sharing of their geolocation, and tell them when it's tracked (4(a)(1)(E))

A minor must be able to delete their account and all its data, and limit the amount of time they spend there (4(a)(2)).

If a platform knows a user is a minor, these settings must be "the strongest option available" (4(a)(3)).

Parental Tools

A covered platform must provide "readily accessible and easy-to-use" tools for parents of minors to:

  • view their "privacy and account settings" including the above "#Safeguards for Minors", and change the settings for any child (4(b)(2)(A))
  • restrict their purchases (4(b)(2)(B))
  • track the time they spend (4(b)(2)(C))

A minor will be told when those tools are in effect (4(b)(3)).

If a platform "knows or reasonably believes" a user to be a minor, these tools must be in effect by default (4(b)(4)).

Reporting Mechanisms

A covered platform has to have "a readily accessible and easy-to-use" way to report "harms to minors" (4(c)(1)(A)), an "electronic point of contact" specific to them (4(c)(1)(B)), and a way to confirm and track reports (4(c)(1)(C)). It must respond "in a reasonable and timely manner", within 10 days of getting them if it has at least 10 million users in the US (4(c)(2)(A)), and 21 days if it doesn't (4(c)(2)(B)).

Exceptions

This Act doesn't:

  • stop a covered platform from limiting their algorithms from recommending "harmful, obscene, or unlawful content to minors" (4(e)(3)(A)(i)), or blocking or filtering spam or protecting their security (4(e)(3)(A)(ii))
  • force a covered platform to disclose the content of a minor's communications (4(e)(3)(B))
  • stop a covered platform from cooperating with law enforcement agencies regarding activity that it "reasonably and in good faith" believes is illegal (4(e)(3)(C)(i))

Disclosure

Before a person "a covered platform reasonably believes is a minor" signs up for it, it must get them and their parents to acknowledge

  • how it looks after their personal data (5(a)(1)(A))
  • how to access the #Restrictions (5(a)(1)(B))
  • whether it poses any "heightened risk of harm" to them (5(a)(1)(C))

The terms and conditions of any covered platform that uses an algorithm for recommendations must describe

  • how the system uses personal data belonging to minors (5(b)(1))
  • how minors and their parents can control the system (5(b)(2))

Any "advertising aimed at minors" on a covered platform must

  • tell them why they've been targeted for the ad including what personal data was used (5(c)(1)(B))
  • explicitly say that it is an ad (5(c)(1)(C))

All these disclosures must be "clear, accessible, and easy-to-understand".

Transparency

Every year a covered platform that has at least 10 million "active users on a monthly basis in the United States" has to get an "independent, third-party...reasonable inspection" (6(a)(1)).