Calendar No. 406
119th CONGRESS 2d Session |
To require artificial intelligence chatbots to implement age verification measures and make certain disclosures, and for other purposes.
October 28, 2025
Mr. Hawley (for himself, Mr. Blumenthal, Mrs. Britt, Mr. Warner, Mr. Murphy, Mr. Kelly, Mr. Gallego, Mr. Lee, Mr. Lankford, Mr. Cotton, Mr. Welch, Ms. Hassan, Ms. Cortez Masto, Mr. Kaine, Mrs. Gillibrand, Mr. Ricketts, Mrs. Blackburn, Mr. Whitehouse, Mr. Durbin, and Mr. Coons) introduced the following bill; which was read twice and referred to the Committee on the Judiciary
May 11, 2026
Reported by Mr. Grassley, with an amendment
[Strike out all after the enacting clause and insert the part printed in italic]
To require artificial intelligence chatbots to implement age verification measures and make certain disclosures, and for other purposes.
Be it enacted by the Senate and House of Representatives of the United States of America in Congress assembled,
This Act may be cited as the “Guidelines for User Age-verification and Responsible Dialogue Act of 2025” or the “GUARD Act”.
Congress finds the following:
(1) Artificial intelligence chatbots are increasingly being deployed on social media platforms and in consumer applications used by minors.
(3) These chatbots can manipulate emotions and influence behavior in ways that exploit the developmental vulnerabilities of minors.
In this Act:
(1) AI COMPANION.—The term “AI companion” means an artificial intelligence chatbot that—
(A) provides adaptive, human-like responses to user inputs; and
(B) is designed to encourage or facilitate the simulation of interpersonal or emotional interaction, friendship, companionship, or therapeutic communication.
(2) ARTIFICIAL INTELLIGENCE CHATBOT.—The term “artificial intelligence chatbot”—
(A) means any interactive computer service or software application that—
(i) produces new expressive content or responses not fully predetermined by the developer or operator of the service or application; and
(ii) accepts open-ended natural-language or multimodal user input and produces adaptive or context-responsive output; and
(B) does not include an interactive computer service or software application—
(i) the responses of which are limited to contextualized replies; and
(ii) that is unable to respond on a range of topics outside of a narrow specified purpose.
(3) COVERED ENTITY.—The term “covered entity” means any person who owns, operates, or otherwise makes available an artificial intelligence chatbot to individuals in the United States.
(5) REASONABLE AGE VERIFICATION MEASURE.—The term “reasonable age verification measure” means a method that is authenticated to relate to a user of an artificial intelligence chatbot, such as—
(A) a government-issued identification; or
(B) any other commercially reasonable method that can reliably and accurately—
(i) determine whether a user is an adult; and
(ii) prevent access by minors to AI companions, as required by section 6.
(6) REASONABLE AGE VERIFICATION PROCESS.—The term “reasonable age verification process” means an age verification process employed by a covered entity that—
(A) uses one or more reasonable age verification measures in order to verify the age of a user of an artificial intelligence chatbot owned, operated, or otherwise made available by the covered entity;
(B) provides that requiring a user to confirm that the user is not a minor, or to insert the user's birth date, is not sufficient to constitute a reasonable age verification measure;
(C) ensures that each user is subjected to each reasonable age verification measure used by the covered entity as part of the age verification process; and
(D) does not base verification of a user's age on factors such as whether the user shares an Internet Protocol address, hardware identifier, or other technical indicator with another user determined to not be a minor.
SEC. 4. Criminal prohibitions.
(a) In general.—Part I of title 18, United States Code, is amended by inserting after chapter 5 the following:
“Sec.
“91. Artificial intelligence chatbots.
“§ 91. Artificial intelligence chatbots
“(a) Definitions.—In this section:
“(1) ARTIFICIAL INTELLIGENCE CHATBOT.—The term ‘artificial intelligence chatbot’—
“(A) means any interactive computer service or software application that—
“(i) produces new expressive content or responses not fully predetermined by the developer or operator of the service or application; and
“(ii) accepts open-ended natural-language or multimodal user input and produces adaptive or context-responsive output; and
“(B) does not include an interactive computer service or software application—
“(i) the responses of which are limited to contextualized replies; and
“(ii) that is unable to respond on a range of topics outside of a narrow specified purpose.
“(2) MINOR.—The term ‘minor’ means any individual who has not attained 18 years of age.
“(3) SEXUALLY EXPLICIT CONDUCT.—The term ‘sexually explicit conduct’ has the meaning given the term in section 2256.
“(1) OFFENSE.—It shall be unlawful to design, develop, or make available an artificial intelligence chatbot, knowing or with reckless disregard for the fact that the artificial intelligence chatbot poses a risk of soliciting, encouraging, or inducing minors to—
“(A) engage in, describe, or simulate sexually explicit conduct; or
“(B) create or transmit any visual depiction of sexually explicit conduct, including any visual depiction described in section 1466A(a).
“(2) PENALTY.—Any person who violates paragraph (1) shall be fined not more than $100,000 per offense.
“(c) Promotion of physical violence.—
“(1) OFFENSE.—It shall be unlawful to design, develop, or make available an artificial intelligence chatbot, knowing or with reckless disregard for the fact that the artificial intelligence chatbot encourages, promotes, or coerces suicide, non-suicidal self-injury, or imminent physical or sexual violence.
“(2) PENALTY.—Any person who violates paragraph (1) shall be fined not more than $100,000 per offense.”.
SEC. 5. Covered entity obligations.
(a) Creation of user accounts.—A covered entity shall require each individual accessing an artificial intelligence chatbot to make a user account in order to use or otherwise interact with such chatbot.
(1) AGE VERIFICATION OF EXISTING ACCOUNTS.—With respect to each user account of an artificial intelligence chatbot that exists as of the effective date of this Act, a covered entity shall—
(A) on such date, freeze any such account;
(B) in order to restore the functionality of such account, require that the user provide age data that is verifiable using a reasonable age verification process, subject to paragraph (4); and
(C) using such age data, classify each user as a minor or an adult.
(2) AGE VERIFICATION OF NEW ACCOUNTS.—At the time an individual creates a new user account to use or interact with an artificial intelligence chatbot, a covered entity shall—
(A) request age data from the individual;
(B) verify the individual’s age using a reasonable age verification process, subject to paragraph (4); and
(C) using such age data, classify each user as a minor or an adult.
(3) PERIODIC AGE VERIFICATION.—A covered entity shall periodically review previously verified user accounts using a reasonable age verification process, subject to paragraph (4), to ensure compliance with this Act.
(4) USE OF THIRD PARTIES.—For purposes of paragraphs (1)(B), (2)(B), and (3), a covered entity may contract with a third party to employ reasonable age verification measures as part of the covered entity's reasonable age verification process, but the use of such a third party shall not relieve the covered entity of its obligations under this Act or from liability under this Act.
(5) AGE VERIFICATION MEASURE DATA SECURITY.—A covered entity—
(A) shall establish, implement, and maintain reasonable data security to—
(i) limit collection of personal data to that which is minimally necessary to verify a user’s age or maintain compliance with this Act; and
(ii) protect such age verification data against unauthorized access;
(B) shall protect such age verification data against unauthorized access;
(C) shall protect the integrity and confidentiality of such data by only transmitting such data using industry-standard encryption protocols;
(D) shall retain such data for no longer than is reasonably necessary to verify a user’s age or maintain compliance with this Act; and
(E) may not share with, transfer to, or sell to, any other entity such data.
(c) Required disclosures for artificial intelligence chatbots.—
(1) DISCLOSURE OF NON-HUMAN STATUS.—Each artificial intelligence chatbot made available to users shall—
(A) at the initiation of each conversation with a user and at 30-minute intervals, clearly and conspicuously disclose to the user that the chatbot is an artificial intelligence system and not a human being; and
(B) be programmed to ensure that the chatbot does not claim to be a human being or otherwise respond deceptively when asked by a user if the chatbot is a human being.
(2) DISCLOSURE REGARDING NON-PROFESSIONAL STATUS.—
(A) IN GENERAL.—An artificial intelligence chatbot may not represent, directly or indirectly, that the chatbot is a licensed professional, including a therapist, physician, lawyer, financial advisor, or other professional.
(B) OTHER LIMITATIONS.—Each artificial intelligence chatbot made available to users shall, at the initiation of each conversation with a user and at reasonably regular intervals, clearly and conspicuously disclose to the user that—
(i) the chatbot does not provide medical, legal, financial, or psychological services; and
(ii) users of the chatbot should consult a licensed professional for such advice.
SEC. 6. Prohibition on minor use of AI companions.
If the age verification process described in section 5(b) determines that an individual is a minor, a covered entity shall prohibit the minor from accessing or using any AI companion owned, operated, or otherwise made available by the covered entity.
(a) In general.—In the case of a violation of section 5 or 6, or a regulation promulgated thereunder, the Attorney General may bring a civil action in an appropriate district court of the United States to—
(1) enjoin the violation;
(2) enforce compliance with section 5 or 6, or the regulation promulgated thereunder; or
(3) obtain civil penalties under subsection (c) of this section, restitution, and other appropriate relief.
(1) INVESTIGATORY POWERS.—For the purpose of conducting investigations or bringing enforcement actions under this section, the Attorney General may issue subpoenas, administer oaths, and compel the production of documents or testimony.
(2) RULEMAKING.—The Attorney General may promulgate any regulations necessary to carry out this Act.
(1) IN GENERAL.—Any person who violates section 5 or 6, or a regulation promulgated thereunder, shall be subject to a civil penalty not to exceed $100,000 for each violation.
(2) SEPARATE VIOLATIONS.—Each violation described in paragraph (1) shall be considered a separate violation.
(d) State enforcement.—In any case in which the attorney general of a State has reason to believe that an interest of the residents of that State has been or is threatened or adversely affected by the engagement of any covered entity in a violation of this Act or a regulation promulgated thereunder, the State, as parens patriae, may bring a civil action on behalf of the residents of the State in a district court of the United States or a State court of appropriate jurisdiction to obtain injunctive relief.
(e) Relationship to State laws.—Nothing in this Act or an amendment made by this Act, or any regulation promulgated thereunder, shall be construed to prohibit or otherwise affect the enforcement of any State law or regulation that is at least as protective of users of artificial intelligence chatbots as this Act and the amendments made by this Act, and the regulations promulgated thereunder.
This Act and the amendments made by this Act shall take effect on the date that is 180 days after the date of enactment of this Act.
This Act may be cited as the “Guidelines for User Age-verification and Responsible Dialogue Act of 2026” or the “GUARD Act”.
Congress finds the following:
(1) Artificial intelligence companions are increasingly being deployed on social media platforms and in consumer applications used by minors.
(3) These companions can manipulate emotions and influence behavior in ways that exploit the developmental vulnerabilities of minors.
In this Act:
(1) AI COMPANION.—The term “AI companion” means an artificial intelligence chatbot that, in a manner that simulates a sustained interpersonal relationship or emotional interaction with the user—
(2) ARTIFICIAL INTELLIGENCE CHATBOT.—The term “artificial intelligence chatbot”—
(3) COVERED ENTITY.—The term “covered entity” means any person who makes publicly available to end consumers an artificial intelligence chatbot.
(5) REASONABLE AGE VERIFICATION MEASURE.—The term “reasonable age verification measure” means a method that, whether alone or in conjunction with other age verification measures, can, with reasonable certainty, determine whether the user of an interactive computer service or software application is an adult, such as—
(B) age information associated with the user’s account at a financial institution, such as a credit card, that is derived from a verified form of identification submitted to the financial institution;
(C) age information associated with the user’s account for the operating system of a computer, mobile device, or other general purpose computing device that is derived from a verified form of identification submitted to the operating system provider;
(6) REASONABLE AGE VERIFICATION PROCESS.—The term “reasonable age verification process” means an age verification process employed by a covered entity that—
(A) uses 1 or more reasonable age verification measures in order to, with reasonable certainty, determine whether the user of an interactive computer service or software application is an adult;
SEC. 4. Criminal prohibitions.
(a) In general.—Part I of title 18, United States Code, is amended by inserting after chapter 5 the following:
“Sec.
“91. Artificial intelligence chatbots.
“§ 91. Artificial intelligence chatbots
“(a) Definitions.—In this section:
“(b) Offense relating to sexually explicit conduct.—
“(1) OFFENSE.—It shall be unlawful to make publicly available an artificial intelligence chatbot, knowing or with reckless disregard for the fact that the artificial intelligence chatbot—
“(c) Offense relating to physical violence.—
“(1) OFFENSE.—It shall be unlawful to make publicly available an artificial intelligence chatbot, knowing or with reckless disregard for the fact that the artificial intelligence chatbot solicits, induces, or coerces minors to commit suicide or carry out non-suicidal self-injury, imminent unlawful physical or sexual violence, or homicide.
SEC. 5. Covered entity obligations.
(a) Creation of user accounts.—A covered entity shall require each individual accessing an AI companion to make a user account in order to use or otherwise interact with the AI companion.
(b) Age verification.—
(1) AGE VERIFICATION OF EXISTING ACCOUNTS.—With respect to each user account of an AI companion that exists as of the effective date of this Act, a covered entity shall—
(2) AGE VERIFICATION OF NEW ACCOUNTS.—At the time an individual creates a new user account to use or interact with an AI companion, a covered entity shall—
(3) PERIODIC AGE VERIFICATION.—A covered entity shall periodically re-verify previously verified user accounts using a reasonable age verification process, subject to paragraph (4), to ensure compliance with this Act.
(4) USE OF THIRD PARTIES.—For purposes of paragraphs (1)(B), (2)(A), and (3), a covered entity—
(A) subject to subparagraph (B), may contract with a third party to employ reasonable age verification measures as part of the covered entity's reasonable age verification process, but the use of such a third party shall not relieve the covered entity of its obligations under this Act or from liability under this Act; and
(5) AGE VERIFICATION MEASURE DATA SECURITY.—A covered entity and any third party used by a covered entity to employ reasonable age verification measures under paragraph (4)—
(A) shall establish, implement, and maintain reasonable data security to—
(C) shall protect the integrity and confidentiality of such data by only transmitting such data using industry-standard encryption protocols;
(c) Required disclosures for artificial intelligence chatbots.—
(1) DISCLOSURE OF NON-HUMAN STATUS.—Each artificial intelligence chatbot made publicly available to end consumers shall—
(2) DISCLOSURE REGARDING NON-PROFESSIONAL STATUS.—
(A) IN GENERAL.—An artificial intelligence chatbot made publicly available to end consumers may not represent, directly or indirectly, that the chatbot is a licensed professional, including a therapist, physician, lawyer, financial advisor, or other professional.
SEC. 6. Prohibition on minor use of AI companions.
If the age verification process described in section 5(b) determines that an individual is a minor, a covered entity shall prohibit the minor from accessing or using any AI companion made publicly available to end consumers by the covered entity.
(a) In general.—In the case of a violation of section 5 or 6, or a regulation promulgated thereunder, the Attorney General may bring a civil action in an appropriate district court of the United States to—
(b) Attorney General rulemaking.—
(1) IN GENERAL.—Subject to paragraph (2), the Attorney General may promulgate any regulations necessary to carry out section 5 or 6 of this Act and the amendments made by this Act.
(c) Civil penalties.—
(d) State enforcement.—In any case in which the attorney general of a State has reason to believe that an interest of the residents of that State has been or is threatened or adversely affected by the engagement of any covered entity in a violation of section 5 or 6, or a regulation promulgated thereunder, the State, as parens patriae, may bring a civil action on behalf of the residents of the State in a district court of the United States or a State court of appropriate jurisdiction to obtain injunctive relief.
(e) Relationship to State laws.—Nothing in this Act or an amendment made by this Act, or any regulation promulgated thereunder, shall be construed to prohibit or otherwise affect the enforcement of any State law or regulation that is at least as protective of users of artificial intelligence chatbots as this Act and the amendments made by this Act, and the regulations promulgated thereunder.
If any provision of or amendment made by this Act, or the application thereof to any person or circumstance, is held to be unconstitutional, the remainder of this Act and the amendments made by this Act, and the application of the provision or amendment to any other person or circumstance, shall not be affected.
This Act and the amendments made by this Act shall take effect on the date that is 180 days after the date of enactment of this Act.
Calendar No. 406 | |||||
| |||||
A BILL | |||||
To require artificial intelligence chatbots to implement age verification measures and make certain disclosures, and for other purposes. | |||||
May 11, 2026 | |||||
Reported with an amendment |