BILL AS INTRODUCED H.711
2024 Page 1 of 14
1 H.711
2 Introduced by Representatives Priestley of Bradford, Burrows of West
3 Windsor, Chase of Chester, Christie of Hartford, Jerome of
4 Brandon, Masland of Thetford, Roberts of Halifax, Sibilia of
5 Dover, Sims of Craftsbury, Templeman of Brownington, White
6 of Bethel, and Williams of Barre City
7 Referred to Committee on
8 Date:
9 Subject: Commerce and trade; consumer protection; liability for developers
10 and deployers of artificial intelligence systems
11 Statement of purpose of bill as introduced: This bill proposes to create
12 oversight and liability standards for developers and deployers of inherently
13 dangerous artificial intelligence systems.
14 An act relating to creating oversight and liability standards for developers
15 and deployers of inherently dangerous artificial intelligence systems
16 It is hereby enacted by the General Assembly of the State of Vermont:
17 Sec. 1. 9 V.S.A. chapter 63, subchapter 12 is added to read:
18 Subchapter 12. Artificial Intelligence Oversight and Liability
19 § 2495a. LEGISLATIVE INTENT
VT LEG #373232 v.1
BILL AS INTRODUCED H.711
2024 Page 2 of 14
1 (a) Artificial intelligence systems are products that shift decision-making
2 power and responsibility away from persons to software-based systems, often
3 without direct human oversight. An artificial intelligence system can be
4 inherently dangerous due to its capabilities, potential for misuse or
5 exploitation, and ability to unilaterally evolve.
6 (b) Developers of sophisticated artificial intelligence systems have an
7 obligation to make such systems safe when used in reasonably foreseeable
8 ways. Deployers of these products also have an obligation to ensure that the
9 products are safe and used in a way that does not materially affect an
10 individual’s rights.
11 (c) In the artificial intelligence ecosystem, there will typically be multiple
12 suppliers upstream of a consumer. The original developer of an artificial
13 intelligence system should be responsible for harms attributable to the artificial
14 intelligence system, even if the developer is not the deployer of the system to a
15 consumer. Small businesses using off-the-shelf artificial intelligence products
16 according to the product’s terms of use are not intended to be covered by this
17 act.
18 § 2495b. DEFINITIONS
19 As used in this subchapter:
20 (1) “Artificial intelligence system” means a machine-based system that
21 can, for a given set of objectives, make predictions, recommendations, or VT LEG #373232 v.1
BILL AS INTRODUCED H.711
2024 Page 3 of 14
1 decisions influencing real or virtual environments. Artificial intelligence
2 systems use machine- and human-based inputs to perceive real and virtual
3 environments, abstract such perceptions into models through analysis in an
4 automated manner, and use model inference to formulate options for
5 information or action.
6 (2) “Biometric data” means data that depict or describe physical,
7 biological, or behavioral traits, characteristics, or measurements of or relating
8 to an identified or identifiable person’s body. Biometric information includes
9 depictions, images, descriptions, or recordings of an individual’s facial
10 features, iris or retina, finger or handprints, voice, genetics, or characteristic
11 movements or gestures. Biometric information also includes data derived from
12 such depictions, images, descriptions, or recordings, to the extent that it would
13 be reasonably possible to identify the person from whose information the data
14 had been derived.
15 (3) “Consequential decision” means a decision that either has a legal or
16 similarly significant effect on an individual’s access to the criminal justice
17 system, housing, employment, credit, education, health care, or insurance.
18 (4) “Consumer” means any individual who is a resident of this State.
19 (5) “Deployer” means a person, including a developer, who uses or
20 operates an artificial intelligence system for internal use or for use by third
21 parties in the State.
VT LEG #373232 v.1
BILL AS INTRODUCED H.711
2024 Page 4 of 14
1 (6) “Developer” means a person who designs, codes, produces, owns, or
2 substantially modifies an artificial intelligence system for internal use or for
3 use by a third party in the State.
4 (7) “Dual-use foundational model” means an artificial intelligence
5 system that:
6 (A) is trained on broad data;
7 (B) generally uses self-supervision;
8 (C) contains at least 10 billion parameters;
9 (D) is applicable across a wide range of contexts; and
10 (E) exhibits, or could be easily modified to exhibit, high levels of
11 performance at tasks that pose a serious risk to economic security, public
12 health or safety, or any combination of those matters, such as by:
13 (i) substantially lowering the barrier of entry for nonexperts to
14 design, synthesize, acquire, or use chemical, biological, radiological, or nuclear
15 (CBRN) weapons;
16 (ii) enabling powerful offensive cyber operations through
17 automated vulnerability discovery and exploitation against a wide range of
18 potential targets of cyber attacks; or
19 (iii) permitting the evasion of human control or oversight through
20 means of deception or obfuscation.
VT LEG #373232 v.1
BILL AS INTRODUCED H.711
2024 Page 5 of 14
1 (8) “Generative artificial intelligence system” means an artificial
2 intelligence system that is capable of generating output including text,
3 imagery, audio, and synthetic data.
4 (9) “High-risk artificial intelligence system” means any artificial
5 intelligence system, regardless of the number of parameters and supervision
6 structure, that is:
7 (A) used, reasonably foreseeable as being used, or is a controlling
8 factor in making a consequential decision;
9 (B) used, or reasonably foreseeable as being used, to categorize
10 groups of persons by sensitive and protected characteristics, such as race,
11 ethnic origin, or religious belief;
12 (C) used, or reasonably foreseeable as being used, in the direct
13 management or operation of critical infrastructure;
14 (D) used, or reasonably foreseeable as being used, in vehicles,
15 medical devices, or in the safety system of a product;
16 (E) used, or reasonably foreseeable as being used, to influence
17 elections or voters; or
18 (F) used to collect the biometric data of an individual from a
19 biometric identification system without consent.
VT LEG #373232 v.1
BILL AS INTRODUCED H.711
2024 Page 6 of 14
1 (10) “Inherently dangerous artificial intelligence system” means a high-
2 risk artificial intelligence system, dual-use foundational model, or generative
3 artificial intelligence system.
4 § 2495c. OVERSIGHT AND ENFORCEMENT
5 (a) The Division of Artificial Intelligence within the Agency of Digital
6 Services shall collect and review Artificial Intelligence Safety and Impact
7 Assessments pursuant to this subchapter.
8 (b) The Attorney General shall enforce the provisions of this subchapter
9 and may bring an action in the name of the State against a deployer or
10 developer for noncompliance to restrain by temporary or permanent injunction
11 the noncompliance. The action may be brought in the Superior Court of the
12 county in which such person resides, has a place of business, or is doing
13 business. Said courts are authorized to issue temporary or permanent
14 injunctions to restrain and prevent violations of this subchapter, such
15 injunctions to be issued without bonds, or to dissolve, or revoke the certificate
16 of authority of, a deployer or developer.
17 § 2495d. ARTIFICIAL INTELLIGENCE SYSTEM SAFETY AND IMPACT
18 ASSESSMENT
19 (a) Each deployer of an inherently dangerous artificial intelligence system
20 shall:
VT LEG #373232 v.1
BILL AS INTRODUCED H.711
2024 Page 7 of 14
1 (1) submit to the Division of Artificial Intelligence an Artificial
2 Intelligence System Safety and Impact Assessment prior to deploying the
3 inherently dangerous artificial intelligence system in this State, and every two
4 years thereafter; and
5 (2) submit to the Division of Artificial Intelligence an updated Artificial
6 Intelligence System Safety and Impact Assessment if the deployer makes a
7 material and substantial change to the inherently dangerous artificial
8 intelligence system that includes:
9 (A) the purpose for which the system is used for; or
10 (B) the type of data the system processes or uses for training
11 purposes.
12 (b) Each Artificial Intelligence System Impact Assessment pursuant to
13 subsection (a) of this section shall include, with respect to the inherently
14 dangerous artificial intelligence system:
15 (1) the purpose of the system;
16 (2) the deployment context and intended use cases;
17 (3) the benefits of use;
18 (4) any foreseeable risk of unintended or unauthorized uses and the steps
19 taken, to the extent reasonable, to mitigate such risk;
20 (5) whether the model is proprietary;
VT LEG #373232 v.1
BILL AS INTRODUCED H.711
2024 Page 8 of 14
1 (6) a description of the data the system processes or uses for training
2 purposes;
3 (7) a description of transparency measures, including identifying to
4 individuals when the system is in use;
5 (8) identification of any third-party artificial intelligence systems or
6 datasets the deployer relies on to train or operate the system, if applicable;
7 (9) whether the developer of the system, if different than the deployer,
8 disclosed the information pursuant to subsection 2495e(b) of this chapter as
9 well as the results of testing, vulnerabilities, and the parameters for safe and
10 intended use;
11 (10) a description of the data the system, once deployed, processes as
12 inputs;
13 (11) a description of postdeployment monitoring and user safeguards,
14 including a description of the oversight process in place to address issues as
15 issues arise; and
16 (12) a description of how the model impacts consequential decisions or
17 the collection of biometric data.
18 (c) Each deployer of a high-risk artificial intelligence system must submit a
19 one-, six-, and 12-month testing result to the Division of Artificial Intelligence
20 showing the reliability of the results generated by the system, any variance in
VT LEG #373232 v.1
BILL AS INTRODUCED H.711
2024 Page 9 of 14
1 those results over the testing periods, and any mitigation strategies for
2 variances, in the first year of deployment.
3 (d) Upon the Division of Artificial Intelligence receiving notice that a
4 deployer of an inherently dangerous artificial intelligence system is not in
5 compliance with the requirements under this section, the Division shall
6 immediately notify the deployer of the finding in writing and order the
7 deployer to submit the assessment required pursuant to subsection (a) of this
8 section. If the deployer fails to submit the assessment within 45 days after
9 receiving the notice, the Division of Artificial Intelligence shall notify the
10 Attorney General in writing of the violation.
11 § 2495e. STANDARD OF CARE
12 (a) Each developer or deployer of any inherently dangerous artificial
13 intelligence system that could be reasonably expected to impact consumers
14 shall exercise reasonable care to avoid any reasonably foreseeable risk arising
15 out of the development, intentional and substantial modification, or
16 deployment of an artificial intelligence system that causes or is likely to cause:
17 (1) the commission of a crime or unlawful act;
18 (2) any unfair or deceptive treatment of or unlawful impact on an
19 individual;
20 (3) any physical, financial, relational, or reputational injury on an
21 individual;
VT LEG #373232 v.1
BILL AS INTRODUCED H.711
2024 Page 10 of 14
1 (4) psychological injuries that would be highly offensive to a reasonable
2 person;
3 (5) any physical or other intrusion upon the solitude or seclusion, or the
4 private affairs or concerns, of a person if such intrusion would be offensive to a
5 reasonable person;
6 (6) any violation to the intellectual property rights of persons under
7 applicable State and federal laws;
8 (7) discrimination on the basis of a person’s or class of person’s actual
9 or perceived race, color, ethnicity, sex, sexual orientation, gender identity, sex
10 characteristics, religion, national origin, familial status, biometric information,
11 or disability status;
12 (8) distortion of a person’s behavior in a manner that causes or is likely
13 to cause that person or another person physical or psychological harm; or
14 (9) the exploitation of the vulnerabilities of a specific group of persons
15 due to their age or physical or mental disability in order to materially distort
16 the behavior of a person pertaining to that group in a manner that causes or is
17 likely to cause that person or another person physical or psychological harm.
18 (b) Each developer of an inherently dangerous artificial intelligence system
19 shall:
20 (1) document and disclose to any actual or potential deployer of the
21 artificial intelligence system any reasonably foreseeable risk, including by
VT LEG #373232 v.1
BILL AS INTRODUCED H.711
2024 Page 11 of 14
1 unintended or unauthorized uses, that causes or is likely to cause any of the
2 injuries as set forth in subsection (a) of this section; and
3 (2) document and disclose to any actual or potential deployer of the
4 artificial intelligence system any risk mitigation processes that are reasonably
5 foreseeable to mitigate any injury as set forth in subsection (a) of this section.
6 § 2495f. UNSAFE ARTIFICIAL INTELLIGENCE PRODUCTS,
7 PROHIBITIONS
8 (a) No developer shall offer, sell, lease, give, or otherwise place in the
9 stream of commerce:
10 (1) an inherently dangerous artificial intelligence system, unless the
11 developer has conducted a documented testing, evaluation, verification, and
12 validation of that system at least as stringent as the latest version of the
13 Artificial Intelligence Risk Management Framework published by the National
14 Institute of Standards and Technology (NIST); or
15 (2) an artificial intelligence system that creates reasonably foreseeable
16 risks pursuant to section 2495e of this chapter, unless the developer mitigates
17 these risks to the extent possible, considers alternatives, and discloses
18 vulnerabilities and mitigation tactics to a deployer.
19 (b) No deployer shall deploy an inherently dangerous artificial intelligence
20 system or an artificial intelligence system that creates reasonably foreseeable
21 risks pursuant to section 2495e of this chapter unless the deployer has designed VT LEG #373232 v.1
BILL AS INTRODUCED H.711
2024 Page 12 of 14
1 and implemented a risk management policy and program for such model or
2 system. The risk management policy shall specify the principles, processes,
3 and personnel that the deployer shall use in maintaining the risk management
4 program to identify, mitigate, and document any risk that is a reasonably
5 foreseeable consequence of deploying or