HB 1709

AN ACT relating to the regulation and reporting on the use of artificial

House Bill Capriglione
Filed

Filed

Bill introduced by legislator

Committee

Hearing

Passed Cmte

Calendar

Passed

Sent

Enrolled

Governor

Signed

89th Regular Session

Jan 14, 2025 - Jun 2, 2025 • Session ended

Awaiting Committee Assignment

Bill filed, pending referral to House committee

← Back to Bills

Committee

Not yet assigned

Fiscal Note

Not available

What This Bill Does

Here's a concise summary of the Texas Artificial Intelligence Governance Act: This bill establishes comprehensive regulations for high-risk artificial intelligence (AI) systems in Texas. It requires developers, distributors, and deployers of high-risk AI systems to: 1. Conduct risk assessments and impact evaluations 2. Disclose AI system usage to consumers 3. Implement risk management policies 4. Prevent algorithmic discrimination 5. Protect consumer privacy and rights The bill prohibits AI systems from: - Manipulating human behavior - Conducting social scoring - Capturing biometric identifiers without consent - Producing harmful or explicit content Key enforcement mechanisms include: - Attorney General oversight - Potential fines up to $100,000 per violation - Consumer right to appeal AI-driven decisions The legislation also creates an AI Regulatory Sandbox Program to facilitate responsible AI innovation and establishes an

Subject Areas

Bill Text

relating to the regulation and reporting on the use of artificial
intelligence systems by certain business entities and state
agencies; providing civil penalties.
BE IT ENACTED BY THE LEGISLATURE OF THE STATE OF TEXAS:
SECTION 1.  This Act may be cited as the Texas Responsible
Artificial Intelligence Governance Act
SECTION 2.  Title 11, Business & Commerce Code, is amended by
adding Subtitle D to read as follows:
SUBTITLE D.  ARTIFICIAL INTELLIGENCE PROTECTION
CHAPTER 551.  ARTIFICIAL INTELLIGENCE PROTECTION
SUBCHAPTER A.  GENERAL PROVISIONS
Sec. 551.001.  DEFINITIONS.  In this chapter:
(1)  "Algorithmic discrimination" means any condition
in which an artificial intelligence system when deployed creates an
unlawful discrimination of a protected classification in violation
of the laws of this state or federal law.
(A)  "Algorithmic discrimination" does not
include the offer, license, or use of a high-risk artificial
intelligence system by a developer or deployer for the sole purpose
of the developer's or deployer's self-testing, for a non-deployed
purpose, to identify, mitigate, or prevent discrimination or
otherwise ensure compliance with state and federal law.
(2)  "Artificial intelligence system" means the use of
machine learning and related technologies that use data to train
statistical models for the purpose of enabling computer systems to
perform tasks normally associated with human intelligence or
perception, such as computer vision, speech or natural language
processing, and content generation.
(3)  "Biometric identifier" means a retina or iris
scan, fingerprint, voiceprint, or record of hand or face geometry.
(4)  "Council" means the Artificial Intelligence
Council established under Chapter 553.
(5)  "Consequential decision" means any decision that
has a material, legal, or similarly significant, effect on a
consumer's access to, cost of, or terms or conditions of:
(A)  a criminal case assessment, a sentencing or
plea agreement analysis, or a pardon, parole, probation, or release
(B)  education enrollment or an education
(C)  employment or an employment opportunity;
(E)  an essential government service;
(F)  residential utility services;
(G)  a health-care service or treatment;
(L)  constitutionally protected services or
(M)  elections or voting process.
(6)  "Consumer" means an individual who is a resident
of this state acting only in an individual or household context.
The term does not include an individual acting in a commercial or
(7)  "Deploy" means to put into effect or
(8)  "Deployer" means a person doing business in this
state that deploys a high-risk artificial intelligence system.
(9)  "Developer" means a person doing business in this
state that develops a high-risk artificial intelligence system or
substantially or intentionally modifies an artificial intelligence
(10)  "Digital service" means a website, an
application, a program, or software that collects or processes
personal identifying information with Internet connectivity.
(11)  "Digital service provider" means a person who:
(A)  owns or operates a digital service;
(B)  determines the purpose of collecting and
processing the personal identifying information of users of the
(C)  determines the means used to collect and
process the personal identifying information of users of the
(12)  "Distributor" means a person, other than the
Developer, that makes an artificial intelligence system available
in the market for a commercial purpose.
(13)  "Generative artificial intelligence" means
artificial intelligence models that can emulate the structure and
characteristics of input data in order to generate derived
synthetic content.  This can include images, videos, audio, text,
(14)  "High-risk artificial intelligence system" means
any artificial intelligence system that is a substantial factor to
a consequential decision.  The term does not include:
(A)  an artificial intelligence system if the
artificial intelligence system is intended to detect
decision-making patterns or deviations from prior decision-making
patterns and is not intended to replace or influence a previously
completed human assessment without sufficient human review;
(B)  an artificial intelligence system that
violates a provision of Subchapter B; or
(C)  the following technologies, unless the
technologies, when deployed, make, or are a substantial factor in
making, a consequential decision:
(viii)  fraud detection systems;
(ix)  internet domain registration;
(xiii)  spam- and robocall-filtering;
(xviii)  web hosting or any similar
(xviv)  any technology that solely
communicates in natural language for the sole purpose of providing
users with information, making referrals or recommendations
relating to customer service, and answering questions and is
subject to an acceptable use policy that prohibits generating
content that is discriminatory or harmful, as long as the system
does not violate any provision listed in Subchapter B.
(15)  "Open source artificial intelligence system"
means an artificial intelligence system that:
(A)  can be used or modified for any purpose
without securing permission from the owner or creator of such an
artificial intelligence system;
(B)  can be shared for any use with or without
(C)  includes information about the data used to
train such system that is sufficiently detailed such that a person
skilled in artificial intelligence could create a substantially
equivalent system when the following are made available freely or
through a non-restrictive license:
(ii)  the source code used to train and run
(iii)  the model weights and parameters of
(16)  "Operational technology" means hardware and
software that detects or causes a change through the direct
monitoring or control of physical devices, processes, and events in
(17)  "Personal data" has the meaning assigned to it by
Section 541.001, Business and Commerce Code.
(18)  "Risk" means the composite measure of an event's
probability of occurring and the magnitude or degree of the
consequences of the corresponding event.
(19)  "Sensitive personal attribute" means race,
political opinions, religious or philosophical beliefs, ethnic
orientation, mental health diagnosis, or sex.  The term does not
include conduct that would be classified as an offense under
(20)  "Social media platform" has the meaning assigned
by Section 120.001, Business and Commerce Code.
(21)  "Substantial factor" means a factor that is:
(A)  considered when making a consequential
(B)  likely to alter the outcome of a
(C)  weighed more heavily than any other factor
contributing to the consequential decision.
(22)  "Intentional and substantial modification" or
"Substantial modification" means a deliberate change made to an
artificial intelligence system that reasonably increases the risk
Sec. 551.002.  APPLICABILITY OF CHAPTER.  This chapter
applies only to a person that is not a small business as defined by
the United States Small Business Administration, and:
(1)  conducts business, promotes, or advertises in this
state or produces a product or service consumed by residents of this
(2)  engages in the development, distribution, or
deployment of a high-risk artificial intelligence system in this
Sec. 551.003.  DEVELOPER DUTIES.  (a)  A developer of a
high-risk artificial intelligence system shall use reasonable care
to protect consumers from any known or reasonably foreseeable risks
of algorithmic discrimination arising from the intended and
contracted uses of the high-risk artificial intelligence system.
(b)  Prior to providing a high-risk artificial intelligence
system to a deployer, a developer shall provide to the deployer, in
writing, a High-Risk Report that consists of:
(1)  a statement describing how the high-risk
artificial intelligence system should be used or not be used;
(2)  any known limitations of the system that could
lead to algorithmic discrimination, the metrics used to measure the
system's performance, which shall include at a minimum, metrics
related to accuracy, explainability, transparency, reliability,
and security set forth in the most recent version of the "Artificial
Intelligence Risk Management Framework: Generative Artificial
Intelligence Profile" published by the National Institute of
Standards and Technology, and how the system performs under those
metrics in its intended use contexts;
(3)  any known or reasonably foreseeable risks of
algorithmic discrimination, arising from its intended or likely
(4)  a high-level summary of the type of data used to
program or train the high-risk artificial intelligence system;
(5)  the data governance measures used to cover the
training datasets and their collection, and the measures used to
examine the suitability of data sources and prevent unlawful
(6)  appropriate principles, processes, and personnel
for the deployers' risk management policy.
(c)  If a high-risk artificial intelligence system is
intentionally or substantially modified after a developer provides
it to a deployer, a developer shall make necessary information in
subsection (b) available to deployers within 30 days of the
(d)  If a developer believes or has reason to believe, that
it deployed a high-risk artificial intelligence system that does
not comply with a requirement of this chapter, the developer shall
immediately take the necessary corrective actions to bring that
system into compliance, including by withdrawing it, disabling it,
and recalling it, as appropriate.  Where applicable, the developer
shall inform the distributors or deployers of the high-risk
artificial intelligence system concerned.
(e)  Where the high-risk artificial intelligence system
presents risks of algorithmic discrimination, unlawful use or
disclosure of personal data, or deceptive manipulation or coercion
of human behavior and the developer knows or should reasonably know
of that risk, it shall immediately investigate the causes, in
collaboration with the deployer, where applicable, and inform the
attorney general in writing of the nature of the non-compliance and
of any relevant corrective action taken.
(f)  Developers shall keep detailed records of any
generative artificial intelligence training data used to develop a
generative artificial intelligence system or service, consistent
with the suggested actions under GV-1.2-007 of the "Artificial
Intelligence Risk Management Framework: Generative Artificial
Intelligence Profile" by the National Institute of Standards and
Technology, or any subsequent versions thereof.
Sec. 551.004.  DISTRIBUTOR DUTIES.  A distributor of a
high-risk artificial intelligence system shall use reasonable care
to protect consumers from any known or reasonably foreseeable risks
of algorithmic discrimination.  If a distributor of a high-risk
artificial intelligence system knows or has reason to know that a
high-risk artificial intelligence system is not in compliance with
any requirement in this chapter, it shall immediately withdraw,
disable, or recall as appropriate, the high-risk artificial
intelligence system from the market until the system has been
brought into compliance with the requirements of this chapter.  The
distributor shall inform the developers of the high-risk artificial
intelligence system concerned and, where applicable, the
Sec. 551.005.  DEPLOYER DUTIES.  A deployer of a high-risk
artificial intelligence system shall use reasonable care to protect
consumers from any known or reasonably foreseeable risks of
algorithmic discrimination.  If a deployer of a high-risk
artificial intelligence system knows or has reason to know that a
high-risk artificial intelligence system is not in compliance with
any requirement in this chapter, it shall immediately suspend the
use of the high-risk artificial intelligence system from the market
until the system has been brought into compliance with the
requirements of this chapter.  The deployer shall inform the
developers of the high-risk artificial intelligence system
concerned and, where applicable, the distributors.
Sec. 551.006.  IMPACT ASSESSMENTS.  (a)  A deployer that
deploys a high-risk artificial intelligence system shall complete
an impact assessment for the high-risk artificial intelligence
system.  A deployer, or a third-party contracted by the deployer for
such purposes, shall complete an impact assessment annually and
within ninety days after any intentional and substantial
modification to the high-risk artificial intelligence system is
made available.  An impact assessment must include, at a minimum,
and to the extent reasonably known by or available to the deployer:
(1)  a statement by the deployer disclosing the
purpose, intended use cases, and deployment context of, and
benefits afforded by, the high-risk artificial intelligence
(2)  an analysis of whether the deployment of the
high-risk artificial intelligence system poses any known or
reasonably foreseeable risks of algorithmic discrimination and, if
so, the nature of the algorithmic discrimination and the steps that
have been taken to mitigate the risks;
(3)  a description of the categories of data the
high-risk artificial intelligence system processes as inputs and
the outputs the high-risk artificial intelligence system produces;
(4)  if the deployer used data to customize the
high-risk artificial intelligence system, an overview of the
categories of data the deployer used to customize the high-risk
artificial intelligence system;
(5)  any metrics used to evaluate the performance and
known limitations of the high-risk artificial intelligence system;
(6)  a description of any transparency measures taken
concerning the high-risk artificial intelligence system, including
any measures taken to disclose to a consumer that the high-risk
artificial intelligence system will be used;
(7)  a description of the post-deployment monitoring
and user safeguards provided concerning the high-risk artificial
intelligence system, including the oversight, use, and learning
process established by the deployer to address issues arising from
the deployment of the high-risk artificial intelligence system; and
(8)  a description of cybersecurity measures and threat
modeling conducted on the system.
(b)  Following an intentional and substantial modification
to a high-risk artificial intelligence system, a deployer must
disclose the extent to which the high-risk artificial intelligence
system was used in a manner that was consistent with, or varied
from, the developer's intended uses of the high-risk artificial
(c)  A single impact assessment may address a comparable set
of high-risk artificial intelligence systems deployed by a
(d)  A deployer shall maintain the most recently completed
impact assessment for a high-risk artificial intelligence system,
all records concerning each impact assessment, and all prior impact
assessments, if any, for at least three years following the final
deployment of the high-risk artificial intelligence system.
(e)  If a deployer, or a third party contracted by the
deployer, completes an impact assessment for the purpose of
complying with another applicable law or regulation, such impact
assessment shall be deemed to satisfy the requirements established
in this subsection if such impact assessment is reasonably similar
in scope and effect to the impact assessment that would otherwise be
completed pursuant to this subsection.
(f)  A deployer may redact any trade secrets as defined by
Section 541.001(33), Business & Commerce Code or information
protected from disclosure by state or federal law.
(g)  Except as provided in subsection (e) of this section, a
developer that makes a high-risk artificial intelligence system
available to a deployer shall make available to the deployer the
documentation and information necessary for a deployer to complete
an impact assessment pursuant to this section.
(h)  A developer that also serves as a deployer for a
high-risk artificial intelligence system is not required to
generate and store an impact assessment unless the high-risk
artificial intelligence system is provided to an unaffiliated
Sec. 551.007.  DISCLOSURE OF A HIGH-RISK ARTIFICIAL
INTELLIGENCE SYSTEM TO CONSUMERS.  (a)  A deployer or developer that
deploys, offers, sells, leases, licenses, gives, or otherwise makes
available a high-risk artificial intelligence system that is
intended to interact with consumers shall disclose to each
consumer, before or at the time of interaction:
(1)  that the consumer is interacting with an
artificial intelligence system;
(2)  the purpose of the system;
(3)  that the system may or will make a consequential
decision affecting the consumer;
(4)  the nature of any consequential decision in which
the system is or may be a substantial factor;
(5)  the factors to be used in making any consequential
(6)  contact information of the deployer;
(A)  any human components of the system;
(B)  any automated components of the system; and
(C)  how human and automated components are used
to inform a consequential decision; and
(8)  a declaration of the consumer's rights under
(b)  Disclosure is required under subsection (a) of this
section regardless of whether it would be obvious to a reasonable
person that the person is interacting with an artificial
(c)  All disclosures under subsection (a) shall be clear and
conspicuous and written in plain language, and avoid the use of a
dark pattern as defined by 541.001, Business & Commerce Code.
(d)  All disclosures under subsection (a) may be linked to a
separate webpage of the developer or deployer.
(e)  Any requirement in this section that may conflict with
state or federal law may be exempt.
Sec. 551.008.  RISK IDENTIFICATION AND MANAGEMENT POLICY.
(a)  A developer or deployer of a high-risk artificial intelligence
system shall, prior to deployment, assess potential risks of
algorithmic discrimination and implement a risk management policy
to govern the development or deployment of the high-risk artificial
intelligence system.  The risk management policy shall:
(1)  specify and incorporate the principles and
processes that the developer or deployer uses to identify,
document, and mitigate, in the development or deployment of a
high-risk artificial intelligence system:
(A)  known or reasonably foreseeable risks of
algorithmic discrimination; and
(B)  prohibited uses and unacceptable risks under
(2)  be reasonable in size, scope, and breadth,
(A)  guidance and standards set forth in the most
recent version of the "Artificial Intelligence Risk Management
Framework: Generative Artificial Intelligence Profile" published
by the National Institute of Standards and Technology;
(B)  any existing risk management guidance,
standards or framework applicable to artificial intelligence
systems designated by the Banking Commissioner or Insurance
Commissioner, if the developer or deployer is regulated by the
Department of Banking or Department of Insurance;
(C)  the size and complexity of the developer or
(D)  the nature, scope, and intended use of the
high-risk artificial intelligence systems developed or deployed;
(E)  the sensitivity and volume of personal data
processed in connection with the high-risk artificial intelligence
(b)  A risk management policy implemented pursuant to this
section may apply to more than one high-risk artificial
intelligence system developed or deployed, so long as the developer
or deployer complies with all of the forgoing requirements and
considerations in adopting and implementing the risk management
policy with respect to each high-risk artificial intelligence
(c)  A developer or deployer may redact or omit any trade
secrets as defined by Section 541.001(33), Business & Commerce Code
or information protected from disclosure by state or federal law.
Sec. 551.009.  RELATIONSHIPS BETWEEN ARTIFICIAL
INTELLIGENCE PARTIES.  Any distributor or deployer, shall be
considered to be a developer of a high-risk artificial intelligence
system for the purposes of this chapter and shall be subject to the
obligations and duties of a developer under this chapter in any of
(1)  they put their name or trademark on a high-risk
artificial intelligence system already placed in the market or put
(2)  they intentionally and substantially modify a
high-risk artificial intelligence system that has already been
placed in the market or has already been put into service in such a
way that it remains a high-risk artificial intelligence system
(3)  they modify the intended purpose of an artificial
intelligence system which has not previously been classified as
high-risk and has already been placed in the market or put into
service in such a way that the artificial intelligence system
concerned becomes a high-risk artificial intelligence system in
accordance with this chapter of a high-risk artificial intelligence
Sec. 551.010.  DIGITAL SERVICE PROVIDER AND SOCIAL MEDIA
PLATFORM DUTIES REGARDING ARTIFICIAL INTELLIGENCE SYSTEMS.  A
digital service provider as defined by Section 509.001(2), Business &
Commerce Code or a social media platform as defined by Section
120.001(1), Business & Commerce Code, shall require advertisers on
the service or platform to agree to terms preventing the deployment
of a high-risk artificial intelligence system on the service or
platform that could expose the users of the service or platform to
algorithmic discrimination or prohibited uses under Subchapter B.
Sec. 551.011.  REPORTING REQUIREMENTS.  (a)  A deployer must
notify, in writing, the council, the attorney general, or the
director of the appropriate state agency that regulates the
deployer's industry, and affected consumers as soon as practicable
after the date on which the deployer discovers or is made aware that
a deployed high-risk artificial intelligence system has caused
algorithmic discrimination of an individual or group of
(b)  If a developer discovers or is made aware that a
deployed high-risk artificial intelligence system is using inputs
or providing outputs that constitute a violation of Subchapter B,
the deployer must cease operation of the offending system as soon as
technically feasible and provide notice to the council and the
attorney general as soon as practicable and not later than the 10th
day after the date on which the developer discovers or is made aware
Sec. 551.012.  SANDBOX PROGRAM EXCEPTION.  (a)  Excluding
violations of Subchapter B, this chapter does not apply to the
development of an artificial intelligence system that is used
exclusively for research, training, testing, or other
pre-deployment activities performed by active participants of the
sandbox program in compliance with Chapter 552.
SUBCHAPTER B.  PROHIBITED USES AND UNACCEPTABLE RISK
Sec. 551.051.  MANIPULATION OF HUMAN BEHAVIOR TO CIRCUMVENT
INFORMED DECISION-MAKING.  An artificial intelligence system shall
not be developed or deployed that uses subliminal techniques beyond
a person's consciousness, or purposefully manipulative or
deceptive techniques, with the objective or the effect of
materially distorting the behavior of a person or a group of persons
by appreciably impairing their ability to make an informed
decision, thereby causing a person to make a decision that the
person would not have otherwise made, in a manner that causes or is
likely to cause significant harm to that person or another person or
Sec. 551.052.  SOCIAL SCORING.  An artificial intelligence
system shall not be developed or deployed for the evaluation or
classification of natural persons or groups of natural persons
based on their social behavior or known, inferred, or predicted
personal characteristics with the intent to determine a social
score or similar categorical estimation or valuation of a person or
Sec. 551.053.  CAPTURE OF BIOMETRIC IDENTIFIERS USING
ARTIFICIAL INTELLIGENCE.  An artificial intelligence system
developed with biometric identifiers of individuals and the
targeted or untargeted gathering of images or other media from the
internet or any other publicly available source shall not be
deployed for the purpose of uniquely identifying a specific
individual. An individual is not considered to be informed nor to
have provided consent for such purpose pursuant to Section 503.001,
Business and Commerce Code, based solely upon the existence on the
internet, or other publicly available source, of an image or other
media containing one or more biometric identifiers.
Sec. 551.054.  CATEGORIZATION BASED ON SENSITIVE
ATTRIBUTES.  An artificial intelligence system shall not be
developed or deployed with the specific purpose of inferring or
interpreting, sensitive personal attributes of a person or group of
persons using biometric identifiers, except for the labeling or
filtering of lawfully acquired biometric identifier data.
Sec. 551.055.  UTILIZATION OF PERSONAL ATTRIBUTES FOR HARM.
An artificial intelligence system shall not utilize
characteristics of a person or a specific group of persons based on
their race, color, disability, religion, sex, national origin, age,
or a specific social or economic situation, with the objective, or
the effect, of materially distorting the behavior of that person or
a person belonging to that group in a manner that causes or is
reasonably likely to cause that person or another person harm.
Sec. 551.056.  CERTAIN SEXUALLY EXPLICIT VIDEOS, IMAGES, AND
CHILD PORNOGRAPHY.  An artificial intelligence system shall not be
developed or deployed that produces, assists, or aids in producing,
or is capable of producing unlawful visual material in violation of
Section 43.26, Penal Code or an unlawful deep fake video or image in
violation of Section 21.165, Penal Code.
SUBCHAPTER C.  ENFORCEMENT AND CONSUMER PROTECTIONS
Sec. 551.101.  CONSTRUCTION AND APPLICATION.  (a)  This
chapter shall be broadly construed and applied to promote its
underlying purposes, which are:
(1)  to facilitate and advance the responsible
development and use of artificial intelligence systems;
(2)  to protect individuals and groups of individuals
from known, and unknown but reasonably foreseeable, risks,
including unlawful algorithmic discrimination;
(3)  to provide transparency regarding those risks in
the development, deployment, or use of artificial intelligence
(4)  to provide reasonable notice regarding the use or
considered use of artificial intelligence systems by state
(b)  this chapter does not apply to the developer of an open
source artificial intelligence system, provided that:
(1)  the system is not deployed as a high-risk
artificial intelligence system and the developer has taken
reasonable steps to ensure that the system cannot be used as a
high-risk artificial intelligence system without substantial
(2)  the weights and technical architecture of the
system are made publicly available.
Sec. 551.102.  ENFORCEMENT AUTHORITY.  The attorney general
has authority to enforce this chapter. Excluding violations of
Subchapter B, researching, training, testing, or the conducting of
other pre-deployment activities by active participants of the
sandbox program, in compliance with Chapter 552, does not subject a
developer or deployer to penalties or actions.
Sec. 551.103.  INTERNET WEBSITE AND COMPLAINT MECHANISM.
The attorney general shall post on the attorney general's Internet
(A)  the responsibilities of a developer,
distributor, and deployer under Subchapter A; and
(B)  an online mechanism through which a consumer
may submit a complaint under this chapter to the attorney general.
Sec. 551.104.  INVESTIGATIVE AUTHORITY.  (a)  If the
attorney general has reasonable cause to believe that a person has
engaged in or is engaging in a violation of this chapter, the
attorney general may issue a civil investigative demand.  The
attorney general shall issue such demands in accordance with and
under the procedures established under Section 15.10.
(b)  The attorney general may request, pursuant to a civil
investigative demand issued under Subsection (a), that a developer
or deployer of a high-risk artificial intelligence system disclose
their risk management policy and impact assessments required under
Subchapter A.  The attorney general may evaluate the risk
management policy and impact assessments for compliance with the
requirements set forth in Subchapter A.
(c)  The attorney general may not institute an action for a
civil penalty against a developer or deployer for artificial
intelligence systems that remain isolated from customer
interaction in a pre-deployment environment.
Sec. 551.105.  NOTICE OF VIOLATION OF CHAPTER; OPPORTUNITY
TO CURE.  Before bringing an action under Section 551.106, the
attorney general shall notify a developer, distributor, or deployer
in writing, not later than the 30th day before bringing the action,
identifying the specific provisions of this chapter the attorney
general alleges have been or are being violated.  The attorney
general may not bring an action against the developer or deployer
(1)  within the 30-day period, the developer or
deployer cures the identified violation; and
(2)  the developer or deployer provides the attorney
general a written statement that the developer or deployer:
(A)  cured the alleged violation;
(B)  notified the consumer, if technically
feasible, and the council that the developer or deployer's
violation was addressed, if the consumer's contact information has
been made available to the developer or deployer and the attorney
(C)  provided supportive documentation to show
how the violation was cured; and
(D)  made changes to internal policies, if
necessary, to reasonably ensure that no such further violations are
Sec. 551.106.  CIVIL PENALTY; INJUNCTION.  (a)  The attorney
general may bring an action in the name of this state to restrain or
enjoin the person from violating this chapter and seek injunctive
(b)  The attorney general may recover reasonable attorney's
fees and other reasonable expenses incurred in investigating and
bringing an action under this section.
(c)  The attorney general may assess and collect an
administrative fine against a developer or deployer who fails to
timely cure a violation or who breaches a written statement
provided to the attorney general, other than those for a prohibited
use, of not less than $50,000 and not more than $100,000 per uncured
(d)  The attorney general may assess and collect an
administrative fine against a developer or deployer who fails to
timely cure a violation of a prohibited use, or whose violation is
determined to be uncurable, of not less than $80,000 and not more
(e)  A developer or deployer who was found in violation of
and continues to operate with the provisions of this chapter shall
be assessed an administrative fine of not less than $2,000 and not
(f)  There is a rebuttable presumption that a developer,
distributor, or deployer used reasonable care as required under
this chapter if the developer, distributor, or deployer complied
with their duties under Subchapter A.
Sec. 551.107.  ENFORCEMENT ACTIONS BY STATE AGENCIES.  A
state agency may sanction an individual licensed, registered, or
certified by that agency for violations of Subchapter B, including:
(1)  the suspension, probation, or revocation of a
license, registration, certificate, or other form of permission to
(2)  monetary penalties up to $100,000.
Sec. 551.108.  CONSUMER RIGHTS AND REMEDIES.  A consumer may
appeal a consequential decision made by a high-risk artificial
intelligence system which has an adverse impact on their health,
safety, or fundamental rights, and shall have the right to obtain
from the deployer clear and meaningful explanations of the role of
the high-risk artificial intelligence system in the
decision-making procedure and the main elements of the decision
SUBCHAPTER D.  CONSTRUCTION OF CHAPTER; LOCAL PREEMPTION
Sec. 551.151.  CONSTRUCTION OF CHAPTER.  This chapter may
not be construed as imposing a requirement on a developer, a
deployer, or other person that adversely affects the rights or
freedoms of any person, including the right of free speech.
Sec. 551.152.  LOCAL PREEMPTION.  This chapter supersedes
and preempts any ordinance, resolution, rule, or other regulation
adopted by a political subdivision regarding the use of high-risk
artificial intelligence systems.
CHAPTER 552. ARTIFICIAL INTELLIGENCE REGULATORY SANDBOX PROGRAM
SUBCHAPTER A.  GENERAL PROVISIONS
Sec. 552.001.  DEFINITIONS.  In this chapter:
(1)  "Applicable agency" means a state agency
responsible for regulating a specific sector impacted by an
artificial intelligence system.
(2)  "Consumer" means a person who engages in
transactions involving an artificial intelligence system or is
directly affected by the use of such a system.
(3)  "Council" means the Artificial Intelligence
Council established by Chapter 553.
(4)  "Department" means the Texas Department of
(5)  "Program participant" means a person or business
entity approved to participate in the sandbox program.
(6)  "Sandbox program" means the regulatory framework
established under this chapter that allows temporary testing of
artificial intelligence systems in a controlled, limited manner
without full regulatory compliance.
SUBCHAPTER B.  SANDBOX PROGRAM FRAMEWORK
Sec. 552.051.  ESTABLISHMENT OF SANDBOX PROGRAM.  (a)  The
department, in coordination with the council, shall administer the
Artificial Intelligence Regulatory Sandbox Program to facilitate
the development, testing, and deployment of innovative artificial
(b)  The sandbox program is designed to:
(1)  promote the safe and innovative use of artificial
intelligence across various sectors including healthcare, finance,
education, and public services;
(2)  encourage the responsible deployment of
artificial intelligence systems while balancing the need for
consumer protection, privacy, and public safety; and
(3)  provide clear guidelines for artificial
intelligence developers to test systems while temporarily exempt
from certain regulatory requirements.
Sec. 552.052.  APPLICATION PROCESS.  (a)  A person or
business entity seeking to participate in the sandbox program must
submit an application to the council.
(b)  The application must include:
(1)  a detailed description of the artificial
intelligence system and its intended use;
(2)  a risk assessment that addresses potential impacts
on consumers, privacy, and public safety;
(3)  a plan for mitigating any adverse consequences
(4)  proof of compliance with federal artificial
intelligence laws and regulations, where applicable.
Sec. 552.053.  DURATION AND SCOPE OF PARTICIPATION.  A
participant may test an artificial intelligence system under the
sandbox program for a period of up to 36 months, unless extended by
SUBCHAPTER C.  OVERSIGHT AND COMPLIANCE
Sec. 552.101.  AGENCY COORDINATION.  (a)  The department
shall coordinate with all relevant state regulatory agencies to
oversee the operations of the sandbox participants.
(b)  A relevant agency may recommend to the department that a
participant's sandbox privileges be revoked if the artificial
(1)  poses undue risk to public safety or welfare;
(2)  violates any federal or state laws that the
sandbox program cannot override.
Sec. 552.102.  REPORTING REQUIREMENTS.  (a)  Each sandbox
participant must submit quarterly reports to the department, which
(1)  system performance metrics;
(2)  updates on how the system mitigates any risks
associated with its operation; and
(3)  feedback from consumers and affected stakeholders
that are using a product that has been deployed from this section.
(b)  The department must submit an annual report to the
(1)  the number of participants in the sandbox program;
(2)  the overall performance and impact of artificial
intelligence systems tested within the program; and
(3)  recommendations for future legislative or
CHAPTER 553.  TEXAS ARTIFICIAL INTELLIGENCE COUNCIL
SUBCHAPTER A.  CREATION AND ORGANIZATION OF COUNCIL
Sec. 553.001.  CREATION OF COUNCIL.  (a)  The Artificial
Intelligence Council is administratively attached to the office of
the governor, and the office of the governor shall provide
administrative support to the council as provided by this section.
(b)  The office of the governor and the council shall enter
into a memorandum of understanding detailing:
(1)  the administrative support the council requires
from the office of the governor to fulfill the purposes of this
(2)  the reimbursement of administrative expenses to
the office of the governor; and
(3)  any other provisions available by law to ensure
the efficient operation of the council as attached to the office of
(c)  The purpose of the council is to:
(1)  ensure artificial intelligence systems are
ethical and in the public's best interest and do not harm public
safety or undermine individual freedoms by finding gaps in the
Penal Code and Chapter 82, Civil Practice and Remedies Code and
making recommendations to the Legislature.
(2)  identify existing laws and regulations that impede
innovation in artificial intelligence development and recommend
(3)  analyze opportunities to improve the efficiency
and effectiveness of state government operations through the use of
artificial intelligence systems;
(4)  investigate and evaluate potential instances of
regulatory capture, including undue influence by technology
companies or disproportionate burdens on smaller innovators;
(5)  investigate and evaluate the influence of
technology companies on other companies and determine the existence
or use of tools or processes designed to censor competitors or
(6)  offer guidance and recommendations to state
agencies including advisory opinions on the ethical and legal use
Sec. 553.002.  COUNCIL MEMBERSHIP.  (a)  The council is
composed of 10 members as follows:
(1)  four members of the public appointed by the
(2)  two members of the public appointed by the
(3)  two members of the public appointed by the speaker
of the house of representatives;
(4)  one senator appointed by the lieutenant governor
(5)  one member of the house of representatives
appointed by the speaker of the house of representatives as a
(b)  Voting members of the council serve staggered four-year
terms, with the terms of four members expiring every two years.
(c)  The governor shall appoint a chair from among the
members, and the council shall elect a vice chair from its
(d)  The council may establish an advisory board composed of
individuals from the public who possess expertise directly related
to the council's functions, including technical, ethical,
regulatory, and other relevant areas.
Sec. 553.003.  QUALIFICATIONS.  (a)  Members of the council
must be Texas residents and have knowledge or expertise in one or
(1)  artificial intelligence technologies;
(2)  data privacy and security;
(3)  ethics in technology or law;
(4)  public policy and regulation; or
(5)  risk management or safety related to artificial
(b)  Members must not hold an office or profit under the
state or federal government at the time of appointment.
Sec. 553.004.  STAFF AND ADMINISTRATION.  The council may
employ an executive director and other personnel as necessary to
SUBCHAPTER B.  POWERS AND DUTIES OF THE COUNCIL
Sec. 553.101.  ISSUANCE OF ADVISORY OPINIONS.  (a)  A state
agency may request a written advisory opinion from the council
regarding the use of artificial intelligence systems in the state.
(b)  The council may issue advisory opinions on state use of
artificial intelligence systems regarding:
(1)  the compliance of artificial intelligence systems
(2)  the ethical implications of artificial
intelligence deployments in the state;
(3)  data privacy and security concerns related to
artificial intelligence systems; or
(4)  potential liability or legal risks associated with
Sec. 553.102.  RULEMAKING AUTHORITY.  (a)  The council may
adopt rules necessary to administer its duties under this chapter,
(1)  procedures for requesting advisory opinions;
(2)  standards for ethical artificial intelligence
(3)  guidelines for evaluating the safety, privacy, and
fairness of artificial intelligence systems.
(b)  The council's rules shall align with state laws on
artificial intelligence, technology, data security, and consumer
Sec. 553.103.  TRAINING AND EDUCATIONAL OUTREACH.  The
council shall conduct training programs for state agencies and
local governments on the ethical use of artificial intelligence
SECTION 3.  Section 503.001, Business & Commerce Code is
amended by adding Subsection (c-3) to read as follows:
(c-3)  This section does not apply to the training,
processing, or storage of biometric identifiers involved in machine
learning or artificial intelligence systems, unless performed for
the purpose of uniquely identifying a specific individual.  If a
biometric identifier captured for the purpose of training an
artificial intelligence system is subsequently used for a
commercial purpose, the person possessing the biometric identifier
is subject to this section's provisions for the possession and
destruction of a biometric identifier and the associated penalties.
SECTION 4.  Sections 541.051(b), 541.101(a), 541.102(a),
and Sec.541.104(a), Business & Commerce Code, are amended to read
Sec. 541.051.  CONSUMER'S PERSONAL DATA RIGHTS; REQUEST TO
EXERCISE RIGHTS.  (a)  A consumer is entitled to exercise the
consumer rights authorized by this section at any time by
submitting a request to a controller specifying the consumer rights
the consumer wishes to exercise.  With respect to the processing of
personal data belonging to a known child, a parent or legal guardian
of the child may exercise the consumer rights on behalf of the
(b)  A controller shall comply with an authenticated
consumer request to exercise the right to:
(1)  confirm whether a controller is processing the
consumer's personal data and to access the personal data;
(2)  correct inaccuracies in the consumer's personal
data, taking into account the nature of the personal data and the
purposes of the processing of the consumer's personal data;
(3)  delete personal data provided by or obtained about
(4)  if the data is available in a digital format,
obtain a copy of the consumer's personal data that the consumer
previously provided to the controller in a portable and, to the
extent technically feasible, readily usable format that allows the
consumer to transmit the data to another controller without
(5)  know if the consumer's personal data is or will be
used in any artificial intelligence system and for what purposes;
([5]6)  opt out of the processing of the personal data
(B)  the sale of personal data; [or]
(C)  the sale of personal data for use in
artificial intelligence systems prior to being collected; or
([C]D)  profiling in furtherance of a decision
that produces a legal or similarly significant effect concerning
Sec. 541.101.  CONTROLLER DUTIES; TRANSPARENCY.  (a)  A
(1)  shall limit the collection of personal data to
what is adequate, relevant, and reasonably necessary in relation to
the purposes for which that personal data is processed, as
disclosed to the consumer; [and]
(2)  for purposes of protecting the confidentiality,
integrity, and accessibility of personal data, shall establish,
implement, and maintain reasonable administrative, technical, and
physical data security practices that are appropriate to the volume
and nature of the personal data at issue.; and
(3)  for purposes of protecting the unauthorized
access, disclosure, alteration, or destruction of data collected,
stored, and processed by artificial intelligence systems, shall
establish, implement, and maintain, reasonable administrative,
technical, and physical data security practices that are
appropriate to the volume and nature of the data collected, stored,
and processed by artificial intelligence systems.
Sec.541.102.  PRIVACY NOTICE.  (a)  A controller shall
provide consumers with a reasonably accessible and clear privacy
(1)  the categories of personal data processed by the
controller, including, if applicable, any sensitive data processed
(2)  the purpose for processing personal data;
(3)  how consumers may exercise their consumer rights
under Subchapter B, including the process by which a consumer may
appeal a controller's decision with regard to the consumer's
(4)  if applicable, the categories of personal data
that the controller shares with third parties;
(5)  if applicable, the categories of third parties
with whom the controller shares personal data; [and]
(6)  if applicable, an acknowledgement of the
collection, use, and sharing of personal data for artificial
([6]7)  a description of the methods required under
Section 541.055 through which consumers can submit requests to
exercise their consumer rights under this chapter.
Sec. 541.104.  DUTIES OF PROCESSOR.  (a)  A processor shall
adhere to the instructions of a controller and shall assist the
controller in meeting or complying with the controller's duties or
requirements under this chapter, including:
(1)  assisting the controller in responding to consumer
rights requests submitted under Section 541.051 by using
appropriate technical and organizational measures, as reasonably
practicable, taking into account the nature of processing and the
information available to the processor;
(2)  assisting the controller with regard to complying
with the [requirement]requirements relating to the security of
processing personal data, and if applicable, the data collected,
stored, and processed by artificial intelligence systems and to the
notification of a breach of security of the processor's system
under Chapter 521, taking into account the nature of processing and
the information available to the processor; and
(3)  providing necessary information to enable the
controller to conduct and document data protection assessments
SECTION 5.  Subtitle E, Title 4, Labor Code, is amended by
adding Chapter 319 to read as follows:
CHAPTER 319.  TEXAS ARTIFICIAL INTELLIGENCE WORKFORCE DEVELOPMENT
SUBCHAPTER A.  GENERAL PROVISIONS
Sec. 319.001.  DEFINITIONS.  In this chapter:
(1)  "Artificial intelligence industry" means
businesses, research organizations, governmental entities, and
educational institutions engaged in the development, deployment,
or use of artificial intelligence technologies in Texas.
(2)  "Commission" means the Texas Workforce
(3)  "Eligible entity" means Texas-based businesses in
the artificial intelligence industry, public school districts,
community colleges, public technical institutes, and workforce
(4)  "Program" means the Texas Artificial Intelligence
Workforce Development Grant Program established under this
SUBCHAPTER B.  ARTIFICIAL INTELLIGENCE WORKFORCE DEVELOPMENT GRANT
Sec. 319.051.  ESTABLISHMENT OF GRANT PROGRAM.  (a)  The
commission shall establish the Texas Artificial Intelligence
Workforce Development Grant Program to:
(1)  support and assist Texas-based artificial
intelligence companies in developing a skilled workforce;
(2)  provide grants to local community colleges and
public high schools to implement or expand career and technical
education programs focused on artificial intelligence readiness
(3)  offer opportunities to retrain and reskill workers
through partnerships with the artificial intelligence industry and
workforce development programs.
(b)  The program is intended to:
(1)  prepare Texas workers and students for employment
in the rapidly growing artificial intelligence industry;
(2)  support the creation of postsecondary programs and
certifications relevant to current artificial intelligence
(3)  ensure that Texas maintains a competitive edge in
artificial intelligence innovation and workforce development; and
(4)  address workforce gaps in artificial
intelligence-related fields, including data science,
cybersecurity, machine learning, robotics, and automation.
(c)  The commission shall adopt rules necessary to implement
Sec. 319.052.  FEDERAL FUNDS AND GIFTS, GRANTS, AND
In addition to other money appropriated by the legislature,
for the purpose of providing artificial intelligence workforce
opportunities under the program established under this subchapter
(1)  seek and apply for any available federal funds;
(2)  solicit and accept gifts, grants, and donations
from any other source, public or private, as necessary to ensure
effective implementation of the program.
Sec. 319.053.  ELIGIBILITY FOR GRANTS.  (a)  The following
entities are eligible to apply for grants under this program:
(1)  Texas-based businesses engaged in the development
or deployment of artificial intelligence technologies;
(2)  public school districts and charter schools
offering or seeking to offer career and technical education
programs in artificial intelligence-related fields or to update
existing curricula to address these fields;
(3)  public community colleges and technical
institutes that develop artificial intelligence-related curricula
or training programs or update existing curricula or training
programs to incorporate artificial intelligence training; and
(4)  workforce development organizations in
partnership with artificial intelligence companies to reskill and
retrain workers in artificial intelligence competencies.
(b)  To be eligible, the entity must:
(1)  submit an application to the commission in the
form and manner prescribed by the commission; and
(2)  demonstrate the capacity to develop and implement
training, educational, or workforce development programs that
align with the needs of the artificial intelligence industry in
Texas and lead to knowledge, skills, and work-based experiences
that are transferable to similar employment opportunities in the
artificial intelligence industry.
Sec. 319.054.  USE OF GRANTS.  (a)  Grants awarded under the
(1)  developing or expanding workforce training
programs for artificial intelligence-related skills, including but
not limited to machine learning, data analysis, software
(2)  creating or enhancing career and technical
education programs in artificial intelligence for high school
students, with a focus on preparing them for careers in 

Bill Sponsors

Legislators who authored or co-sponsored this bill.

Bill History

filed

Bill filed: AN ACT relating to the regulation and reporting on the use of artificial