|
|
|
A BILL TO BE ENTITLED
|
|
AN ACT
|
|
relating to the regulation and reporting on the use of artificial |
|
intelligence systems by certain business entities and state |
|
agencies; providing civil penalties. |
|
BE IT ENACTED BY THE LEGISLATURE OF THE STATE OF TEXAS: |
|
SECTION 1. This Act may be cited as the Texas Responsible |
|
Artificial Intelligence Governance Act |
|
SECTION 2. Title 11, Business & Commerce Code, is amended by |
|
adding Subtitle D to read as follows: |
|
SUBTITLE D. ARTIFICIAL INTELLIGENCE PROTECTION |
|
CHAPTER 551. ARTIFICIAL INTELLIGENCE PROTECTION |
|
SUBCHAPTER A. GENERAL PROVISIONS |
|
Sec. 551.001. DEFINITIONS. In this chapter: |
|
(1) "Algorithmic discrimination" means any condition |
|
in which an artificial intelligence system when deployed creates an |
|
unlawful discrimination of a protected classification in violation |
|
of the laws of this state or federal law. |
|
(A) "Algorithmic discrimination" does not |
|
include the offer, license, or use of a high-risk artificial |
|
intelligence system by a developer or deployer for the sole purpose |
|
of the developer's or deployer's self-testing, for a non-deployed |
|
purpose, to identify, mitigate, or prevent discrimination or |
|
otherwise ensure compliance with state and federal law. |
|
(2) "Artificial intelligence system" means the use of |
|
machine learning and related technologies that use data to train |
|
statistical models for the purpose of enabling computer systems to |
|
perform tasks normally associated with human intelligence or |
|
perception, such as computer vision, speech or natural language |
|
processing, and content generation. |
|
(3) "Biometric identifier" means a retina or iris |
|
scan, fingerprint, voiceprint, or record of hand or face geometry. |
|
(4) "Council" means the Artificial Intelligence |
|
Council established under Chapter 553. |
|
(5) "Consequential decision" means any decision that |
|
has a material, legal, or similarly significant, effect on a |
|
consumer's access to, cost of, or terms or conditions of: |
|
(A) a criminal case assessment, a sentencing or |
|
plea agreement analysis, or a pardon, parole, probation, or release |
|
decision; |
|
(B) education enrollment or an education |
|
opportunity; |
|
(C) employment or an employment opportunity; |
|
(D) a financial service; |
|
(E) an essential government service; |
|
(F) residential utility services; |
|
(G) a health-care service or treatment; |
|
(H) housing; |
|
(I) insurance; |
|
(J) a legal service; |
|
(K) a transportation service; |
|
(L) constitutionally protected services or |
|
products; or |
|
(M) elections or voting process. |
|
(6) "Consumer" means an individual who is a resident |
|
of this state acting only in an individual or household context. |
|
The term does not include an individual acting in a commercial or |
|
employment context. |
|
(7) "Deploy" means to put into effect or |
|
commercialize. |
|
(8) "Deployer" means a person doing business in this |
|
state that deploys a high-risk artificial intelligence system. |
|
(9) "Developer" means a person doing business in this |
|
state that develops a high-risk artificial intelligence system or |
|
substantially or intentionally modifies an artificial intelligence |
|
system. |
|
(10) "Digital service" means a website, an |
|
application, a program, or software that collects or processes |
|
personal identifying information with Internet connectivity. |
|
(11) "Digital service provider" means a person who: |
|
(A) owns or operates a digital service; |
|
(B) determines the purpose of collecting and |
|
processing the personal identifying information of users of the |
|
digital service; and |
|
(C) determines the means used to collect and |
|
process the personal identifying information of users of the |
|
digital service. |
|
(12) "Distributor" means a person, other than the |
|
Developer, that makes an artificial intelligence system available |
|
in the market for a commercial purpose. |
|
(13) "Generative artificial intelligence" means |
|
artificial intelligence models that can emulate the structure and |
|
characteristics of input data in order to generate derived |
|
synthetic content. This can include images, videos, audio, text, |
|
and other digital content. |
|
(14) "High-risk artificial intelligence system" means |
|
any artificial intelligence system that is a substantial factor to |
|
a consequential decision. The term does not include: |
|
(A) an artificial intelligence system if the |
|
artificial intelligence system is intended to detect |
|
decision-making patterns or deviations from prior decision-making |
|
patterns and is not intended to replace or influence a previously |
|
completed human assessment without sufficient human review; |
|
(B) an artificial intelligence system that |
|
violates a provision of Subchapter B; or |
|
(C) the following technologies, unless the |
|
technologies, when deployed, make, or are a substantial factor in |
|
making, a consequential decision: |
|
(i) anti-malware; |
|
(ii) anti-virus; |
|
(iii) calculators; |
|
(iv) cybersecurity; |
|
(v) databases; |
|
(vi) data storage; |
|
(vii) firewall; |
|
(viii) fraud detection systems; |
|
(ix) internet domain registration; |
|
(x) internet website loading; |
|
(xi) networking; |
|
(xii) operational technology; |
|
(xiii) spam- and robocall-filtering; |
|
(xiv) spell-checking; |
|
(xv) spreadsheets; |
|
(xvi) web caching; |
|
(xvii) web scraping; |
|
(xviii) web hosting or any similar |
|
technology; or |
|
(xviv) any technology that solely |
|
communicates in natural language for the sole purpose of providing |
|
users with information, making referrals or recommendations |
|
relating to customer service, and answering questions and is |
|
subject to an acceptable use policy that prohibits generating |
|
content that is discriminatory or harmful, as long as the system |
|
does not violate any provision listed in Subchapter B. |
|
(15) "Open source artificial intelligence system" |
|
means an artificial intelligence system that: |
|
(A) can be used or modified for any purpose |
|
without securing permission from the owner or creator of such an |
|
artificial intelligence system; |
|
(B) can be shared for any use with or without |
|
modifications; and |
|
(C) includes information about the data used to |
|
train such system that is sufficiently detailed such that a person |
|
skilled in artificial intelligence could create a substantially |
|
equivalent system when the following are made available freely or |
|
through a non-restrictive license: |
|
(i) the same or similar data; |
|
(ii) the source code used to train and run |
|
such system; and |
|
(iii) the model weights and parameters of |
|
such system. |
|
(16) "Operational technology" means hardware and |
|
software that detects or causes a change through the direct |
|
monitoring or control of physical devices, processes, and events in |
|
the enterprise. |
|
(17) "Personal data" has the meaning assigned to it by |
|
Section 541.001, Business and Commerce Code. |
|
(18) "Risk" means the composite measure of an event's |
|
probability of occurring and the magnitude or degree of the |
|
consequences of the corresponding event. |
|
(19) "Sensitive personal attribute" means race, |
|
political opinions, religious or philosophical beliefs, ethnic |
|
orientation, mental health diagnosis, or sex. The term does not |
|
include conduct that would be classified as an offense under |
|
Chapter 21, Penal Code. |
|
(20) "Social media platform" has the meaning assigned |
|
by Section 120.001, Business and Commerce Code. |
|
(21) "Substantial factor" means a factor that is: |
|
(A) considered when making a consequential |
|
decision; |
|
(B) likely to alter the outcome of a |
|
consequential decision; and |
|
(C) weighed more heavily than any other factor |
|
contributing to the consequential decision. |
|
(22) "Intentional and substantial modification" or |
|
"Substantial modification" means a deliberate change made to an |
|
artificial intelligence system that reasonably increases the risk |
|
of algorithmic discrimination. |
|
Sec. 551.002. APPLICABILITY OF CHAPTER. This chapter |
|
applies only to a person that is not a small business as defined by |
|
the United States Small Business Administration, and: |
|
(1) conducts business, promotes, or advertises in this |
|
state or produces a product or service consumed by residents of this |
|
state; or |
|
(2) engages in the development, distribution, or |
|
deployment of a high-risk artificial intelligence system in this |
|
state. |
|
Sec. 551.003. DEVELOPER DUTIES. (a) A developer of a |
|
high-risk artificial intelligence system shall use reasonable care |
|
to protect consumers from any known or reasonably foreseeable risks |
|
of algorithmic discrimination arising from the intended and |
|
contracted uses of the high-risk artificial intelligence system. |
|
(b) Prior to providing a high-risk artificial intelligence |
|
system to a deployer, a developer shall provide to the deployer, in |
|
writing, a High-Risk Report that consists of: |
|
(1) a statement describing how the high-risk |
|
artificial intelligence system should be used or not be used; |
|
(2) any known limitations of the system that could |
|
lead to algorithmic discrimination, the metrics used to measure the |
|
system's performance, which shall include at a minimum, metrics |
|
related to accuracy, explainability, transparency, reliability, |
|
and security set forth in the most recent version of the "Artificial |
|
Intelligence Risk Management Framework: Generative Artificial |
|
Intelligence Profile" published by the National Institute of |
|
Standards and Technology, and how the system performs under those |
|
metrics in its intended use contexts; |
|
(3) any known or reasonably foreseeable risks of |
|
algorithmic discrimination, arising from its intended or likely |
|
use; |
|
(4) a high-level summary of the type of data used to |
|
program or train the high-risk artificial intelligence system; |
|
(5) the data governance measures used to cover the |
|
training datasets and their collection, and the measures used to |
|
examine the suitability of data sources and prevent unlawful |
|
discriminatory biases; and |
|
(6) appropriate principles, processes, and personnel |
|
for the deployers' risk management policy. |
|
(c) If a high-risk artificial intelligence system is |
|
intentionally or substantially modified after a developer provides |
|
it to a deployer, a developer shall make necessary information in |
|
subsection (b) available to deployers within 30 days of the |
|
modification. |
|
(d) If a developer believes or has reason to believe, that |
|
it deployed a high-risk artificial intelligence system that does |
|
not comply with a requirement of this chapter, the developer shall |
|
immediately take the necessary corrective actions to bring that |
|
system into compliance, including by withdrawing it, disabling it, |
|
and recalling it, as appropriate. Where applicable, the developer |
|
shall inform the distributors or deployers of the high-risk |
|
artificial intelligence system concerned. |
|
(e) Where the high-risk artificial intelligence system |
|
presents risks of algorithmic discrimination, unlawful use or |
|
disclosure of personal data, or deceptive manipulation or coercion |
|
of human behavior and the developer knows or should reasonably know |
|
of that risk, it shall immediately investigate the causes, in |
|
collaboration with the deployer, where applicable, and inform the |
|
attorney general in writing of the nature of the non-compliance and |
|
of any relevant corrective action taken. |
|
(f) Developers shall keep detailed records of any |
|
generative artificial intelligence training data used to develop a |
|
generative artificial intelligence system or service, consistent |
|
with the suggested actions under GV-1.2-007 of the "Artificial |
|
Intelligence Risk Management Framework: Generative Artificial |
|
Intelligence Profile" by the National Institute of Standards and |
|
Technology, or any subsequent versions thereof. |
|
Sec. 551.004. DISTRIBUTOR DUTIES. A distributor of a |
|
high-risk artificial intelligence system shall use reasonable care |
|
to protect consumers from any known or reasonably foreseeable risks |
|
of algorithmic discrimination. If a distributor of a high-risk |
|
artificial intelligence system knows or has reason to know that a |
|
high-risk artificial intelligence system is not in compliance with |
|
any requirement in this chapter, it shall immediately withdraw, |
|
disable, or recall as appropriate, the high-risk artificial |
|
intelligence system from the market until the system has been |
|
brought into compliance with the requirements of this chapter. The |
|
distributor shall inform the developers of the high-risk artificial |
|
intelligence system concerned and, where applicable, the |
|
deployers. |
|
Sec. 551.005. DEPLOYER DUTIES. A deployer of a high-risk |
|
artificial intelligence system shall use reasonable care to protect |
|
consumers from any known or reasonably foreseeable risks of |
|
algorithmic discrimination. If a deployer of a high-risk |
|
artificial intelligence system knows or has reason to know that a |
|
high-risk artificial intelligence system is not in compliance with |
|
any requirement in this chapter, it shall immediately suspend the |
|
use of the high-risk artificial intelligence system from the market |
|
until the system has been brought into compliance with the |
|
requirements of this chapter. The deployer shall inform the |
|
developers of the high-risk artificial intelligence system |
|
concerned and, where applicable, the distributors. |
|
Sec. 551.006. IMPACT ASSESSMENTS. (a) A deployer that |
|
deploys a high-risk artificial intelligence system shall complete |
|
an impact assessment for the high-risk artificial intelligence |
|
system. A deployer, or a third-party contracted by the deployer for |
|
such purposes, shall complete an impact assessment annually and |
|
within ninety days after any intentional and substantial |
|
modification to the high-risk artificial intelligence system is |
|
made available. An impact assessment must include, at a minimum, |
|
and to the extent reasonably known by or available to the deployer: |
|
(1) a statement by the deployer disclosing the |
|
purpose, intended use cases, and deployment context of, and |
|
benefits afforded by, the high-risk artificial intelligence |
|
system; |
|
(2) an analysis of whether the deployment of the |
|
high-risk artificial intelligence system poses any known or |
|
reasonably foreseeable risks of algorithmic discrimination and, if |
|
so, the nature of the algorithmic discrimination and the steps that |
|
have been taken to mitigate the risks; |
|
(3) a description of the categories of data the |
|
high-risk artificial intelligence system processes as inputs and |
|
the outputs the high-risk artificial intelligence system produces; |
|
(4) if the deployer used data to customize the |
|
high-risk artificial intelligence system, an overview of the |
|
categories of data the deployer used to customize the high-risk |
|
artificial intelligence system; |
|
(5) any metrics used to evaluate the performance and |
|
known limitations of the high-risk artificial intelligence system; |
|
(6) a description of any transparency measures taken |
|
concerning the high-risk artificial intelligence system, including |
|
any measures taken to disclose to a consumer that the high-risk |
|
artificial intelligence system will be used; |
|
(7) a description of the post-deployment monitoring |
|
and user safeguards provided concerning the high-risk artificial |
|
intelligence system, including the oversight, use, and learning |
|
process established by the deployer to address issues arising from |
|
the deployment of the high-risk artificial intelligence system; and |
|
(8) a description of cybersecurity measures and threat |
|
modeling conducted on the system. |
|
(b) Following an intentional and substantial modification |
|
to a high-risk artificial intelligence system, a deployer must |
|
disclose the extent to which the high-risk artificial intelligence |
|
system was used in a manner that was consistent with, or varied |
|
from, the developer's intended uses of the high-risk artificial |
|
intelligence system. |
|
(c) A single impact assessment may address a comparable set |
|
of high-risk artificial intelligence systems deployed by a |
|
deployer. |
|
(d) A deployer shall maintain the most recently completed |
|
impact assessment for a high-risk artificial intelligence system, |
|
all records concerning each impact assessment, and all prior impact |
|
assessments, if any, for at least three years following the final |
|
deployment of the high-risk artificial intelligence system. |
|
(e) If a deployer, or a third party contracted by the |
|
deployer, completes an impact assessment for the purpose of |
|
complying with another applicable law or regulation, such impact |
|
assessment shall be deemed to satisfy the requirements established |
|
in this subsection if such impact assessment is reasonably similar |
|
in scope and effect to the impact assessment that would otherwise be |
|
completed pursuant to this subsection. |
|
(f) A deployer may redact any trade secrets as defined by |
|
Section 541.001(33), Business & Commerce Code or information |
|
protected from disclosure by state or federal law. |
|
(g) Except as provided in subsection (e) of this section, a |
|
developer that makes a high-risk artificial intelligence system |
|
available to a deployer shall make available to the deployer the |
|
documentation and information necessary for a deployer to complete |
|
an impact assessment pursuant to this section. |
|
(h) A developer that also serves as a deployer for a |
|
high-risk artificial intelligence system is not required to |
|
generate and store an impact assessment unless the high-risk |
|
artificial intelligence system is provided to an unaffiliated |
|
deployer. |
|
Sec. 551.007. DISCLOSURE OF A HIGH-RISK ARTIFICIAL |
|
INTELLIGENCE SYSTEM TO CONSUMERS. (a) A deployer or developer that |
|
deploys, offers, sells, leases, licenses, gives, or otherwise makes |
|
available a high-risk artificial intelligence system that is |
|
intended to interact with consumers shall disclose to each |
|
consumer, before or at the time of interaction: |
|
(1) that the consumer is interacting with an |
|
artificial intelligence system; |
|
(2) the purpose of the system; |
|
(3) that the system may or will make a consequential |
|
decision affecting the consumer; |
|
(4) the nature of any consequential decision in which |
|
the system is or may be a substantial factor; |
|
(5) the factors to be used in making any consequential |
|
decisions; |
|
(6) contact information of the deployer; |
|
(7) a description of: |
|
(A) any human components of the system; |
|
(B) any automated components of the system; and |
|
(C) how human and automated components are used |
|
to inform a consequential decision; and |
|
(8) a declaration of the consumer's rights under |
|
Section 551.108. |
|
(b) Disclosure is required under subsection (a) of this |
|
section regardless of whether it would be obvious to a reasonable |
|
person that the person is interacting with an artificial |
|
intelligence system. |
|
(c) All disclosures under subsection (a) shall be clear and |
|
conspicuous and written in plain language, and avoid the use of a |
|
dark pattern as defined by 541.001, Business & Commerce Code. |
|
(d) All disclosures under subsection (a) may be linked to a |
|
separate webpage of the developer or deployer. |
|
(e) Any requirement in this section that may conflict with |
|
state or federal law may be exempt. |
|
Sec. 551.008. RISK IDENTIFICATION AND MANAGEMENT POLICY. |
|
(a) A developer or deployer of a high-risk artificial intelligence |
|
system shall, prior to deployment, assess potential risks of |
|
algorithmic discrimination and implement a risk management policy |
|
to govern the development or deployment of the high-risk artificial |
|
intelligence system. The risk management policy shall: |
|
(1) specify and incorporate the principles and |
|
processes that the developer or deployer uses to identify, |
|
document, and mitigate, in the development or deployment of a |
|
high-risk artificial intelligence system: |
|
(A) known or reasonably foreseeable risks of |
|
algorithmic discrimination; and |
|
(B) prohibited uses and unacceptable risks under |
|
Subchapter B; and |
|
(2) be reasonable in size, scope, and breadth, |
|
considering: |
|
(A) guidance and standards set forth in the most |
|
recent version of the "Artificial Intelligence Risk Management |
|
Framework: Generative Artificial Intelligence Profile" published |
|
by the National Institute of Standards and Technology; |
|
(B) any existing risk management guidance, |
|
standards or framework applicable to artificial intelligence |
|
systems designated by the Banking Commissioner or Insurance |
|
Commissioner, if the developer or deployer is regulated by the |
|
Department of Banking or Department of Insurance; |
|
(C) the size and complexity of the developer or |
|
deployer; |
|
(D) the nature, scope, and intended use of the |
|
high-risk artificial intelligence systems developed or deployed; |
|
and |
|
(E) the sensitivity and volume of personal data |
|
processed in connection with the high-risk artificial intelligence |
|
systems. |
|
(b) A risk management policy implemented pursuant to this |
|
section may apply to more than one high-risk artificial |
|
intelligence system developed or deployed, so long as the developer |
|
or deployer complies with all of the forgoing requirements and |
|
considerations in adopting and implementing the risk management |
|
policy with respect to each high-risk artificial intelligence |
|
system covered by the policy. |
|
(c) A developer or deployer may redact or omit any trade |
|
secrets as defined by Section 541.001(33), Business & Commerce Code |
|
or information protected from disclosure by state or federal law. |
|
Sec. 551.009. RELATIONSHIPS BETWEEN ARTIFICIAL |
|
INTELLIGENCE PARTIES. Any distributor or deployer, shall be |
|
considered to be a developer of a high-risk artificial intelligence |
|
system for the purposes of this chapter and shall be subject to the |
|
obligations and duties of a developer under this chapter in any of |
|
the following circumstances: |
|
(1) they put their name or trademark on a high-risk |
|
artificial intelligence system already placed in the market or put |
|
into service; |
|
(2) they intentionally and substantially modify a |
|
high-risk artificial intelligence system that has already been |
|
placed in the market or has already been put into service in such a |
|
way that it remains a high-risk artificial intelligence system |
|
under this chapter; or |
|
(3) they modify the intended purpose of an artificial |
|
intelligence system which has not previously been classified as |
|
high-risk and has already been placed in the market or put into |
|
service in such a way that the artificial intelligence system |
|
concerned becomes a high-risk artificial intelligence system in |
|
accordance with this chapter of a high-risk artificial intelligence |
|
system. |
|
Sec. 551.010. DIGITAL SERVICE PROVIDER AND SOCIAL MEDIA |
|
PLATFORM DUTIES REGARDING ARTIFICIAL INTELLIGENCE SYSTEMS. A |
|
digital service provider as defined by Section 509.001(2), Business & |
|
Commerce Code or a social media platform as defined by Section |
|
120.001(1), Business & Commerce Code, shall require advertisers on |
|
the service or platform to agree to terms preventing the deployment |
|
of a high-risk artificial intelligence system on the service or |
|
platform that could expose the users of the service or platform to |
|
algorithmic discrimination or prohibited uses under Subchapter B. |
|
Sec. 551.011. REPORTING REQUIREMENTS. (a) A deployer must |
|
notify, in writing, the council, the attorney general, or the |
|
director of the appropriate state agency that regulates the |
|
deployer's industry, and affected consumers as soon as practicable |
|
after the date on which the deployer discovers or is made aware that |
|
a deployed high-risk artificial intelligence system has caused |
|
algorithmic discrimination of an individual or group of |
|
individuals. |
|
(b) If a developer discovers or is made aware that a |
|
deployed high-risk artificial intelligence system is using inputs |
|
or providing outputs that constitute a violation of Subchapter B, |
|
the deployer must cease operation of the offending system as soon as |
|
technically feasible and provide notice to the council and the |
|
attorney general as soon as practicable and not later than the 10th |
|
day after the date on which the developer discovers or is made aware |
|
of the unacceptable risk. |
|
Sec. 551.012. SANDBOX PROGRAM EXCEPTION. (a) Excluding |
|
violations of Subchapter B, this chapter does not apply to the |
|
development of an artificial intelligence system that is used |
|
exclusively for research, training, testing, or other |
|
pre-deployment activities performed by active participants of the |
|
sandbox program in compliance with Chapter 552. |
|
SUBCHAPTER B. PROHIBITED USES AND UNACCEPTABLE RISK |
|
Sec. 551.051. MANIPULATION OF HUMAN BEHAVIOR TO CIRCUMVENT |
|
INFORMED DECISION-MAKING. An artificial intelligence system shall |
|
not be developed or deployed that uses subliminal techniques beyond |
|
a person's consciousness, or purposefully manipulative or |
|
deceptive techniques, with the objective or the effect of |
|
materially distorting the behavior of a person or a group of persons |
|
by appreciably impairing their ability to make an informed |
|
decision, thereby causing a person to make a decision that the |
|
person would not have otherwise made, in a manner that causes or is |
|
likely to cause significant harm to that person or another person or |
|
group of persons. |
|
Sec. 551.052. SOCIAL SCORING. An artificial intelligence |
|
system shall not be developed or deployed for the evaluation or |
|
classification of natural persons or groups of natural persons |
|
based on their social behavior or known, inferred, or predicted |
|
personal characteristics with the intent to determine a social |
|
score or similar categorical estimation or valuation of a person or |
|
groups of persons. |
|
Sec. 551.053. CAPTURE OF BIOMETRIC IDENTIFIERS USING |
|
ARTIFICIAL INTELLIGENCE. An artificial intelligence system |
|
developed with biometric identifiers of individuals and the |
|
targeted or untargeted gathering of images or other media from the |
|
internet or any other publicly available source shall not be |
|
deployed for the purpose of uniquely identifying a specific |
|
individual. An individual is not considered to be informed nor to |
|
have provided consent for such purpose pursuant to Section 503.001, |
|
Business and Commerce Code, based solely upon the existence on the |
|
internet, or other publicly available source, of an image or other |
|
media containing one or more biometric identifiers. |
|
Sec. 551.054. CATEGORIZATION BASED ON SENSITIVE |
|
ATTRIBUTES. An artificial intelligence system shall not be |
|
developed or deployed with the specific purpose of inferring or |
|
interpreting, sensitive personal attributes of a person or group of |
|
persons using biometric identifiers, except for the labeling or |
|
filtering of lawfully acquired biometric identifier data. |
|
Sec. 551.055. UTILIZATION OF PERSONAL ATTRIBUTES FOR HARM. |
|
An artificial intelligence system shall not utilize |
|
characteristics of a person or a specific group of persons based on |
|
their race, color, disability, religion, sex, national origin, age, |
|
or a specific social or economic situation, with the objective, or |
|
the effect, of materially distorting the behavior of that person or |
|
a person belonging to that group in a manner that causes or is |
|
reasonably likely to cause that person or another person harm. |
|
Sec. 551.056. CERTAIN SEXUALLY EXPLICIT VIDEOS, IMAGES, AND |
|
CHILD PORNOGRAPHY. An artificial intelligence system shall not be |
|
developed or deployed that produces, assists, or aids in producing, |
|
or is capable of producing unlawful visual material in violation of |
|
Section 43.26, Penal Code or an unlawful deep fake video or image in |
|
violation of Section 21.165, Penal Code. |
|
SUBCHAPTER C. ENFORCEMENT AND CONSUMER PROTECTIONS |
|
Sec. 551.101. CONSTRUCTION AND APPLICATION. (a) This |
|
chapter shall be broadly construed and applied to promote its |
|
underlying purposes, which are: |
|
(1) to facilitate and advance the responsible |
|
development and use of artificial intelligence systems; |
|
(2) to protect individuals and groups of individuals |
|
from known, and unknown but reasonably foreseeable, risks, |
|
including unlawful algorithmic discrimination; |
|
(3) to provide transparency regarding those risks in |
|
the development, deployment, or use of artificial intelligence |
|
systems; and |
|
(4) to provide reasonable notice regarding the use or |
|
considered use of artificial intelligence systems by state |
|
agencies. |
|
(b) this chapter does not apply to the developer of an open |
|
source artificial intelligence system, provided that: |
|
(1) the system is not deployed as a high-risk |
|
artificial intelligence system and the developer has taken |
|
reasonable steps to ensure that the system cannot be used as a |
|
high-risk artificial intelligence system without substantial |
|
modifications; and |
|
(2) the weights and technical architecture of the |
|
system are made publicly available. |
|
Sec. 551.102. ENFORCEMENT AUTHORITY. The attorney general |
|
has authority to enforce this chapter. Excluding violations of |
|
Subchapter B, researching, training, testing, or the conducting of |
|
other pre-deployment activities by active participants of the |
|
sandbox program, in compliance with Chapter 552, does not subject a |
|
developer or deployer to penalties or actions. |
|
Sec. 551.103. INTERNET WEBSITE AND COMPLAINT MECHANISM. |
|
The attorney general shall post on the attorney general's Internet |
|
website: |
|
(1) information relating to: |
|
(A) the responsibilities of a developer, |
|
distributor, and deployer under Subchapter A; and |
|
(B) an online mechanism through which a consumer |
|
may submit a complaint under this chapter to the attorney general. |
|
Sec. 551.104. INVESTIGATIVE AUTHORITY. (a) If the |
|
attorney general has reasonable cause to believe that a person has |
|
engaged in or is engaging in a violation of this chapter, the |
|
attorney general may issue a civil investigative demand. The |
|
attorney general shall issue such demands in accordance with and |
|
under the procedures established under Section 15.10. |
|
(b) The attorney general may request, pursuant to a civil |
|
investigative demand issued under Subsection (a), that a developer |
|
or deployer of a high-risk artificial intelligence system disclose |
|
their risk management policy and impact assessments required under |
|
Subchapter A. The attorney general may evaluate the risk |
|
management policy and impact assessments for compliance with the |
|
requirements set forth in Subchapter A. |
|
(c) The attorney general may not institute an action for a |
|
civil penalty against a developer or deployer for artificial |
|
intelligence systems that remain isolated from customer |
|
interaction in a pre-deployment environment. |
|
Sec. 551.105. NOTICE OF VIOLATION OF CHAPTER; OPPORTUNITY |
|
TO CURE. Before bringing an action under Section 551.106, the |
|
attorney general shall notify a developer, distributor, or deployer |
|
in writing, not later than the 30th day before bringing the action, |
|
identifying the specific provisions of this chapter the attorney |
|
general alleges have been or are being violated. The attorney |
|
general may not bring an action against the developer or deployer |
|
if: |
|
(1) within the 30-day period, the developer or |
|
deployer cures the identified violation; and |
|
(2) the developer or deployer provides the attorney |
|
general a written statement that the developer or deployer: |
|
(A) cured the alleged violation; |
|
(B) notified the consumer, if technically |
|
feasible, and the council that the developer or deployer's |
|
violation was addressed, if the consumer's contact information has |
|
been made available to the developer or deployer and the attorney |
|
general; |
|
(C) provided supportive documentation to show |
|
how the violation was cured; and |
|
(D) made changes to internal policies, if |
|
necessary, to reasonably ensure that no such further violations are |
|
likely to occur. |
|
Sec. 551.106. CIVIL PENALTY; INJUNCTION. (a) The attorney |
|
general may bring an action in the name of this state to restrain or |
|
enjoin the person from violating this chapter and seek injunctive |
|
relief. |
|
(b) The attorney general may recover reasonable attorney's |
|
fees and other reasonable expenses incurred in investigating and |
|
bringing an action under this section. |
|
(c) The attorney general may assess and collect an |
|
administrative fine against a developer or deployer who fails to |
|
timely cure a violation or who breaches a written statement |
|
provided to the attorney general, other than those for a prohibited |
|
use, of not less than $50,000 and not more than $100,000 per uncured |
|
violation. |
|
(d) The attorney general may assess and collect an |
|
administrative fine against a developer or deployer who fails to |
|
timely cure a violation of a prohibited use, or whose violation is |
|
determined to be uncurable, of not less than $80,000 and not more |
|
than $200,000 per violation. |
|
(e) A developer or deployer who was found in violation of |
|
and continues to operate with the provisions of this chapter shall |
|
be assessed an administrative fine of not less than $2,000 and not |
|
more than $40,000 per day. |
|
(f) There is a rebuttable presumption that a developer, |
|
distributor, or deployer used reasonable care as required under |
|
this chapter if the developer, distributor, or deployer complied |
|
with their duties under Subchapter A. |
|
Sec. 551.107. ENFORCEMENT ACTIONS BY STATE AGENCIES. A |
|
state agency may sanction an individual licensed, registered, or |
|
certified by that agency for violations of Subchapter B, including: |
|
(1) the suspension, probation, or revocation of a |
|
license, registration, certificate, or other form of permission to |
|
engage in an activity; and |
|
(2) monetary penalties up to $100,000. |
|
Sec. 551.108. CONSUMER RIGHTS AND REMEDIES. A consumer may |
|
appeal a consequential decision made by a high-risk artificial |
|
intelligence system which has an adverse impact on their health, |
|
safety, or fundamental rights, and shall have the right to obtain |
|
from the deployer clear and meaningful explanations of the role of |
|
the high-risk artificial intelligence system in the |
|
decision-making procedure and the main elements of the decision |
|
taken. |
|
SUBCHAPTER D. CONSTRUCTION OF CHAPTER; LOCAL PREEMPTION |
|
Sec. 551.151. CONSTRUCTION OF CHAPTER. This chapter may |
|
not be construed as imposing a requirement on a developer, a |
|
deployer, or other person that adversely affects the rights or |
|
freedoms of any person, including the right of free speech. |
|
Sec. 551.152. LOCAL PREEMPTION. This chapter supersedes |
|
and preempts any ordinance, resolution, rule, or other regulation |
|
adopted by a political subdivision regarding the use of high-risk |
|
artificial intelligence systems. |
|
CHAPTER 552. ARTIFICIAL INTELLIGENCE REGULATORY SANDBOX PROGRAM |
|
SUBCHAPTER A. GENERAL PROVISIONS |
|
Sec. 552.001. DEFINITIONS. In this chapter: |
|
(1) "Applicable agency" means a state agency |
|
responsible for regulating a specific sector impacted by an |
|
artificial intelligence system. |
|
(2) "Consumer" means a person who engages in |
|
transactions involving an artificial intelligence system or is |
|
directly affected by the use of such a system. |
|
(3) "Council" means the Artificial Intelligence |
|
Council established by Chapter 553. |
|
(4) "Department" means the Texas Department of |
|
Information Resources. |
|
(5) "Program participant" means a person or business |
|
entity approved to participate in the sandbox program. |
|
(6) "Sandbox program" means the regulatory framework |
|
established under this chapter that allows temporary testing of |
|
artificial intelligence systems in a controlled, limited manner |
|
without full regulatory compliance. |
|
SUBCHAPTER B. SANDBOX PROGRAM FRAMEWORK |
|
Sec. 552.051. ESTABLISHMENT OF SANDBOX PROGRAM. (a) The |
|
department, in coordination with the council, shall administer the |
|
Artificial Intelligence Regulatory Sandbox Program to facilitate |
|
the development, testing, and deployment of innovative artificial |
|
intelligence systems in Texas. |
|
(b) The sandbox program is designed to: |
|
(1) promote the safe and innovative use of artificial |
|
intelligence across various sectors including healthcare, finance, |
|
education, and public services; |
|
(2) encourage the responsible deployment of |
|
artificial intelligence systems while balancing the need for |
|
consumer protection, privacy, and public safety; and |
|
(3) provide clear guidelines for artificial |
|
intelligence developers to test systems while temporarily exempt |
|
from certain regulatory requirements. |
|
Sec. 552.052. APPLICATION PROCESS. (a) A person or |
|
business entity seeking to participate in the sandbox program must |
|
submit an application to the council. |
|
(b) The application must include: |
|
(1) a detailed description of the artificial |
|
intelligence system and its intended use; |
|
(2) a risk assessment that addresses potential impacts |
|
on consumers, privacy, and public safety; |
|
(3) a plan for mitigating any adverse consequences |
|
during the testing phase; and |
|
(4) proof of compliance with federal artificial |
|
intelligence laws and regulations, where applicable. |
|
Sec. 552.053. DURATION AND SCOPE OF PARTICIPATION. A |
|
participant may test an artificial intelligence system under the |
|
sandbox program for a period of up to 36 months, unless extended by |
|
the department for good cause. |
|
SUBCHAPTER C. OVERSIGHT AND COMPLIANCE |
|
Sec. 552.101. AGENCY COORDINATION. (a) The department |
|
shall coordinate with all relevant state regulatory agencies to |
|
oversee the operations of the sandbox participants. |
|
(b) A relevant agency may recommend to the department that a |
|
participant's sandbox privileges be revoked if the artificial |
|
intelligence system: |
|
(1) poses undue risk to public safety or welfare; |
|
(2) violates any federal or state laws that the |
|
sandbox program cannot override. |
|
Sec. 552.102. REPORTING REQUIREMENTS. (a) Each sandbox |
|
participant must submit quarterly reports to the department, which |
|
shall include: |
|
(1) system performance metrics; |
|
(2) updates on how the system mitigates any risks |
|
associated with its operation; and |
|
(3) feedback from consumers and affected stakeholders |
|
that are using a product that has been deployed from this section. |
|
(b) The department must submit an annual report to the |
|
legislature detailing: |
|
(1) the number of participants in the sandbox program; |
|
(2) the overall performance and impact of artificial |
|
intelligence systems tested within the program; and |
|
(3) recommendations for future legislative or |
|
regulatory reforms. |
|
CHAPTER 553. TEXAS ARTIFICIAL INTELLIGENCE COUNCIL |
|
SUBCHAPTER A. CREATION AND ORGANIZATION OF COUNCIL |
|
Sec. 553.001. CREATION OF COUNCIL. (a) The Artificial |
|
Intelligence Council is administratively attached to the office of |
|
the governor, and the office of the governor shall provide |
|
administrative support to the council as provided by this section. |
|
(b) The office of the governor and the council shall enter |
|
into a memorandum of understanding detailing: |
|
(1) the administrative support the council requires |
|
from the office of the governor to fulfill the purposes of this |
|
chapter; |
|
(2) the reimbursement of administrative expenses to |
|
the office of the governor; and |
|
(3) any other provisions available by law to ensure |
|
the efficient operation of the council as attached to the office of |
|
the governor. |
|
(c) The purpose of the council is to: |
|
(1) ensure artificial intelligence systems are |
|
ethical and in the public's best interest and do not harm public |
|
safety or undermine individual freedoms by finding gaps in the |
|
Penal Code and Chapter 82, Civil Practice and Remedies Code and |
|
making recommendations to the Legislature. |
|
(2) identify existing laws and regulations that impede |
|
innovation in artificial intelligence development and recommend |
|
appropriate reforms; |
|
(3) analyze opportunities to improve the efficiency |
|
and effectiveness of state government operations through the use of |
|
artificial intelligence systems; |
|
(4) investigate and evaluate potential instances of |
|
regulatory capture, including undue influence by technology |
|
companies or disproportionate burdens on smaller innovators; |
|
(5) investigate and evaluate the influence of |
|
technology companies on other companies and determine the existence |
|
or use of tools or processes designed to censor competitors or |
|
users; and |
|
(6) offer guidance and recommendations to state |
|
agencies including advisory opinions on the ethical and legal use |
|
of artificial intelligence; |
|
Sec. 553.002. COUNCIL MEMBERSHIP. (a) The council is |
|
composed of 10 members as follows: |
|
(1) four members of the public appointed by the |
|
governor; |
|
(2) two members of the public appointed by the |
|
lieutenant governor; |
|
(3) two members of the public appointed by the speaker |
|
of the house of representatives; |
|
(4) one senator appointed by the lieutenant governor |
|
as a nonvoting member; and |
|
(5) one member of the house of representatives |
|
appointed by the speaker of the house of representatives as a |
|
nonvoting member. |
|
(b) Voting members of the council serve staggered four-year |
|
terms, with the terms of four members expiring every two years. |
|
(c) The governor shall appoint a chair from among the |
|
members, and the council shall elect a vice chair from its |
|
membership. |
|
(d) The council may establish an advisory board composed of |
|
individuals from the public who possess expertise directly related |
|
to the council's functions, including technical, ethical, |
|
regulatory, and other relevant areas. |
|
Sec. 553.003. QUALIFICATIONS. (a) Members of the council |
|
must be Texas residents and have knowledge or expertise in one or |
|
more of the following areas: |
|
(1) artificial intelligence technologies; |
|
(2) data privacy and security; |
|
(3) ethics in technology or law; |
|
(4) public policy and regulation; or |
|
(5) risk management or safety related to artificial |
|
intelligence systems. |
|
(b) Members must not hold an office or profit under the |
|
state or federal government at the time of appointment. |
|
Sec. 553.004. STAFF AND ADMINISTRATION. The council may |
|
employ an executive director and other personnel as necessary to |
|
perform its duties. |
|
SUBCHAPTER B. POWERS AND DUTIES OF THE COUNCIL |
|
Sec. 553.101. ISSUANCE OF ADVISORY OPINIONS. (a) A state |
|
agency may request a written advisory opinion from the council |
|
regarding the use of artificial intelligence systems in the state. |
|
(b) The council may issue advisory opinions on state use of |
|
artificial intelligence systems regarding: |
|
(1) the compliance of artificial intelligence systems |
|
with Texas law; |
|
(2) the ethical implications of artificial |
|
intelligence deployments in the state; |
|
(3) data privacy and security concerns related to |
|
artificial intelligence systems; or |
|
(4) potential liability or legal risks associated with |
|
the use of AI. |
|
Sec. 553.102. RULEMAKING AUTHORITY. (a) The council may |
|
adopt rules necessary to administer its duties under this chapter, |
|
including: |
|
(1) procedures for requesting advisory opinions; |
|
(2) standards for ethical artificial intelligence |
|
development and deployment; |
|
(3) guidelines for evaluating the safety, privacy, and |
|
fairness of artificial intelligence systems. |
|
(b) The council's rules shall align with state laws on |
|
artificial intelligence, technology, data security, and consumer |
|
protection. |
|
Sec. 553.103. TRAINING AND EDUCATIONAL OUTREACH. The |
|
council shall conduct training programs for state agencies and |
|
local governments on the ethical use of artificial intelligence |
|
systems. |
|
SECTION 3. Section 503.001, Business & Commerce Code is |
|
amended by adding Subsection (c-3) to read as follows: |
|
(c-3) This section does not apply to the training, |
|
processing, or storage of biometric identifiers involved in machine |
|
learning or artificial intelligence systems, unless performed for |
|
the purpose of uniquely identifying a specific individual. If a |
|
biometric identifier captured for the purpose of training an |
|
artificial intelligence system is subsequently used for a |
|
commercial purpose, the person possessing the biometric identifier |
|
is subject to this section's provisions for the possession and |
|
destruction of a biometric identifier and the associated penalties. |
|
SECTION 4. Sections 541.051(b), 541.101(a), 541.102(a), |
|
and Sec.541.104(a), Business & Commerce Code, are amended to read |
|
as follows: |
|
Sec. 541.051. CONSUMER'S PERSONAL DATA RIGHTS; REQUEST TO |
|
EXERCISE RIGHTS. (a) A consumer is entitled to exercise the |
|
consumer rights authorized by this section at any time by |
|
submitting a request to a controller specifying the consumer rights |
|
the consumer wishes to exercise. With respect to the processing of |
|
personal data belonging to a known child, a parent or legal guardian |
|
of the child may exercise the consumer rights on behalf of the |
|
child. |
|
(b) A controller shall comply with an authenticated |
|
consumer request to exercise the right to: |
|
(1) confirm whether a controller is processing the |
|
consumer's personal data and to access the personal data; |
|
(2) correct inaccuracies in the consumer's personal |
|
data, taking into account the nature of the personal data and the |
|
purposes of the processing of the consumer's personal data; |
|
(3) delete personal data provided by or obtained about |
|
the consumer; |
|
(4) if the data is available in a digital format, |
|
obtain a copy of the consumer's personal data that the consumer |
|
previously provided to the controller in a portable and, to the |
|
extent technically feasible, readily usable format that allows the |
|
consumer to transmit the data to another controller without |
|
hindrance; [or] |
|
(5) know if the consumer's personal data is or will be |
|
used in any artificial intelligence system and for what purposes; |
|
or |
|
([5]6) opt out of the processing of the personal data |
|
for purposes of: |
|
(A) targeted advertising; |
|
(B) the sale of personal data; [or] |
|
(C) the sale of personal data for use in |
|
artificial intelligence systems prior to being collected; or |
|
([C]D) profiling in furtherance of a decision |
|
that produces a legal or similarly significant effect concerning |
|
the consumer. |
|
Sec. 541.101. CONTROLLER DUTIES; TRANSPARENCY. (a) A |
|
controller: |
|
(1) shall limit the collection of personal data to |
|
what is adequate, relevant, and reasonably necessary in relation to |
|
the purposes for which that personal data is processed, as |
|
disclosed to the consumer; [and] |
|
(2) for purposes of protecting the confidentiality, |
|
integrity, and accessibility of personal data, shall establish, |
|
implement, and maintain reasonable administrative, technical, and |
|
physical data security practices that are appropriate to the volume |
|
and nature of the personal data at issue.; and |
|
(3) for purposes of protecting the unauthorized |
|
access, disclosure, alteration, or destruction of data collected, |
|
stored, and processed by artificial intelligence systems, shall |
|
establish, implement, and maintain, reasonable administrative, |
|
technical, and physical data security practices that are |
|
appropriate to the volume and nature of the data collected, stored, |
|
and processed by artificial intelligence systems. |
|
Sec.541.102. PRIVACY NOTICE. (a) A controller shall |
|
provide consumers with a reasonably accessible and clear privacy |
|
notice that includes: |
|
(1) the categories of personal data processed by the |
|
controller, including, if applicable, any sensitive data processed |
|
by the controller; |
|
(2) the purpose for processing personal data; |
|
(3) how consumers may exercise their consumer rights |
|
under Subchapter B, including the process by which a consumer may |
|
appeal a controller's decision with regard to the consumer's |
|
request; |
|
(4) if applicable, the categories of personal data |
|
that the controller shares with third parties; |
|
(5) if applicable, the categories of third parties |
|
with whom the controller shares personal data; [and] |
|
(6) if applicable, an acknowledgement of the |
|
collection, use, and sharing of personal data for artificial |
|
intelligence purposes; and |
|
([6]7) a description of the methods required under |
|
Section 541.055 through which consumers can submit requests to |
|
exercise their consumer rights under this chapter. |
|
Sec. 541.104. DUTIES OF PROCESSOR. (a) A processor shall |
|
adhere to the instructions of a controller and shall assist the |
|
controller in meeting or complying with the controller's duties or |
|
requirements under this chapter, including: |
|
(1) assisting the controller in responding to consumer |
|
rights requests submitted under Section 541.051 by using |
|
appropriate technical and organizational measures, as reasonably |
|
practicable, taking into account the nature of processing and the |
|
information available to the processor; |
|
(2) assisting the controller with regard to complying |
|
with the [requirement]requirements relating to the security of |
|
processing personal data, and if applicable, the data collected, |
|
stored, and processed by artificial intelligence systems and to the |
|
notification of a breach of security of the processor's system |
|
under Chapter 521, taking into account the nature of processing and |
|
the information available to the processor; and |
|
(3) providing necessary information to enable the |
|
controller to conduct and document data protection assessments |
|
under Section 541.105. |
|
SECTION 5. Subtitle E, Title 4, Labor Code, is amended by |
|
adding Chapter 319 to read as follows: |
|
CHAPTER 319. TEXAS ARTIFICIAL INTELLIGENCE WORKFORCE DEVELOPMENT |
|
GRANT PROGRAM |
|
SUBCHAPTER A. GENERAL PROVISIONS |
|
Sec. 319.001. DEFINITIONS. In this chapter: |
|
(1) "Artificial intelligence industry" means |
|
businesses, research organizations, governmental entities, and |
|
educational institutions engaged in the development, deployment, |
|
or use of artificial intelligence technologies in Texas. |
|
(2) "Commission" means the Texas Workforce |
|
Commission. |
|
(3) "Eligible entity" means Texas-based businesses in |
|
the artificial intelligence industry, public school districts, |
|
community colleges, public technical institutes, and workforce |
|
development organizations. |
|
(4) "Program" means the Texas Artificial Intelligence |
|
Workforce Development Grant Program established under this |
|
chapter. |
|
SUBCHAPTER B. ARTIFICIAL INTELLIGENCE WORKFORCE DEVELOPMENT GRANT |
|
PROGRAM |
|
Sec. 319.051. ESTABLISHMENT OF GRANT PROGRAM. (a) The |
|
commission shall establish the Texas Artificial Intelligence |
|
Workforce Development Grant Program to: |
|
(1) support and assist Texas-based artificial |
|
intelligence companies in developing a skilled workforce; |
|
(2) provide grants to local community colleges and |
|
public high schools to implement or expand career and technical |
|
education programs focused on artificial intelligence readiness |
|
and skill development; and |
|
(3) offer opportunities to retrain and reskill workers |
|
through partnerships with the artificial intelligence industry and |
|
workforce development programs. |
|
(b) The program is intended to: |
|
(1) prepare Texas workers and students for employment |
|
in the rapidly growing artificial intelligence industry; |
|
(2) support the creation of postsecondary programs and |
|
certifications relevant to current artificial intelligence |
|
opportunities; |
|
(3) ensure that Texas maintains a competitive edge in |
|
artificial intelligence innovation and workforce development; and |
|
(4) address workforce gaps in artificial |
|
intelligence-related fields, including data science, |
|
cybersecurity, machine learning, robotics, and automation. |
|
(c) The commission shall adopt rules necessary to implement |
|
this subchapter. |
|
Sec. 319.052. FEDERAL FUNDS AND GIFTS, GRANTS, AND |
|
DONATIONS. |
|
In addition to other money appropriated by the legislature, |
|
for the purpose of providing artificial intelligence workforce |
|
opportunities under the program established under this subchapter |
|
the commission may: |
|
(1) seek and apply for any available federal funds; |
|
and |
|
(2) solicit and accept gifts, grants, and donations |
|
from any other source, public or private, as necessary to ensure |
|
effective implementation of the program. |
|
Sec. 319.053. ELIGIBILITY FOR GRANTS. (a) The following |
|
entities are eligible to apply for grants under this program: |
|
(1) Texas-based businesses engaged in the development |
|
or deployment of artificial intelligence technologies; |
|
(2) public school districts and charter schools |
|
offering or seeking to offer career and technical education |
|
programs in artificial intelligence-related fields or to update |
|
existing curricula to address these fields; |
|
(3) public community colleges and technical |
|
institutes that develop artificial intelligence-related curricula |
|
or training programs or update existing curricula or training |
|
programs to incorporate artificial intelligence training; and |
|
(4) workforce development organizations in |
|
partnership with artificial intelligence companies to reskill and |
|
retrain workers in artificial intelligence competencies. |
|
(b) To be eligible, the entity must: |
|
(1) submit an application to the commission in the |
|
form and manner prescribed by the commission; and |
|
(2) demonstrate the capacity to develop and implement |
|
training, educational, or workforce development programs that |
|
align with the needs of the artificial intelligence industry in |
|
Texas and lead to knowledge, skills, and work-based experiences |
|
that are transferable to similar employment opportunities in the |
|
artificial intelligence industry. |
|
Sec. 319.054. USE OF GRANTS. (a) Grants awarded under the |
|
program may be used for: |
|
(1) developing or expanding workforce training |
|
programs for artificial intelligence-related skills, including but |
|
not limited to machine learning, data analysis, software |
|
development, and robotics; |
|
(2) creating or enhancing career and technical |
|
education programs in artificial intelligence for high school |
|
students, with a focus on preparing them for careers in artificial |
|
intelligence or related fields; |
|
(3) providing financial support for instructors, |
|
equipment, and technology necessary for artificial |
|
intelligence-related workforce training; |
|
(4) partnering with local businesses to develop |
|
internship programs, on-the-job training opportunities, instructor |
|
externships, and apprenticeships in the artificial intelligence |
|
industry; |
|
(5) funding scholarships or stipends for students, |
|
instructors, and workers participating in artificial intelligence |
|
training programs, particularly for individuals from underserved |
|
or underrepresented communities; or |
|
(6) reskilling and retraining workers displaced by |
|
technological changes or job automation, with an emphasis on |
|
artificial intelligence-related job roles. |
|
(b) The commission shall prioritize funding for: |
|
(1) initiatives that partner with rural and |
|
underserved communities to promote artificial intelligence |
|
education and career pathways; |
|
(2) programs that lead to credentials of value in |
|
artificial intelligence or related fields; and |
|
(3) proposals that include partnerships between the |
|
artificial intelligence industry, a public or private institution |
|
of higher education in this state, and workforce development |
|
organizations. |
|
SECTION 6. Section 325.011, Government Code, is amended to |
|
read as follows: |
|
Sec. 325.011. CRITERIA FOR REVIEW. The commission and its |
|
staff shall consider the following criteria in determining whether |
|
a public need exists for the continuation of a state agency or its |
|
advisory committees or for the performance of the functions of the |
|
agency or its advisory committees: |
|
(1) the efficiency and effectiveness with which the |
|
agency or the advisory committee operates; |
|
(2)(A) an identification of the mission, goals, and |
|
objectives intended for the agency or advisory committee and of the |
|
problem or need that the agency or advisory committee was intended |
|
to address; and |
|
(B) the extent to which the mission, goals, and |
|
objectives have been achieved and the problem or need has been |
|
addressed; |
|
(3)(A) an identification of any activities of the |
|
agency in addition to those granted by statute and of the authority |
|
for those activities; and |
|
(B) the extent to which those activities are |
|
needed; |
|
(4) an assessment of authority of the agency relating |
|
to fees, inspections, enforcement, and penalties; |
|
(5) whether less restrictive or alternative methods of |
|
performing any function that the agency performs could adequately |
|
protect or provide service to the public; |
|
(6) the extent to which the jurisdiction of the agency |
|
and the programs administered by the agency overlap or duplicate |
|
those of other agencies, the extent to which the agency coordinates |
|
with those agencies, and the extent to which the programs |
|
administered by the agency can be consolidated with the programs of |
|
other state agencies; |
|
(7) the promptness and effectiveness with which the |
|
agency addresses complaints concerning entities or other persons |
|
affected by the agency, including an assessment of the agency's |
|
administrative hearings process; |
|
(8) an assessment of the agency's rulemaking process |
|
and the extent to which the agency has encouraged participation by |
|
the public in making its rules and decisions and the extent to which |
|
the public participation has resulted in rules that benefit the |
|
public; |
|
(9) the extent to which the agency has complied with: |
|
(A) federal and state laws and applicable rules |
|
regarding equality of employment opportunity and the rights and |
|
privacy of individuals; and |
|
(B) state law and applicable rules of any state |
|
agency regarding purchasing guidelines and programs for |
|
historically underutilized businesses; |
|
(10) the extent to which the agency issues and |
|
enforces rules relating to potential conflicts of interest of its |
|
employees; |
|
(11) the extent to which the agency complies with |
|
Chapters 551 and 552 and follows records management practices that |
|
enable the agency to respond efficiently to requests for public |
|
information; |
|
(12) the effect of federal intervention or loss of |
|
federal funds if the agency is abolished; |
|
(13) the extent to which the purpose and effectiveness |
|
of reporting requirements imposed on the agency justifies the |
|
continuation of the requirement; [and] |
|
(14) an assessment of the agency's cybersecurity |
|
practices using confidential information available from the |
|
Department of Information Resources or any other appropriate state |
|
agency; and |
|
(15) an assessment, using information available from |
|
the Department of Information Resources, the Attorney General, or |
|
any other appropriate state agency, of the agency's use of |
|
artificial intelligence systems, high-risk artificial intelligence |
|
systems, in its operations and its oversight of the use of |
|
artificial intelligence systems by entities or persons under the |
|
agency's jurisdiction, and any related impact on the agency's |
|
ability to achieve its mission, goals, and objectives. |
|
SECTION 7. Section 2054.068(b), Government Code, is amended |
|
to read as follows: |
|
(b) The department shall collect from each state agency |
|
information on the status and condition of the agency's information |
|
technology infrastructure, including information regarding: |
|
(1) the agency's information security program; |
|
(2) an inventory of the agency's servers, mainframes, |
|
cloud services, and other information technology equipment; |
|
(3) identification of vendors that operate and manage |
|
the agency's information technology infrastructure; [and] |
|
(4) any additional related information requested by |
|
the department; and |
|
(5) an evaluation of the use, or considered use, of |
|
artificial intelligence systems and high-risk artificial |
|
intelligence systems by each state agency. |
|
SECTION 8. Section 2054.0965(b), Government Code, is |
|
amended to read as follows: |
|
Sec. 2054.0965. INFORMATION RESOURCES DEPLOYMENT REVIEW. |
|
(b) Except as otherwise modified by rules adopted by the |
|
department, the review must include: |
|
(1) an inventory of the agency's major information |
|
systems, as defined by Section 2054.008, and other operational or |
|
logistical components related to deployment of information |
|
resources as prescribed by the department; |
|
(2) an inventory of the agency's major databases, |
|
artificial intelligence systems, and applications; |
|
(3) a description of the agency's existing and planned |
|
telecommunications network configuration; |
|
(4) an analysis of how information systems, |
|
components, databases, applications, and other information |
|
resources have been deployed by the agency in support of: |
|
(A) applicable achievement goals established |
|
under Section 2056.006 and the state strategic plan adopted under |
|
Section 2056.009; |
|
(B) the state strategic plan for information |
|
resources; and |
|
(C) the agency's business objectives, mission, |
|
and goals; |
|
(5) agency information necessary to support the state |
|
goals for interoperability and reuse; and |
|
(6) confirmation by the agency of compliance with |
|
state statutes, rules, and standards relating to information |
|
resources. |
|
SECTION 9. Not later than September 1, 2025, the attorney |
|
general shall post on the attorney general's Internet website the |
|
information and online mechanism required by Section 551.041, |
|
Business & Commerce Code, as added by this Act. |
|
SECTION 10. This Act takes effect September 1, 2025. |