New AI Regulations Come into Play with the Texas Responsible Artificial Intelligence Governance Act

Blank Rome LLP
Contact

Blank Rome LLP

The rapid advancement of artificial intelligence (“AI”) has outpaced existing U.S. regulatory frameworks. At present, AI regulation occurs primarily on a state-by-state basis. Most states that have enacted AI laws rely on targeted regulations for particular use cases or fields. Texas has established one of the more comprehensive approaches with its Texas Responsible Artificial Intelligence Governance Act (“TRAIGA”). TRAIGA was signed into law on June 22, 2025, and took effect on January 1, 2026, and has implications beyond the borders of Texas.  

TRAIGA addresses developing or deploying AI systems, and prohibits the following: 

  1. Developing or deploying an AI system with the intent to manipulate human behavior to incite/encourage self‑harm, harm to others, or criminal activity.
  2. Developing or deploying an AI system with the sole intent to infringe, restrict, or impair rights guaranteed under the Constitution.
  3. Developing or deploying an AI system with the intent to unlawfully discriminate against a protected class in violation of state or federal law.
  4. Developing or deploying an AI system with the sole intent of producing or distributing certain sexually explicit content. 

The Texas Business & Commerce Code defines AI systems as “any machine-base system that, for any explicit or implicit objective, infers from the inputs the system receives how to generate outputs, including content, decisions, predictions, or recommendations, that can influence physical or virtual environments.” TRAIGA applies to any person or entity who does business in Texas or with Texans, thereby expanding its reach far beyond the Texas border. Another point of interest is that TRAIGA focuses on both development and deployment. In doing so, the law affects not only AI developers, but any entity that may use such AI.

The emphasis on intent represents a significant shift from EU-style risk-based assessments, which have also been adopted in Colorado. By taking this approach, TRAIGA offers a clear and potentially more easily operationalized framework compared to traditional risk-based methodologies.

TRAIGA also creates a state advisory body in the form of an AI Counsel to provide oversight and guidance. In addition, TRAIGA creates a regulatory sandbox program, in which companies can test AI systems in a controlled environment for 36 months, while being protected from certain types of prosecution.

Enforcement

The Texas Attorney General has the exclusive right to bring actions under TRAIGA, as TRAIGA provides no private right of action. The Attorney General must provide notice and an opportunity to cure before bringing an action, and penalties range from $10,000 to $200,000 per violation, depending in part on whether the violation is determined to be “curable,” or $2,000 to $40,000 per day for a continued violation.

How Does TRAIGA Compare to Other States?

Colorado has enacted comprehensive AI legislation through the Colorado AI Act (the “CAIA”). This statute implements a risk-based framework, requiring developers and deployers to conduct impact assessments, manage risks, and provide clear consumer notifications. Thus, the CAIA includes more roadblocks to compliance than TRAIGA. Utah’s Utah AI Policy Act primarily addresses consumer notification and deceptive practices, resulting in a more limited scope than TRAIGA. California has adopted several targeted regulations addressing specific AI applications, such as chatbot oversight, election integrity measures, and deepfake restrictions. Texas, meanwhile, has established AI-related laws that are more straightforward compared to those in California.

Practical Steps

Any entity conducting business in Texas or with Texas residents should carefully assess their risk exposure and review their business policies accordingly. In general, organizations are advised to consider the following points.

  1. Map your Texas exposure: Identify AI systems developed, offered, or deployed in Texas.
  2. Update AI policies to expressly forbid AI uses that could manipulate self‑harm/violence/crime, discriminate intentionally, violate rights, or produce child sexual content.
  3. Evaluate the Texas sandbox: Assess if it is appropriate to pilot your next-generation AI features within the Texas sandbox to mitigate regulatory risk.

[View source.]

DISCLAIMER: Because of the generality of this update, the information provided herein may not be applicable in all situations and should not be acted upon without specific legal advice based on particular situations. Attorney Advertising.

© Blank Rome LLP

Written by:

Blank Rome LLP
Contact
more
less

What do you want from legal thought leadership?

Please take our short survey – your perspective helps to shape how firms create relevant, useful content that addresses your needs:

Blank Rome LLP on:

Reporters on Deadline

"My best business intelligence, in one easy email…"

Your first step to building a free, personalized, morning email brief covering pertinent authors and topics on JD Supra:
*By using the service, you signify your acceptance of JD Supra's Privacy Policy.
Custom Email Digest
- hide
- hide