AI & You: Key Considerations for a Company’s Use of Artificial Intelligence Systems

Fox Rothschild LLP
Contact

Fox Rothschild LLP

Businesses increasingly rely on technology containing Artificial Intelligence (AI) to streamline operations, enhance employee experiences and drive strategic decision-making. The integration of AI into work systems, including Human Resources Information Systems (HRIS) and other platforms presents, presents both unprecedented opportunities and challenges, though. As companies embrace AI, it is crucial to navigate legal and ethical considerations effectively.

This article provides actionable insights for companies to evaluate AI systems.

Regulations

An important place to start is the current and anticipated regulatory framework.

On February 13, 2024, the European Parliament’s Committee on the Internal Market and Consumer Protection adopted the AI Act. It will be submitted for a plenary vote that is provisionally scheduled for April 10-11, 2024. The Act is the world's first comprehensive law on artificial intelligence.

Various U.S. local and state governments also have passed and are passing legislation to govern the use of AI tools in employment, and the White House Office of Science and Technology Policy has issued the Blueprint for an AI Bill of Rights. The focus of the laws, proposed regulations and the bill of rights is on maintaining safe systems, ensuring data privacy, offering notice and explanation of the use of any automated systems and providing opt-outs.[1]

While the EU AI Act may not apply to all U.S. companies and will not be enforced until 2025, it provides an excellent overarching framework and guide to what companies can expect to see in future AI regulation. Also, any company doing business in the EU should ensure compliance with the Act because once enforceable, companies that violate the EU AI Act’s rules could face fines of up to 35 million euros or between 1.5% to 7% of their global sales in the preceding financial year.

What is AI?

To comply with the act, a company must first identify the presence of AI in its systems. The publicly available EU AI Act Compliance Checker [2] is a good resource.

The EU AI Act provides a good definition of AI, which follows the Organization for Economic Co-Operation and Development’s latest definition of an AI system as “a machine-based system designed to operate with varying levels of autonomy and that may exhibit adaptiveness after deployment and that, for explicit or implicit objectives, infers, from the input it receives, how to generate outputs such as predictions, content, recommendations, or decisions that can influence physical or virtual environments.”

Risk Management

After identifying the presence of AI, a company must understand the level of risk associated with the use of such technology. The EU AI Act follows a risk-based approach to regulation, categorizing AI applications into four levels. The higher the risk, the stricter the governance.

The levels include: (1) Unacceptable Risk: a system that poses a threat to humans such as cognitive manipulation; (2) High Risk: a system that affects the safety of individuals and fundamental rights, such as credit scoring systems and automated insurance claims; (3) Limited Risk: AI tools like chatbots; (4) Minimal Risk: applications such as video games or spam filters.

AI systems identified as high-risk include AI technology used in employment, management of workers and access to self-employment, such as an HRIS system. Other high-risk systems would include systems involving critical infrastructure, education and vocational training, essential private and public services (e.g. healthcare, banking), certain systems in law enforcement, migration and border management, justice and democratic processes (e.g. influencing elections).

Any company that is considering procuring high-risk AI should establish a risk management system (including a record keeping component). Such a system should include a plan to conduct data governance and ensure that training, validation and testing datasets are relevant, sufficiently representative and, to the best extent possible, free of errors and complete according to the intended purpose. Any such program should include human oversight to establish a quality management system to ensure compliance.

Key Questions

In addition to any internal steps for the use of AI, a company seeking to use a vendor selling AI (such as an HRIS) should ask the vendor whether they have taken the required steps to comply with the EU AI Act.

Some examples of key questions to consider asking:

  • How do your AI systems comply with relevant regulations, such as GDPR, CCPA, or industry-specific standards?
  • How do you address ethical considerations and ensure compliance with emerging AI regulations, such as the EU AI Act or similar legislation?
  • What measures are in place to ensure data privacy and security?
  • Can you explain your approach to data encryption, access controls, and data breach prevention?
  • How do you handle return of data at the conclusion or termination of a contractual relationship?
  • How do you mitigate biases in your AI algorithms?
  • What measures do you take to ensure the accuracy of AI predictions, recommendations, or decisions?
  • How do you improve the performance of your AI systems over time and monitor performance?
  • Do you offer support and training to assist companies in implementing and using your AI systems effectively?
  • What measures do you take to ensure your AI solutions remain adaptable to future technological advancements?

When selecting AI system vendors, it is important to conduct due diligence to assess their compliance efforts and commitment to ethical AI practices.

Key Considerations

In general, key considerations include: (1) ensuring the data stored is encrypted; (2) inquiring about what the vendor has done to comply with the EU AI Act; (3) obtaining explicit consent from employees before collecting and processing their personal data; (4) ensuring vendor agreement includes timeline and requirement that any personal data is returned to the company upon termination of such agreement; and (5) providing training on any new systems.

[1] Blueprint for an AI Bill of Rights | OSTP | The White House, https://www.whitehouse.gov/ostp/ai-bill-of-rights/ (last visited March 15, 2024).

[2] EU AI Act Compliance Checker | EU Artificial Intelligence Act https://artificialintelligenceact.eu/assessment/eu-ai-act-compliance-checker/ (last visited March 15, 2024).

[View source.]

DISCLAIMER: Because of the generality of this update, the information provided herein may not be applicable in all situations and should not be acted upon without specific legal advice based on particular situations.

© Fox Rothschild LLP | Attorney Advertising

Written by:

Fox Rothschild LLP
Contact
more
less

Fox Rothschild LLP on:

Reporters on Deadline

"My best business intelligence, in one easy email…"

Your first step to building a free, personalized, morning email brief covering pertinent authors and topics on JD Supra:
*By using the service, you signify your acceptance of JD Supra's Privacy Policy.
Custom Email Digest
- hide
- hide