- Generative AI systems have already found widespread application in the business world and their capacity to disrupt a broad range of industries is now apparent.
- The technology has enormous potential, but comes with many risks, including the potential for copyright and privacy infringements, contractual violations, the disclosure of trade secrets and untrustworthy outputs.
- Directors weighing opportunities for a company to use, build or contribute content for these AI platforms need to understand in broad terms what generative AI systems are, as well as the legal issues they pose and what steps companies can take to mitigate the risks.
What Is Generative AI and How Does It Work?
OpenAI’s ChatGPT platform is reportedly the fastest-growing consumer application in history. It uses generative AI models, which produce new content based on training on vast quantities of data. The system underlying ChatGPT reportedly trained for months on hundreds of billions of words pulled from the Internet. Through this process, OpenAI’s systems and similar “large language models” have achieved near-human level abilities to answer questions, write poetry, compose essays and even perform in the 90th-99th percentile across a wide range of college, graduate and post-graduate exams.
But boards should be aware of the risks associated with models trained on data from the Internet, including the intellectual property, privacy and contractual risks of relying on the outputs of such models.
How Generative AI Is Being Used Today
ChatGPT and other text-generating models are far from the only uses of generative AI. Companies across different industries are using the technology:
- Financial services firms are leveraging generative AI to streamline backend operations, bolster cybersecurity, support service chatbots, accelerate software development, enhance fraud detection and provide personalized financial advice.
- Entertainment companies are using text-to-image generators to create art for storyboards and visual content, including special effects, for films and video games.
- Pharmaceutical researchers are using generative AI to better understand the structure of proteins and design them specifically for medicines. For example, Canadian researchers trained an AI system on images of known proteins to generate new proteins with specific difficult-to-replicate protein folding.
- The materials science industry uses the technology to compose new materials with the desired physical properties.
- In healthcare, AI is being applied to electronic health records systems, and generative AI systems are also being used to produce synthetic data (i.e., fictitious data that mimics real-world data without personally identifiable information) to allow data sharing and analysis otherwise restricted by privacy laws.
Boards will need to understand the risks these innovations bring as this technology expands across all sectors.
Frequently Asked Questions About the Use of Generative AI
1. What are the risks when inputting information into third-party generative AI platforms?
But if the information input is owned by a third party, you may breach confidentiality obligations to them. Furthermore, it may be hard to anticipate the impact of supplying information for the model. The platform’s use of your information to improve its model could result in that information being incorporated into a training dataset published by the platform provider, or used to train a model that ultimately discloses your information in response to another user’s prompt. For instance, if your employees ask a generative AI system to debug confidential software source code, that source code could be used to train an improved model that releases the code in some form to subsequent users.
Moreover, AI systems typically store information on an external server. If the security of that server is breached, the user’s information could be disclosed publicly, which potentially could be devastating to the user, the owner of the information or both.
2. Who owns the output (or results) of generative AI systems?
In March 2023, the Copyright Office published guidance stating that that registrability of works including AI-created content depends on factors such as how the AI tool operates and how it is used to create the final work. Complex written, visual or musical works generated from simple human prompts by an AI system are not registrable, it said. In copyright registration forms, applicants now must disclose any AI-generated material in a work and explain the human author’s contribution.
Companies will therefore need to document the use of AI- and non-AI-generated content to ensure that their products can be copyrighted, and be alert to the possibility that the output of the models could infringe on the rights of others.
3. How trustworthy are the results of generative AI?
Today’s technologies are far from perfect. Because these systems are trained to generate responses that appear similar to the training data based on probabilities, they are prone to “hallucinations,” where the system generates inaccurate content and presents it as fact — often convincingly to non-experts.
Such inaccuracies could impact business outcomes or create liability issues if, say, false information is communicated to the public. That could result in reputational or operational damage, and even defamation claims.
AI text-to-image generators allow users to create amazingly realistic but fictitious images, such as the “deep fake” photos of Pope Francis wearing a white puffer jacket that went viral. The possibility of misuse of generative AI to spread disinformation and misinformation is a concern not just to organizations, but to society as a whole.
Bias also continues to be an important concern because generative AI systems may perpetuate or amplify biases in the training data or in the algorithms or user prompts. This is particularly concerning if the technology perpetuates or amplifies biases based on legally protected characteristics such as race, gender or sexual orientation. Regulators are already focused on this issue. The EU and U.K.’s General Data Protection Regulation and the California Consumer Privacy Act both give individuals the right to opt out of “automated decision-making” where AI may be used to build profiles and make individual decisions, such as extending employment or product offers.
4. Can companies face liability for training generative AI systems?
State-of-the-art generative AI is trained on vast amounts of data (including text or images). Boards considering deploying generative AI should understand if the training process violates copyrights, privacy requirements and/or contractual restrictions.
Copyright. Absent an express license, training generative AI may violate copyrights in the works included in the training data. Copyright owners have already filed infringement suits against generative AI providers in the U.S. and U.K.
There may be a “fair use” defense for a limited and “transformative” purpose, such as commenting on, criticizing or parodying a work. But fair use turns on the facts of a specific case and U.S. courts have not addressed the defense in the AI context, and other jurisdictions could come to different conclusions.
Privacy. An AI model trained on sensitive or personal information might unintentionally generate outputs that reflect this information. Even if the details are not explicitly in the training data, the systems might learn associations between individuals and sensitive attributes like race, gender or health status, potentially leading to privacy breaches. Individuals also may not be aware that their personal information is being used to train these systems, and they may not have given consent or been given an opportunity to opt out.
These types of privacy concerns recently led Italy’s data protection authority to briefly halt the use of ChatGPT in Italy while OpenAI responded to inquiries regarding privacy risks. And in several cases where personal data used to train AI models was gathered or used in violation of privacy policies, the Federal Trade Commission has required “algorithmic disgorgement”— the permanent deletion of all models improperly trained on the personal data. After years spent developing AI training datasets and training models, companies facing algorithmic disgorgement need to start from scratch.
What Should a Board Do?
Careful oversight remains critical. For a board, that entails taking reasonable measures to implement and oversee risk management and compliance controls.
Although courts in Delaware, whose law governs most large companies, have yet to weigh in on AI issues, some guidance can be found in a case involving a data breach that exposed customers’ personal information. Stockholders alleged that directors violated their duty of oversight.
Although the breach stemmed in part from significant lapses (including the use of a simple generic password to secure critical data), the Delaware Court of Chancery ruled for the directors. It noted that the board had charged two committees with monitoring the company’s data security processes, that those committees were well-functioning and met regularly, and that the committees set up appropriate reporting structures.
Boards should carefully consider how best to oversee a company’s use of generative AI. That may entail:
- Establishing monitoring and compliance systems and paying ongoing attention to them, perhaps through a committee empowered to evaluate technology-related risks.
- Paying particular attention to “mission critical” issues involving the use of generative AI.
- Discussing with advisers issues on which the board should receive regular reports and identifying what “red flags” (i.e., indications of potential operational deficiencies) may arise and how best to respond.
- Documenting in board minutes and materials the monitoring system reports to the board, and both directors’ and officers’ oversight efforts, so the company can respond to books-and-records demands and defend itself.
- Evaluating how generative AI may be used to enhance a company’s oversight systems and processes, for example, by automating reports or by creating monitoring or analysis tools to spot potential deficiencies.
Jurisdictions including the U.S., U.K., European Union and China are grappling with the question of whether and how to regulate the technology. Boards will also need to stay abreast of those developments.
View other articles from this issue of The Informed Board
- Nine Mistakes To Avoid When Transitioning CEOs
- The FTC’s Plan To Limit Noncompetes Could Pose an Array of Practical Problems
- The Impact of Banking System Turmoil: What’s Next?
- Podcast - What a New Executive Order and Tighter Controls on Tech Exports Mean for Companies Doing Business in China
See all the editions of The Informed Board
This memorandum is provided by Skadden, Arps, Slate, Meagher & Flom LLP and its affiliates for educational and informational purposes only and is not intended and should not be construed as legal advice. This memorandum is considered advertising under applicable state laws.