Challenges and Risks in Generative AI: Considerations for Investors

January 3rd, 2024 | Kate Tong

Since OpenAI’s ChatGPT went viral in late 2022 for its unprecedented ability to engage in human-like conversations and provide articulate responses in wide-ranging domains of knowledge, several competitors have begun introducing their own iterations of the technology. This type of AI technology, known as generative AI, is based on large language models that are trained on massive amounts of data, which could include text, images or other media. The models identify the patterns and structures of the training data and then generate new content that has similar characteristics based on user prompts.

There are various benefits to incorporating generative AI in a business – process improvements, cost reduction and value creation, to name a few. Leveraging these opportunities, companies across different sectors have already begun testing and implementing generative AI tools. Examples range from financial institutions deploying chatbots trained on internal databases to provide financial advice to customers, to healthcare institutions automating the generation of medical documentation based on conversations between patients and physicians. Across industries, companies are also incorporating generative AI tools in marketing, customer service and product development.

As such, investors have to pay attention not only to large tech companies that are building the foundational models, but also to companies that are starting to incorporate generative AI tools into their business. As with most new technologies, there are potential risks that should be adequately considered and safeguarded against before widespread deployment. Regulation will be important in helping reduce these risks. But because the development of regulation occurs at a much slower pace than the development and application of AI, investors should actively consider its risks and seek stewardship opportunities in companies involved in generative AI to address these risks.

Challenges and Risks in Generative AI

Generative AI models have various known issues. These models have the tendency to “hallucinate,” generating false outputs that are not justified by the training data and presenting them as a fact. These errors can be caused by various factors, such as improper model architecture or noise and divergences in the training data. Opaqueness about the generation of model outcomes is also an issue. With billions to trillions of model parameters that determine the probabilities of each part of its response, it is exceedingly difficult to map model outputs to the source data, including in cases of hallucination.

In addition, if the training data contains societal prejudices or if the algorithm design is influenced by human biases, the model may learn and propagate these biases in its outputs. Enterprise applications could also be vulnerable to data privacy issues and cybersecurity threats. This includes leakage of sensitive information within the training data if the model is customer/public-facing, usage of personal or sensitive data in model training that may have needed explicit consent to use, as well as malicious attacks from hackers that aim to manipulate model outputs.

These issues give rise to various legal and reputational risks, the scale of which depends on the criticality of the use case and the company’s industry. For example, the financial and healthcare industries may be subject to severe consequences if problems do arise, due to the high-stakes nature of these industries.

Sample Use Cases in the Financial Industry

In financial advisory use cases, model hallucinations could give inappropriate advice or offer the wrong product to undiscerning clients, which could undermine public trust in AI systems and the financial institutions using them. Lack of transparency about the generation of model outcomes is also a key issue for financial institutions, as these institutions are required to be able to explain their decisions internally and to external stakeholders. Considering all this, it is best practice to implement a degree of separation between direct model outputs and the customer, where internal staff could be trained to recognize potential errors and inconsistencies in model outputs and assume ultimate responsibility for the decision-making process.

Generative AI could also offer a quick and low-cost way for financial institutions to profile their clients for marketing campaigns, risk management and identification of suspicious transactions. However, overreliance on generative AI profiling could violate anti-discrimination laws due to potential bias embedded within the models. Appropriate human judgment will need to complement generative AI models that perform client profiling. Financial institutions will also need to have strong data privacy policies and robust cybersecurity measures to address generative AI’s risks to their sensitive client information and proprietary data.

Questions for Investors to Consider

In view of all these issues and risks, below are questions investors should consider when assessing companies employing generative AI tools:

  • What are the risk-mitigating mechanisms and/or circumstances? Solutions include having trained internal staff act as an intermediary between direct model outputs and the customer; working to understand potential biases in the training data and address them in model design; regular and proactive monitoring of model output to promptly identify and address any signs of hallucinations; implementing robust cybersecurity measures; etc.
  • What is being done to enhance model performance? Solutions include ensuring that training data is of high quality, accurate and up to date; implementing iterative feedback loops to refine and improve model performance; etc.
  • Is there any transparency and oversight of ethical AI principles? This pertains to providing transparency on data sourcing and data privacy concerns; defining clear policies and procedures to ensure compliance with ethical standards and emerging regulations; outlining the roles and responsibilities of individuals involved in the development, operation and oversight of the generative AI model; etc.


Contributor Disclaimer

The information contained herein is for information purposes only. The information has been drawn from sources believed to be reliable. The information does not provide financial, legal, tax or investment advice. Particular investment, tax or trading strategies should be evaluated relative to each individual’s objectives and risk tolerance.

This document may contain forward-looking statements (“FLS”). FLS reflect current expectations and projections about future events and/or outcomes based on data currently available. Such expectations and projections may be incorrect in the future as events which were not anticipated or considered in their formulation may occur and lead to results that differ materially from those expressed or implied. FLS are not guarantees of future performance and reliance on FLS should be avoided.

The statements and opinions contained herein are those of Kate Tong and do not necessarily reflect the opinions of, and are not specifically endorsed by, TD Asset Management Inc.

RIA Disclaimer

The views and opinions expressed in this article are solely those of the authors and do not necessarily reflect the view or position of the Responsible Investment Association (RIA). The RIA does not endorse, recommend, or guarantee any of the claims made by the authors. This article is intended as general information and not investment advice. We recommend consulting with a qualified advisor or investment professional prior to making any investment or investment-related decision.

Author

author's photo

Kate Tong

Analyst
TD Asset Management

Kate is an Analyst with the ESG Research & Engagement Team where she is responsible for researching relevant ESG issues and supporting the firm's ESG policy positions, engagement and proxy voting efforts. Prior to joining the firm, Kate worked as an Asset Management Intern for a realty firm where she analyzed the multi-residential portfolio's energy consumption and greenhouse gas emission trends to support the company's goal to achieve net zero. She also worked as an undergraduate research assistant in the Department of Physical and Environmental Sciences at the University of Toronto. Kate holds a Bachelor of Science with Honours from the University of Toronto.