What's putting you at risk?

We surveyed over 2,000 business leaders like YOU to identify the key risks you’re facing and created the UK Business Risk Report - full of practical insights to help you tackle them. Download your FREE copy today.

Technology artificial intelligence

How AI can be helpful or harmful to your business

Artificial intelligence has embedded itself into the business landscape. It’s no longer used by big tech companies alone. Businesses across various industries are actively integrating AI into their processes, acquiring tech start-ups and exploring opportunities to deploy the technology in the near future. The COVID-19 pandemic has only accelerated this trend as businesses have had to contend with plummeting revenue and workforce restrictions. 

But as companies increasingly look toward AI to solve business challenges and increase their profitability, what risks will they face? How might they mitigate such risks? And what else should business leaders take into consideration? 

Balancing public health and individual rights

Despite the substantial benefits that the technology promises, AI deployment without safeguards poses risks at all levels of business, especially for traditional, non-tech companies. To limit severe financial and reputational harm, it is crucial that companies weigh the many benefits of AI use against the risks intrinsic to its use, as well as associated concerns from the broader community. Consider, as one particularly pertinent example, the countless ways wherein AI has been deployed in response to the global pandemic: from contact tracing to enhanced infection risk profiling, those who develop and use such cutting-edge techniques must carefully balance the dual imperatives of public health and individual liberties

Defending the decisions of algorithms

Given the self-learning and automated nature of AI, a well-known concern associated with the technology is that of ‘explainability,’ especially with public-facing “black box” AI models that make decisions on sensitive or consequential issues such as job recruitment, credit risk assessments and medical diagnoses. A lack of transparency and traceability, particularly when using externally procured applications, exposes businesses to significant reputational harm. 

For instance, numerous controversies in recent years have shown us that AI systems can inadvertently generate biased and potentially discriminatory outputs that extend inequalities. Businesses, especially when such adverse outcomes to customers and staff are possible, must be able to explain and defend algorithm-based decision processes and their output to a range of stakeholders. This can include subject-matter experts and even the legal community in cases of alleged malpractice. Big-name tech firms with dedicated AI specialists on hand have long struggled with this issue; non-tech companies are also at risk of intense public scrutiny and brand damage.

Cybercriminals exploiting AI

Cyber risk is also a significant threat to companies using AI, especially with the rush toward digitisation during the COVID-19 lockdowns. In fact, participants in a survey of more than 12,000 business executives rated cyber risk as the top risk for doing business in the US, the UK and Canada — among other developed economies — over the next decade. The growing use of AI in critical business operations will only increase vulnerability to cybercrime as hackers can gain control of entire systems simply by manipulating their underlying algorithms. AI can moreover directly enhance the arsenal of cybercriminals who can now cause disproportionate levels of harm by leveraging the speed of decision-making enabled by automated programs. Smarter cyber threats, coupled with industry’s growing reliance on digital capabilities, only escalate the risks to operations and revenue streams.

Beyond such technical hazards, businesses that adopt AI solutions, also risk reputational harm and revenue erosion if consumer data is used inappropriately or otherwise exposed. Some major tech companies have drawn sharp criticism over the last few years for allegedly misusing sensitive voice data recorded by their AI-powered digital assistants. Given big tech’s enduring ability to generate insights from big data and exploit personal profiles in ways that consumers have not anticipated or accepted, such scrutiny will surely persist. This public outcry for data privacy will no doubt extend to non-tech firms in the future. 

Lack of holistic governance standards

Finally, due to the emergent nature of this technology, companies may find themselves deploying AI in rapidly evolving regulatory environments, complicating compliance efforts. The global fragmentation of data standards creates additional regulatory discontinuities across jurisdictions. Non-tech firms that are less familiar with international differences in AI-specific legislation may struggle to align their use of AI with shifting regional mandates, thereby necessitating decentralised, and often difficult and costly, policy rollouts.

These are just some of the threats to which businesses expose themselves should they attempt to realise the benefits of AI without implementing effective and holistic governance measures. Given the complexity of the technology and the pervasiveness of its potential perils in all aspects of operations, a multifaceted and dynamic approach to governance is required to manage AI risks. It is important that businesses evaluate their use of AI technology across five areas:

  • Intent: Using data in a principled manner and verifying that AI design and implementation processes are ethically aligned and appropriate.
  • Fairness: Ensuring that the processes and outputs of AI systems do not unwittingly discriminate against any group or individual. 
  • Transparency: Verifying that AI processes are explainable and repeatable. 
  • Safety/Security: Establishing robust capabilities in data governance, threat protection, and user privacy so as to better defend against malicious incursions. 
  • Accountability: Undertaking rigorous audit and compliance assurance processes to assuage the concerns of various stakeholders — lawmakers, auditors, customers, business partners and shareholders, among others.  

To activate effective governance aligned with these principles, businesses must additionally implement supporting infrastructure and processes, including an oversight committee, a risk register and testing and analytics. Training should also be provided for staff involved in development and management of AI such that they can proficiently handle the dynamic risks that this technology presents. 

By framing the management of their AI solutions around the five dimensions outlined above and instituting proper governance mechanisms, businesses can ensure that they do not expose themselves to undue risk, or worse, inadvertently cause harm to broader society. In doing so, they will be able to rest easier when procuring, developing and implementing new AI solutions. 

 

Read the full, unedited article and subscribe to get BRINK in your inbox

***

Marsh Commercial is part of Marsh & McLennan Companies. BRINK combines knowledge and expertise from across Marsh & McLennan and is managed by Atlantic Media Strategies, the digital consultancy of The Atlantic. The content is subject to BRINK’s Terms and Conditions of Use.

***