According to a recent report by McKinsey, “Generative AI has the potential to change the anatomy of work” and will “substantially increase labor productivity across the economy”. At the time of writing, they have identified 63 Generative AI use cases spanning 16 business functions, which gives you some idea of the level of potential transformation. Of course, productivity, efficiency and effectiveness are the holy grails for organisations who want to achieve long-term success in a changing world, so this news has been met with great excitement.
However, amidst the excitement, one word rears its head time and again: risk. Generative AI is still in its very early days, after all. And so, businesses need to think carefully about how it is deployed in their organisations – and what the potential outcomes could be. “This comes down to risk tolerance and risk appetite,” says Quentyn Taylor, Canon EMEA’s Senior Director – Information Security and Global Response. “For example, are we willing to accept an AI responding to customer queries, when it will very occasionally get it wrong?” This is just one way in which GenAI could be used to increase efficiency, but as Quentyn points out, what is the reputational risk when a customer receives incorrect information? It very much depends on the product and the customer, and this is the big challenge for businesses – having a clear understanding of the places where Gen AI can add true value, and those where there is no level of acceptable risk. In order to even begin to make these judgements, it’s therefore critical to understand just what the risks could be.
Protecting Intellectual Property and commercially sensitive information
This is the area which has been most immediately addressed by most organisations, with some putting a blanket ban on the use of any Generative AI tools and services to protect and preserve their corporate privacy. Essentially, everything you enter into a GenAI model becomes training data. So, if you were to ask it to write a speech for an exciting product release, supplying all details of that product in the request, then you’ve essentially just uploaded business critical, embargoed information to a globally used tool. If your business lives and dies by its IP, then this is a level of unacceptable risk. “On the other hand,” says Quentyn, “if you used Generative AI to write 300 slightly varying descriptions for existing products, is that a problem? Probably not.” Another perspective to consider is the effort of policing the issue versus outcome: “Is stifling the use of GenAI an effective use of our time? Can we fully block access when there are thousands of new tools being released every day?” he asks.
view
Four GenAI risks every business should know
Decision-making, bias and communication
Who’s in charge? Well, of course decision-making starts at the top, but when the board is drawing conclusions with the help of Generative AI tools, the process has to be crystal clear. Equally, it’s vital that bias is taken into consideration when using GenAI as a tool through which to analyse options for increased productivity and profitability. It’s widely understood that training data must be vast for an AI model to be even remotely fair, but even so, bias exists. This is why many organisations choose not to use such tools in key decision-making areas, hiring often being cited as a problematic area. Quentyn underscores this, saying “You must always understand the context in which any AI-supported decisions are made,” says Quentyn. “And this must be clearly communicated to the rest of the organisation, otherwise you risk creating widespread confusion and mistrust in leadership.” This is particularly important when you consider how often organisational decisions require ‘unpicking’, to understand the often very nuanced basis upon which actions are mandated.
Copyright infringement
Right now, there are some very high-profile lawsuits underway where parties believe their creative work has been used to train an AI without their consent. And there are few graver concerns around Generative AI than those around the legality of the content it creates. Yes, there are some incoming tools (such as Adobe Firefly) which are trained on nothing but legally owned data, but for others, there is little to no clarity right now as to how safe they are to use for, say, creating a suite of images for a social media campaign or designing a new brand identity. When working with third parties on such activities, Quentyn sees value in “adapting or updating contracts to mitigate the risk and making sure that clear guidance and policy is in place internally”.
"Fact check, fact check, fact check. That's one of the key roles for humans when we use AI to generate underlying content. We must be scrupulous and effective editors.”
When GenAI tells lies
You might have heard the term ‘hallucinations’ referenced fairly frequently in the context of GenAI. In simple terms, it is when an AI model generates a false or unrealistic response. It could be something silly, like a completely made-up word or nonsensical sentence. Or it could confidently provide a false piece of information, which is what happened to two lawyers who presented six case citations in court that turned out to be completely fictitious. These were later found to have been generated by ChatGPT and the lawyers were ordered to pay a $5000 fine. AI experts acknowledge the issue and are ‘making progress’, but in the meantime this is an area of exceptional risk for organisations and their leaders. “Fact check, fact check, fact check,” stresses Quentyn. “That's one of the key roles for humans when we use AI to generate underlying content. We must be scrupulous and effective editors.” He also warns of the risks of using Gen AI bots to monitor and respond on social media. “Your bot could begin to give answers that, in theory, may well be correct – just not in the context for the question they were given.”
Overall, Quentyn feels positive that many organisations will adopt AI as a “background benefit”, built into and offered through the services and tools they already use. In this respect, much of the risk is already mitigated through contracts with third parties and by deploying solutions from known, respected and proven partners. “Perhaps a company might use AI to search through emails, looking for phishing scams,” he explains. “That's not something they are likely to specifically code or develop themselves, but they will gain benefit from using tools that include this service.” Ultimately, any business involves a significant amount of risk management, and the new opportunities presented by Generative AI are no different in this respect.
Read more articles like this from Canon VIEW
-
السلعة
هل تحقق مستوياتك الإبداعية قيمة في العالم الإنتاجي؟
مهما كان شعورك حيال ذلك، حان وقت التعرّف على الذكاء الاصطناعي الإنتاجي. لكن ماذا قد يعني ذلك بالنسبة إلى الإبداعي البشري؟ نستكشف ما يمكن أن يتغير، وما يمكن أن يبقى على حاله.
13 Mar 2024
-
Article
Part Two: What might your creativity look like in a Generative world?
In the second part of our dive into Generative AI, we look at some potentially unexpected – and positive – outcomes of its use in creative industries.
26 Jun 2023
-
Article
Clever crowd counting: fast, accurate and helpful
Keeping crowds of people safe is all in the numbers, but manually counting people can be tricky. So, why not use Artificial Intelligence to do it for you?
03 Apr 2023
-
Article
Making progress: two inspiring women leading the way in cybersecurity
‘Most Inspiring Women in Cyber’, Zoë Rose and Sakina Asadova discuss their careers and share words of advice for those following in their footsteps.
06 Mar 2023