As the use of ChatGPT and other AI chatbots becomes more prevalent in the business world, insurers and brokers need to stress the importance of best practices and ask their clients how they plan to adopt these technologies safely. Speaking with Insurance Business, John Farley, the managing director of cyber practice at Gallagher, said it is imperative that policyholders be mindful of the following:
“Organisations tend to adopt new or emerging technologies to promote goods and services, but then realise the cyber risk associated with them.” To stay ahead, brokers must acclimate their policies to the changing cyber insurance landscape in 2023 as new opportunities for cyber-attacks begin to take shape with these emerging technologies.
With ChatGPT being very new to the market, it is imperative to pinpoint exactly what types of threats can impact a business in order to safeguard its digital assets and information. Farley has noticed that since ChatGPT has made its way into the mainstream, that there have been “dark web forums popping up that are detailing how the technology can be used by junior hackers to learn how to develop malware and phishing email campaigns and deploy them.”
“This could potential grow the number of novel attackers into a large army in a short amount of time,” Farley said. It will give individuals without a sophisticated background in hacking to become adept at creating rogue software that will have the potential to wreak havoc on businesses, presenting claims for insurers that could be avoided or lessen their impact.
“As hackers exploit the technology or regulators clamp down on the various usages of it, companies will need to be more proactive in staying ahead of the curve with protective measures.”
When using AI-enabled chatbot services, a business must be aware of the kinds of information that is input into them and whether that material is classified or not. From a risk management standpoint, data hygiene must be top priority to “mitigate any chance of intellectual property or trade secrets being pushed out into the public,” Farley said.
To sidestep this potentially disastrous outcome, “very thoughtful processes need to be deployed about what data should go in,” Farley said. Technology is only as good as the data that is put into them, so the information should be thoroughly vetted, while allocation of this job should be entrusted to a select few — doing so would also thwart any instances of misinformation being spread, which can be damaging to a company’s reputation.
A company should also be concerned with data bias and should abridge the legal, IT and marketing teams to negate this from occurring.
In tandem with internal AI concerns, there should also be a plan put in place to address external regulators at a state, federal and international level for compliance reasons.
“At some point, when dealing with business in cyberspace, at some point you will have a regulator who is going to have something to say about the controls surrounding ChatGPT,” Farley said. Since the service is in its infancy, these parameters have not yet been established, which leaves a lot of variables and vague decision making in the meantime.
Risk managers can write specific AI usage policies that will solidify this highly selective task to ensure that classified data does not get into the wrong hands.
ChatGPT is coming for my firms marketing department. Some pretty solid #insurance and #riskmanagement advice pic.twitter.com/ZbIm80usY3
— Bradley Dlatt (@bdlatt) March 25, 2023
Organisations and brokers should be mindful of the rapidly changing cyber insurance products that could potentially impact the scope of coverage. In 2023, the market is constantly in flux, spurring insurers to adopt new methods to mitigate cascading losses for regulatory risk, especially around the use of novel technologies.
Sub-limits and coinsurance are imposed for certain types of cyber losses, while some carriers have even modified policy language to restrict or exclude coverage for certain incidents that may result in litigation, regulatory investigations and fines.
“Artificial intelligence will significantly deepen our digital footprint, while potentially raising a cyber risk profile almost in lockstep,” Farley said.
“It is the industry’s responsibility to stay on top of any developments and implement or modify policies as a result. Things are going to change, but its beneficial to enact some cautious optimism at how ChatGPT will revolutionise things.”