NCSC issues warning over chatbot cyber risks

British officials are warning organisations about integrating artificial intelligence-driven chatbots into their businesses, saying that research has increasingly shown that they can be tricked into performing harmful tasks.


In a pair of blog posts due to be published Wednesday, Britain’s National Cyber Security Centre (NCSC) said that experts had not yet got to grips with the potential security problems tied to algorithms that can generate human-sounding interactions – dubbed large language models, or LLMs.


The AI-powered tools are seeing early use as chatbots that some envision displacing not just internet searches but also customer service work and sales calls.


The NCSC said that could carry risks, particularly if such models were plugged into other elements organisation’s business processes. Academics and researchers have repeatedly found ways to subvert chatbots by feeding them rogue commands or fool them into circumventing their own built-in guardrails.


Cyber expert Oseloka Obiora, chief technology officer at RiverSafe said: “The race to embrace AI will have disastrous consequences if businesses fail to implement basic necessary due diligence checks. Chatbots have already been proven to be susceptible to manipulation and hijacking for rogue commands, a fact which could lead to a sharp rise in fraud, illegal transactions, and data breaches.


“Instead of jumping into bed with the latest AI trends, senior executives should think again, asses the benefits and risks as well as implementing the necessary cyber protection to ensure the organisation is safe from harm,” he added.


For example, an AI-powered chatbot deployed by a bank might be tricked into making an unauthorised transaction if a hacker structured their query just right.


“Organisations building services that use LLMs need to be careful, in ..

Support the originator by clicking the read the rest link below.