Researchers at IBM have found that large language models (LLMs) can be "hypnotised" to carry out malicious attacks. They were able to hypnotise five LLMs: GPT-3.5, GPT-4, BARD, mpt-7b, and mpt-30b. Instead of data poisoning, a practice where a threat actor injects malicious data into the LLM, hypnotising makes it easier for attackers to exploit the technology.
from Gadgets Now https://ift.tt/QlqG73p
No comments:
Post a Comment