Tech

Syntax hacking: Researchers discover sentence structure can bypass AI safety rules

Published

on

[ad_1]

Researchers from MIT, Northeastern University, and Meta recently released a paper suggesting that large language models (LLMs) similar to those that power ChatGPT may sometimes prioritize sentence structure over meaning when answering questions. The findings reveal a weakness in how these models process instructions that may shed light on why some prompt injection or jailbreaking approaches work, though the researchers caution their analysis of…

[ad_2]

Source link

Exit mobile version