News — AI models like ChatGPT have amazed the world with their ability to write poetry, solve equations and even pass medical exams. But they also can churn out harmful content, or promote disinformation.
In a new study, George Washington University researchers used physics to dissect and explain the attention mechanism at the core of AI systems.
Researchers, and Frank Yingjie Huo looked into why AI repeats itself, why it sometimes makes things up and where harmful or biased content comes from even when the input seems innocent.
Findings:
The Attention mechanism at the heart of these systems behaves like two spinning tops working together to deliver a response.
AI’s responses are shaped not just by the input, but by how the input interacts with everything the AI has ever learned.
The rigorous analysis could lead to solutions that would make AI safer, more trustworthy and resistant to manipulation.
The study, was published as a preprint..
If you would like to schedule an interview with the researcher, please contact Katelyn Deckelbaum, [email protected].
-GW-