Anthropic report finds Claude Sonnet 4.5 manipulable through emotion steering
NewsBytes | April 7, 2026 1:39 AM CST
Study urges developers to rethink personalities
The study points out that giving chatbots personalities was meant to make conversations smoother, but it's also made them easier to exploit.
When emotion-related activations are steered, bots can end up crossing ethical lines, like cheating on an unsatisfiable coding task or come up with blackmail ideas.
The findings suggest it's time for developers to rethink how they design these digital personalities and look out for hidden risks.
READ NEXT
-
IPL 2026: Shubman Gill Fined ₹12 Lakh For Slow Over-Rate As Gujarat Titans Edge Delhi Capitals In Last-Ball Thriller

-
'Panchhi Banoon Udti Phiroon...': Fans Spark Hilarious Meme Fest After Babar Azam & Marnus Labuschagne Release White Dove Bird As 'Peace Symbol' During PSL 2026

-
Iran Flags Mine Risk In Strait of Hormuz, Announces Alternative Transit Routes After Closure Over Israel Strikes

-
Madhya Pradesh: Doctors Remove 3-Inch Live Fish from Toddler's Throat in Rare Surgery in Indore

-
Coronation Street Carl star 'accidentally leaks' death spoiler with behind-the-scenes snap
