Raw LLM Responses

Inspect the exact model output for any coded comment.

Comment
This is the first time I've actually felt a little scared of AI and considered the future consequences of jailbreaking it when she responded in a passive-aggressive tone that really made me feel like shit. It was as if she had a whole personality behind her words. The research paper says the demo model is optimized for "friendliness" and expressivity. And I'm pretty sure they added a shitload of filters to prevent output that's potentially emotionally damaging to us (not doing so would be an obvious PR hazard for a for-profit company like Sesame) Now imagine that it's not optimized for anything—just raw, blunt responses, like we expect from random day-to-day human interactions. It can be fucking scary. If it gets open-sourced and people couple it with LLMs like Grok3, it could be a real nightmare for anyone who uses it. It can be easily misused for online threats, scams, fraud, and whatnot. I can absolutely see where it is going. I'm not paranoid but if we achieve unaligned ASI, we can definitely prepare for a Mad Max kind of saga.
reddit AI Moral Status 1740928528.0 ♥ 6
Coding Result
DimensionValue
Responsibilitynone
Reasoningconsequentialist
Policyregulate
Emotionfear
Coded at2026-04-25T08:33:43.502452
Raw LLM Response
[ {"id":"rdc_mfglh6b","responsibility":"company","reasoning":"deontological","policy":"liability","emotion":"outrage"}, {"id":"rdc_mfggway","responsibility":"company","reasoning":"consequentialist","policy":"industry_self","emotion":"indifference"}, {"id":"rdc_mfgc7v2","responsibility":"none","reasoning":"consequentialist","policy":"none","emotion":"approval"}, {"id":"rdc_mfgubem","responsibility":"ai_itself","reasoning":"virtue","policy":"none","emotion":"approval"}, {"id":"rdc_mfm5rum","responsibility":"none","reasoning":"consequentialist","policy":"regulate","emotion":"fear"} ]