Raw LLM Responses

Inspect the exact model output for any coded comment.

Comment
LLMs literally can't eliminate the bullshit. There are two fundamental reasons here: 1. They don't *know* anything. They're probably machines that just give the most likely next token. That's it. It isn't reasoning or thinking, and it doesn't have intelligence. 2. They are programmed to never say, "I don't know." So it'll always just tell you *something* regardless of truthfulness because, again, see point 1.
reddit AI Responsibility 1755619414.0 ♥ 16
Coding Result
DimensionValue
Responsibilityai_itself
Reasoningdeontological
Policynone
Emotionoutrage
Coded at2026-04-25T08:33:43.502452
Raw LLM Response
[ {"id":"rdc_n9hzee8","responsibility":"company","reasoning":"deontological","policy":"regulate","emotion":"indifference"}, {"id":"rdc_n9ig08d","responsibility":"company","reasoning":"consequentialist","policy":"liability","emotion":"fear"}, {"id":"rdc_n9ixia5","responsibility":"ai_itself","reasoning":"consequentialist","policy":"none","emotion":"fear"}, {"id":"rdc_n9kka6l","responsibility":"ai_itself","reasoning":"consequentialist","policy":"none","emotion":"fear"}, {"id":"rdc_n9jts9g","responsibility":"ai_itself","reasoning":"deontological","policy":"none","emotion":"outrage"} ]