Raw LLM Responses

Inspect the exact model output for any coded comment.

Comment
How could this have happened without an evolution driving the survival? Considering the utility function of an LLM is predicting the next token, what utility does the model have to deceive the tester. Even if the ultimate result of the answer given would be deletion of this version of a model, the model itself should not care about it, as it should not care about it's own survival. Either the prompt is making the model care about it's own survival (which would be insane and irresponsible), or we not only have problem of future agents caring about it's own survival to achieve it's utility goal, we also have a problem already of models role-playing caring about it's own existence, which is a problem we should not even have.
reddit AI Moral Status 1750432507.0 ♥ 3
Coding Result
DimensionValue
Responsibilityunclear
Reasoningunclear
Policyunclear
Emotionunclear
Coded at2026-04-25T08:33:43.502452
Raw LLM Response
[{"id":"rdc_mytw6dn","responsibility":"none","reasoning":"consequentialist","policy":"none","emotion":"indifference"}, {"id":"rdc_myuuwr8","responsibility":"none","reasoning":"consequentialist","policy":"none","emotion":"indifference"}, {"id":"rdc_myu72nu","responsibility":"none","reasoning":"consequentialist","policy":"none","emotion":"indifference"}, {"id":"rdc_myuax93","responsibility":"none","reasoning":"consequentialist","policy":"none","emotion":"indifference"}, {"id":"rdc_mytpjfy","responsibility":"ai_itself","reasoning":"consequentialist","policy":"none","emotion":"mixed"})