Raw LLM Responses

Inspect the exact model output for any coded comment.

Comment
Penrose describes consciousness as an emergent phenomena where the outcome is not only more than the sum of its parts but can be something completely different and thus can't be 'reverse engineered' (as in "2+2 makes neither 4 nor 5 but apples"). I'm just asking myself why those 'emergent qualities' shouldn't apply to AI at some point...especially in terms of Quantum computing.
youtube AI Moral Status 2025-04-21T06:4… ♥ 2
Coding Result
DimensionValue
Responsibilityunclear
Reasoningmixed
Policyunclear
Emotionmixed
Coded at2026-04-26T23:09:12.988011
Raw LLM Response
[{"id":"ytc_Ugy0eEbTZrFd19pjEg94AaABAg","responsibility":"none","reasoning":"consequentialist","policy":"none","emotion":"indifference"}, {"id":"ytc_UgxhVwQl4ejcLQZoxZx4AaABAg","responsibility":"unclear","reasoning":"mixed","policy":"unclear","emotion":"mixed"}, {"id":"ytc_UgyXdM7Fc-6yELJ4D8h4AaABAg","responsibility":"developer","reasoning":"mixed","policy":"regulate","emotion":"fear"}, {"id":"ytc_UgwKbyLaRfwD2i0cJqp4AaABAg","responsibility":"unclear","reasoning":"unclear","policy":"unclear","emotion":"mixed"}, {"id":"ytc_UgxKaargc3XDmnK1B_54AaABAg","responsibility":"none","reasoning":"consequentialist","policy":"none","emotion":"indifference"}, {"id":"ytc_UgzAHwFExuxVyJ6cUKl4AaABAg","responsibility":"ai_itself","reasoning":"deontological","policy":"unclear","emotion":"approval"}, {"id":"ytc_UgzhdkBMF6LiVCJKBgZ4AaABAg","responsibility":"unclear","reasoning":"mixed","policy":"unclear","emotion":"mixed"}, {"id":"ytc_UgxrheKRgQHXL43DWTN4AaABAg","responsibility":"unclear","reasoning":"mixed","policy":"unclear","emotion":"mixed"}, {"id":"ytc_UgxF5Lm4hQRRra1kURB4AaABAg","responsibility":"unclear","reasoning":"mixed","policy":"unclear","emotion":"outrage"}, {"id":"ytc_Ugy3lkG5aidJJS-allN4AaABAg","responsibility":"none","reasoning":"consequentialist","policy":"none","emotion":"indifference"}]