Raw LLM Responses

Inspect the exact model output for any coded comment.

Comment
Yudkowsky: An AI will create subgoals which are not foreseeable by the designers. Most of these subgoals will include killing humans. And for humanity to end, it is sufficient if only one of the subgoals includes killing humanity as a side effect. Problems with this argument (in arbitrary order): 1) It is not clear that the jump from AI as smart as humanity, to AI at a level ridiculously smarter than humanity, happens so fast that humanity could not adapt ( e.g. shut it off). 2) Killing humanity is not impossible, but really hard, if humanity can have a look into your brain at any time, analyze you with the help of other AIs, and shut you down or modify you whenever they want. 3) High Intelligence alone does not automatically mean that entities are effective at reaching goals. Look at many intelligent humans ... 4) Humans follow goals and generate subgoals in a similar unforeseeable manner. But these subgoals almost never include killing all of humanity. 5) It should be possible to add high-level goals which eliminate a lot of all potential dangerous subgoals by means of contradiction. 6) Just because the subgoals are not foreseeable, does not mean that most of them will include killing humans. They are unforeseeable and not arbitrary. 7) It is not clear that subgoals are not balanced out by other subgoals. 8) There might not be one AI, but actually several of them balancing each other. 9) .. Probably there are more .. This discussion should have addressed this kind of counter-arguments. Instead Yudkowsky did not actually try to convince anyone, and Wolfram did not address the obvious problems with Yudkowsky's theory, but instead explored all kinds of "1+1 may not be 4 because who knows what a 1 is, in the first place?"
youtube AI Governance 2024-11-14T09:2… ♥ 1
Coding Result
DimensionValue
Responsibilitydeveloper
Reasoningdeontological
Policyregulate
Emotionmixed
Coded at2026-04-27T06:24:53.388235
Raw LLM Response
[ {"id":"ytc_UgwRvWP_k7v_jN9-Te14AaABAg","responsibility":"none","reasoning":"unclear","policy":"unclear","emotion":"indifference"}, {"id":"ytc_Ugyksdh6rn-4hBjfu214AaABAg","responsibility":"developer","reasoning":"deontological","policy":"unclear","emotion":"outrage"}, {"id":"ytc_UgxlTd1d2AkohR8lVSZ4AaABAg","responsibility":"user","reasoning":"consequentialist","policy":"regulate","emotion":"fear"}, {"id":"ytc_UgxF1_HmuOODIl8KiOF4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"liability","emotion":"fear"}, {"id":"ytc_UgzTMs1seu-Hm2wg1tB4AaABAg","responsibility":"developer","reasoning":"consequentialist","policy":"regulate","emotion":"fear"}, {"id":"ytc_Ugzck-R6lKxbvEb8M5Z4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"ban","emotion":"fear"}, {"id":"ytc_UgyyLzF6cJe301DdxjF4AaABAg","responsibility":"developer","reasoning":"deontological","policy":"regulate","emotion":"mixed"}, {"id":"ytc_UgyL07Rq-EVfO1ActR94AaABAg","responsibility":"developer","reasoning":"deontological","policy":"regulate","emotion":"mixed"}, {"id":"ytc_Ugz6Llf_yDF9Gc34V9B4AaABAg","responsibility":"ai_itself","reasoning":"virtue","policy":"industry_self","emotion":"approval"}, {"id":"ytc_UgzaIf0jFeodxvBJt2d4AaABAg","responsibility":"developer","reasoning":"deontological","policy":"ban","emotion":"outrage"} ]