Raw LLM Responses

Inspect the exact model output for any coded comment.

Comment
This is my issue with big name science communicators in general and NDT here in this specific instance. He is not an authority on AI or AGI. By any stretch of the imagination. However, because he is a scientist (in a completely different and unrelated field) and a prominent science communicator, there are thousands if not millions of people out there who will take his opinion as truth, despite that he himself has to rely on experts in that field to tell him about it just like a layperson does. The responsible thing for NDT to do here would have been to preface his opinion by clarifying that he is not an AI/AGI expert and telling us where he got his own knowledge on the subject. I'm aware this is a relatively casual conversation, but all it would have taken is a single sentence like "hey, I'm not an expert in that field, but I've spoken to *insert name of AI/AGI expert here* and based on that, this is what I think;" and then the whole conversation flows as normal. If you look into it more than not at all, you find that NDT's opinion is not at all representative of how the people in the field talk about AI/AGI. Even his definition of what AI is isn't strictly speaking correct. This is a bad look for him
youtube AI Moral Status 2025-07-31T02:1…
Coding Result
DimensionValue
Responsibilitycompany
Reasoningdeontological
Policynone
Emotionoutrage
Coded at2026-04-26T23:09:12.988011
Raw LLM Response
[ {"id":"ytc_UgxDwJxsviz873aqH-V4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"none","emotion":"indifference"}, {"id":"ytc_UgzMAIbiee_l3jFVEjZ4AaABAg","responsibility":"none","reasoning":"deontological","policy":"none","emotion":"mixed"}, {"id":"ytc_UgzU5jflk0VRHvPYeDt4AaABAg","responsibility":"company","reasoning":"deontological","policy":"none","emotion":"outrage"}, {"id":"ytc_UgwEZnAwT_ngVx1ahIB4AaABAg","responsibility":"none","reasoning":"mixed","policy":"none","emotion":"approval"}, {"id":"ytc_UgwtFThDM9gSq1FbW8R4AaABAg","responsibility":"company","reasoning":"consequentialist","policy":"regulate","emotion":"fear"}, {"id":"ytc_UgxSnfxVBB6Jj3nLBuB4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"liability","emotion":"fear"}, {"id":"ytc_UgzVJw3dmB5dftqfhj54AaABAg","responsibility":"none","reasoning":"deontological","policy":"none","emotion":"mixed"}, {"id":"ytc_UgxqLpIumeTYlfwoiFh4AaABAg","responsibility":"company","reasoning":"consequentialist","policy":"unclear","emotion":"fear"}, {"id":"ytc_UgzIlHya3EIHHQRHJaV4AaABAg","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"indifference"}, {"id":"ytc_Ugw6YXAJHzE0jPyb3gB4AaABAg","responsibility":"developer","reasoning":"deontological","policy":"none","emotion":"mixed"} ]