Raw LLM Responses

Inspect the exact model output for any coded comment.

Comment
I tried ChatGPT to help research a sermon, and it gave very fine sounding theological quotes. I would double check each, and in almost ALL cases the quotes were made up. The one quote that was accurate was attributed to the wrong theologian. When challanged initially for refernces ChatGPT would make up refences and then try and explain the quote. When challanged directly it would then admit it makes stuff up. It was easy to see through the AI facade if you had the simple discipline to always check quotes and reference from original sources. Adoption of AI without thorough research will lead to loads of this. Although, to be fair, ChatGPT does a great job at emulating a real human.... if that human was a liar who just makes up stuff to try and bamboozle you :)
youtube AI Responsibility 2023-06-22T09:4…
Coding Result
DimensionValue
Responsibilityai_itself
Reasoningconsequentialist
Policyliability
Emotionoutrage
Coded at2026-04-27T06:26:44.938723
Raw LLM Response
[ {"id":"ytc_UgyvONssAtPiQd8nQ754AaABAg","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"approval"}, {"id":"ytc_UgzKUTDvS_WODcFuMkx4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"unclear","emotion":"fear"}, {"id":"ytc_Ugw1GGxyjlVbhEngQ_J4AaABAg","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"indifference"}, {"id":"ytc_UgwX48gPD3tgbVlbHpB4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"liability","emotion":"outrage"}, {"id":"ytc_Ugz4fr9MCbpwqkVXswl4AaABAg","responsibility":"unclear","reasoning":"unclear","policy":"unclear","emotion":"fear"}, {"id":"ytc_Ugyq0kAvNFGLZw3rWLd4AaABAg","responsibility":"user","reasoning":"deontological","policy":"none","emotion":"indifference"}, {"id":"ytc_UgxbGThnct8U6zDYUO54AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"regulate","emotion":"outrage"}, {"id":"ytc_UgygKaXohmripXAyaZh4AaABAg","responsibility":"user","reasoning":"deontological","policy":"liability","emotion":"outrage"}, {"id":"ytc_Ugxtd8u5Wll6kqVg8W94AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"none","emotion":"resignation"}, {"id":"ytc_Ugz_DzXBW3yOtTiuqId4AaABAg","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"indifference"} ]