Raw LLM Responses

Inspect the exact model output for any coded comment.

Comment
The AI2027 authors obviously do not understand AI - they project human behaviors on their proposed AGIs. That is bs.
youtube AI Governance 2025-08-12T00:2…
Coding Result
DimensionValue
Responsibilitynone
Reasoningunclear
Policynone
Emotionoutrage
Coded at2026-04-27T06:26:44.938723
Raw LLM Response
[ {"id":"ytc_UgwbrhswIIZ9lYzYU0t4AaABAg","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"indifference"}, {"id":"ytc_Ugz4dUoxsQQrT1vSsv54AaABAg","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"resignation"}, {"id":"ytc_UgwGULoxnOV0OXjVLZd4AaABAg","responsibility":"company","reasoning":"consequentialist","policy":"regulate","emotion":"outrage"}, {"id":"ytc_UgxMIWTSUOCr5kJlmxl4AaABAg","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"indifference"}, {"id":"ytc_UgyhLiQovEnzF6Uqzm14AaABAg","responsibility":"ai_itself","reasoning":"unclear","policy":"none","emotion":"fear"}, {"id":"ytc_UgwgzaPeSPAD0TMgPlt4AaABAg","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"outrage"}, {"id":"ytc_UgxJqgH_Gwl_M2hFH5d4AaABAg","responsibility":"company","reasoning":"virtue","policy":"none","emotion":"outrage"}, {"id":"ytc_UgwgxzBT2JuOw6JVz0V4AaABAg","responsibility":"developer","reasoning":"deontological","policy":"none","emotion":"outrage"}, {"id":"ytc_Ugyac_FUzKwZiVvMS-h4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"none","emotion":"fear"}, {"id":"ytc_UgznZSBbffqJp_a8-QV4AaABAg","responsibility":"none","reasoning":"consequentialist","policy":"none","emotion":"approval"} ]