Raw LLM Responses

Inspect the exact model output for any coded comment.

Comment
5:05 Hmmm...sounds like we should get rid of AI then while we still have the chance!
youtube AI Harm Incident 2025-11-12T05:1…
Coding Result
DimensionValue
Responsibilityunclear
Reasoningconsequentialist
Policyban
Emotionfear
Coded at2026-04-27T06:26:44.938723
Raw LLM Response
[ {"id":"ytc_UgyXe0hLw6d25m9Rlit4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"unclear","emotion":"fear"}, {"id":"ytc_UgxZJ5t283nspuhP-wl4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"unclear","emotion":"fear"}, {"id":"ytc_Ugw8FmE8x3DLkJP4kcp4AaABAg","responsibility":"unclear","reasoning":"consequentialist","policy":"ban","emotion":"fear"}, {"id":"ytc_UgyEGMbu7UkdKUVGYK94AaABAg","responsibility":"company","reasoning":"consequentialist","policy":"regulate","emotion":"fear"}, {"id":"ytc_UgwpWNd8byf5Y8LLHq94AaABAg","responsibility":"government","reasoning":"consequentialist","policy":"unclear","emotion":"fear"}, {"id":"ytc_UgzbFFFy62O3Twzzjux4AaABAg","responsibility":"developer","reasoning":"virtue","policy":"unclear","emotion":"outrage"}, {"id":"ytc_UgwCEpoUCAtU09zsohN4AaABAg","responsibility":"developer","reasoning":"virtue","policy":"unclear","emotion":"resignation"}, {"id":"ytc_UgxB2cyjRqSyevfD6Id4AaABAg","responsibility":"company","reasoning":"consequentialist","policy":"ban","emotion":"outrage"}, {"id":"ytc_UgxPoX8QtVSSqXfUwu14AaABAg","responsibility":"unclear","reasoning":"consequentialist","policy":"unclear","emotion":"indifference"}, {"id":"ytc_Ugyb1dC6n8peXiK3nAx4AaABAg","responsibility":"unclear","reasoning":"unclear","policy":"unclear","emotion":"unclear"} ]