Raw LLM Responses

Inspect the exact model output for any coded comment.

Comment
While the problem that an algorithm does not work correctly is true in this case, I don't think it's connected to bias at all. But you certainly have to point out where the problem lies within the code, so the people behind it can fix it. Focusing on that issue seems more important to me.
youtube 2017-03-31T18:0… ♥ 1
Coding Result
DimensionValue
Responsibilitydeveloper
Reasoningconsequentialist
Policynone
Emotionresignation
Coded at2026-04-27T06:24:53.388235
Raw LLM Response
[ {"id":"ytc_UgzFZ0_gFoBtkA8A8dl4AaABAg","responsibility":"company","reasoning":"deontological","policy":"regulate","emotion":"fear"}, {"id":"ytc_UgwHeHZdwARg_gZKfkp4AaABAg","responsibility":"developer","reasoning":"virtue","policy":"liability","emotion":"outrage"}, {"id":"ytc_UghAtGyqfRzI23gCoAEC","responsibility":"none","reasoning":"consequentialist","policy":"none","emotion":"indifference"}, {"id":"ytc_UghX-XK9nQJeongCoAEC","responsibility":"company","reasoning":"consequentialist","policy":"industry_self","emotion":"resignation"}, {"id":"ytc_UgjF2qj63lIzHHgCoAEC","responsibility":"developer","reasoning":"consequentialist","policy":"none","emotion":"approval"}, {"id":"ytc_UggDn_u7gXt523gCoAEC","responsibility":"developer","reasoning":"virtue","policy":"none","emotion":"outrage"}, {"id":"ytc_UggQG6eUAXHh13gCoAEC","responsibility":"developer","reasoning":"consequentialist","policy":"none","emotion":"resignation"}, {"id":"ytc_Uggt_f6d7mj4O3gCoAEC","responsibility":"developer","reasoning":"consequentialist","policy":"none","emotion":"indifference"}, {"id":"ytc_Ugh6MY8m_2nnh3gCoAEC","responsibility":"developer","reasoning":"consequentialist","policy":"none","emotion":"resignation"}, {"id":"ytc_Ugi88RwR_1WkJXgCoAEC","responsibility":"developer","reasoning":"consequentialist","policy":"regulate","emotion":"approval"} ]