Raw LLM Responses
Inspect the exact model output for any coded comment.
Look up by comment ID
Random samples — click to inspect
G
Bruh this is from game detroit become human 💀 not a real robot. There is so much…
ytc_UgypVv_D1…
G
Google, right now, "are tesla self driving cars safer than humans" and top resul…
ytc_UgwcfQchZ…
G
I am sorry but this is just another form of post-training no? You are just telli…
ytc_UgzEXBVqT…
G
I can't care that much about using AI for explicit images or video, most people …
ytc_UgwFwLZ61…
G
Why is the person freaking out about Dans' answer on reversing over population. …
ytc_UgzXe_2v-…
G
I've never seen a group of people become more hostile to the idea of self better…
ytc_Ugz9mVaLe…
G
Being polite to AI actually costs the AI companies millions of dollars in wasted…
ytc_Ugx_C3pxh…
G
Generative AI really feels like the media equivalent of the "Gray Goo" doomsday …
ytc_Ugz6IPzu8…
Comment
You yourself are creating, you yourself are advising , you yourself are wanting , and at the end you yourself are dying
Then what about the other people with less opportunities, less privileges, less money
Will they get monopolized by people like you !
It is the asteroid that ended dinosaurs....and created humans
Asteroid is AI , dinosaurs are us and humans are those with AI and power
youtube
AI Responsibility
2025-08-09T11:4…
Coding Result
| Dimension | Value |
|---|---|
| Responsibility | distributed |
| Reasoning | consequentialist |
| Policy | unclear |
| Emotion | fear |
| Coded at | 2026-04-27T06:26:44.938723 |
Raw LLM Response
[
{"id":"ytc_UgxkelPtOCFuII-wmTh4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"unclear","emotion":"fear"},
{"id":"ytc_Ugz3KvYLsvdSfANH2kl4AaABAg","responsibility":"user","reasoning":"virtue","policy":"none","emotion":"mixed"},
{"id":"ytc_UgwVmuChrgBFsKYvpKR4AaABAg","responsibility":"distributed","reasoning":"unclear","policy":"unclear","emotion":"fear"},
{"id":"ytc_UgwUIxi8-g6q-ta4sYZ4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"unclear","emotion":"fear"},
{"id":"ytc_UgyXgOn9X7jETy7rTXR4AaABAg","responsibility":"user","reasoning":"virtue","policy":"unclear","emotion":"mixed"},
{"id":"ytc_UgziFSzKg8okvLk_oP14AaABAg","responsibility":"unclear","reasoning":"unclear","policy":"unclear","emotion":"mixed"},
{"id":"ytc_UgwBZG3YLA7IRzRLgKl4AaABAg","responsibility":"ai_itself","reasoning":"deontological","policy":"unclear","emotion":"mixed"},
{"id":"ytc_UgyVlp5Nq5oOajiZMyh4AaABAg","responsibility":"user","reasoning":"virtue","policy":"liability","emotion":"approval"},
{"id":"ytc_UgwIZpsB8Nk1m9vJlgB4AaABAg","responsibility":"distributed","reasoning":"consequentialist","policy":"unclear","emotion":"fear"},
{"id":"ytc_UgySfs2vp2VvBmYtRCt4AaABAg","responsibility":"developer","reasoning":"virtue","policy":"liability","emotion":"outrage"}
]