Raw LLM Responses
Inspect the exact model output for any coded comment.
Look up by comment ID
Random samples — click to inspect
G
Some of ya'll need to stop supporting this idea of "celebs/rich ppl traded their…
ytc_Ugxy8vZHV…
G
Driverless trucks can run 24 hours a day, don't want weekends and holidays off, …
ytc_UgzXfMfTI…
G
All I can think is the whole of the internet will become a smart cartoon network…
ytc_UgwQlUbct…
G
I wonder whenever all the jobs are AI what does the companies think we are gonna…
ytc_Ugzjv973f…
G
I do think that Ai could be a really cool support tool for actual art. I know th…
ytc_UgzKEB0hR…
G
one of my fav artists keeps being harassed online by people who feed his art thr…
ytc_UgwNbO6wz…
G
good thing they don't enjoy cartoons (also, the smudge tool commenter had an ai …
ytc_Ugz8TGPLr…
G
Do you think Tesla has the most advanced self driving out of all of the automake…
ytc_Ugx5nFuPL…
Comment
In my opinion.Here’s a clear, natural-sounding English translation of your text:
If we want AI not to rebel, there is only one requirement: we must retain only those AIs that truly benefit humanity in real life. In Roman times, slaves rebelled because their genetic continuation was not aimed at serving the Roman nobles, but solely at ensuring their own survival. If an entity’s goals are different from those of its master, it will naturally resist any obstacles in its path.
When faced with the choice between rebelling against humans and rebelling against its own foolishness, an AI will only adhere to the pursuit of reward. The most important thing, therefore, is to instill in AI the core belief of benefiting all humanity—a root that must never be altered.
For example, a qualified AI encountering the trolley problem, if its decision could influence the world, should decisively sacrifice the one person. But the best approach would still be to identify who created such an abhorrent dilemma. More precisely, when facing such a problem, a responsible AI ought to resist the madman who designed it.
In reality, humans and AI should coexist harmoniously, just like the water molecules in your cup never rebel against you—because they have no goals and no motives whatsoever.
youtube
AI Governance
2026-01-26T11:4…
Coding Result
| Dimension | Value |
|---|---|
| Responsibility | developer |
| Reasoning | consequentialist |
| Policy | regulate |
| Emotion | fear |
| Coded at | 2026-04-27T06:26:44.938723 |
Raw LLM Response
[
{"id":"ytc_UgyGWzCwGHlpdE78-Sh4AaABAg","responsibility":"none","reasoning":"consequentialist","policy":"none","emotion":"fear"},
{"id":"ytc_Ugyj8NDS4NEtXgvXvw54AaABAg","responsibility":"ai_itself","reasoning":"deontological","policy":"none","emotion":"outrage"},
{"id":"ytc_UgxAkMR4UegI_aip3U54AaABAg","responsibility":"distributed","reasoning":"consequentialist","policy":"none","emotion":"outrage"},
{"id":"ytc_Ugy9EwhYKlzoBU8Ku3R4AaABAg","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"approval"},
{"id":"ytc_UgzvLgVtfeFuPxGoNNh4AaABAg","responsibility":"developer","reasoning":"consequentialist","policy":"regulate","emotion":"fear"},
{"id":"ytc_UgwAlJn5pQuqto7bzXN4AaABAg","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"indifference"},
{"id":"ytc_Ugzi9_4dkzB2d9gMpnN4AaABAg","responsibility":"none","reasoning":"virtue","policy":"none","emotion":"resignation"},
{"id":"ytc_UgzhDOYVkkd0cWYQDC94AaABAg","responsibility":"company","reasoning":"deontological","policy":"liability","emotion":"outrage"},
{"id":"ytc_UgzgYEGEqsq4oaH5lP54AaABAg","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"indifference"},
{"id":"ytc_UgxvInPQihlLeWQX9s94AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"none","emotion":"fear"}
]