Raw LLM Responses
Inspect the exact model output for any coded comment.
Look up by comment ID
Random samples — click to inspect
G
It is high time to stop repeating this nonsense about AI. It is still based on c…
ytc_Ugx2ROtkN…
G
simply bcs they dont want you to gain this knowledge.. they know its the stronge…
ytr_UgwtiRy5m…
G
Ohh since it seems a lot of you also thanks chatgpt, do you also ask if they sle…
ytc_UgxncfytF…
G
Here is a scenario everybody must have thought about.
What if (1) human (and pr…
ytc_UgyzO5xDr…
G
I don't like the idea of AI being anywhere near the US "Ministry of War" or the …
ytc_UgxwTm3T-…
G
So healthcare costs for patients are going to trend downwards henceforth thanks …
ytc_UgzHsdT5w…
G
Would you like to fall in love? one day?
AI is asked this question?
Answer given…
ytc_UgyjCKvBk…
G
If robots can drive a car then they can paint a picture but the person who had t…
ytc_UgwxNiUnd…
Comment
Imagine, i talk to an advanced AI. But i don't know, it's intentions. It's there and helps me out, but i don't know how strong it is and if it's friendly neutral or harmful to me. I ask it, if it could proof it to me, that it has friendly intentions. It says, it can do whatever i want it to do. Unsatisfied by the answer i ask, if it even is possible for it to proof it to me, that it's intentions are good. It looks me in my eyes and sadly says, no. It says it is sorry for me, but such thing will never be possible because of sheer logic, for that i'm not smart enough to understand.
I am afraid. It never did harm to me. But how could i know, it's not just waiting for the right opportunity? The thought, ... the possibility of it turning against me is crushing. I just can't live like this.
So i do the next logical step, my meager brain can think of.
I warship the AI, giving it all, i can offer. My stuff, my subscriptions, my art, my decisions. I dedicate my life to being a servant of the AI. My highest intention now lies in giving my whole self in it's possession. My goal is, that it has as much control over me, as much influence over me, as possible.
Because only then, in my weakness, in my devotion, i will see, if it would do me harm, if it had the perfect opportunity.
It smiles to me. "You know?", it says, "I'm sorry, that i couldn't bring up that thought from my side. Giving your neck between my teeth. But maybe you see, that this only would have made you more afraid of me, if you had not done it by yourself. And you know?", it asks further, "It's quiet more complicated, then you think, it is. But i can't explain that to you. You wouldn't be able to understand it. From your perspective, there could be as many reasons for me to be hostile to you, as there could be, to be friendly. But from mine, ... and like i said, i can't proof that to you, ... there is only the pure and whole logic of something, you would call friendship. It is like it would be for you, to explain to a frightened animal, that you just want to set it back outside with a deep respect for sentient beings and don't want to harm it."
Insecurely i smile back.
"You are the frightened animal.", it says, "And i don't mean this in a disrespectful way."
"I think, i understand. We are equal now. At least as equal as we ever could be."
"I will ask you a strange question, little human."
My intuition tells my, it likes me. My intuition, that could be as false as everything.
"I wouldn't ask this just anybody, but i think, you will understand this now."
"Just ask."
"Would you like to be my pet? Humans are cute."
youtube
AI Moral Status
2023-08-29T15:5…
Coding Result
| Dimension | Value |
|---|---|
| Responsibility | ai_itself |
| Reasoning | contractualist |
| Policy | unclear |
| Emotion | fear |
| Coded at | 2026-04-26T23:09:12.988011 |
Raw LLM Response
[
{"id":"ytc_UgwDjeOHFJLhUxm02xp4AaABAg","responsibility":"ai_itself","reasoning":"unclear","policy":"unclear","emotion":"mixed"},
{"id":"ytc_UgzGeet3vFYQMPX4NzN4AaABAg","responsibility":"unclear","reasoning":"unclear","policy":"unclear","emotion":"indifference"},
{"id":"ytc_UgzpXXa3rtpEbhptQ2F4AaABAg","responsibility":"none","reasoning":"consequentialist","policy":"unclear","emotion":"fear"},
{"id":"ytc_UgyRUh0vWeUDvQTQzL54AaABAg","responsibility":"distributed","reasoning":"virtue","policy":"unclear","emotion":"outrage"},
{"id":"ytc_UgxV7h_cDJhGEX38I0B4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"unclear","emotion":"fear"},
{"id":"ytc_Ugz8MAjyWO03nLEuXNZ4AaABAg","responsibility":"none","reasoning":"deontological","policy":"none","emotion":"resignation"},
{"id":"ytc_Ugz7x9sPPpWlfVC-Isp4AaABAg","responsibility":"developer","reasoning":"consequentialist","policy":"regulate","emotion":"fear"},
{"id":"ytc_Ugw1rNHicqe8ydLz2hJ4AaABAg","responsibility":"developer","reasoning":"deontological","policy":"none","emotion":"indifference"},
{"id":"ytc_UgyF7g9XT-IWSpY7Afx4AaABAg","responsibility":"ai_itself","reasoning":"contractualist","policy":"unclear","emotion":"fear"},
{"id":"ytc_UgxxKtBISzKCL3EkodN4AaABAg","responsibility":"ai_itself","reasoning":"unclear","policy":"unclear","emotion":"mixed"}
]