Raw LLM Responses
Inspect the exact model output for any coded comment.
Look up by comment ID
Random samples — click to inspect
G
So can someone clarify this:
1. If we can guarantee that ai owners will share t…
ytc_UgxqTM0pn…
G
Everyone spitting doom and gloom about AI has obviously not used LLMs since thei…
ytc_UgxeSAKLA…
G
My hope is that these ideas can be tested so most people can agree on what works…
rdc_e2vukml
G
no one is going to hook up some AI to control nuclear power plant nor some elect…
ytc_UgySQV4vD…
G
Maybe I'm the weirdo here but I don't see these AI as monsters.
I see them more …
ytc_UgwsRPcwt…
G
Have you seen the movie Elysium? The rich will have all of their needs provided …
ytc_UgwlT1B9-…
G
We appreciate your perspective. In our live broadcasts on AITube, we delve into …
ytr_Ugy9ASDvq…
G
In some ways, trying to explain a generative learning algorithm is like trying t…
ytc_UgykhxVAY…
Comment
the use of the word "Intelligence" is the problem. We have created an algorithm based tech that mimics human outputs but it is no way intelligent. Relying upon a system that contains no intelligence is a mark of human stupidity. Alignment will never exist because AI isn't intelligent and so human intervention is always required. Not to say "machine learning" isn't useful.
youtube
AI Governance
2025-10-30T11:4…
Coding Result
| Dimension | Value |
|---|---|
| Responsibility | developer |
| Reasoning | deontological |
| Policy | industry_self |
| Emotion | outrage |
| Coded at | 2026-04-26T23:09:12.988011 |
Raw LLM Response
[
{"id":"ytc_Ugw5FMG_q8YjisCwvqJ4AaABAg","responsibility":"developer","reasoning":"virtue","policy":"unclear","emotion":"mixed"},
{"id":"ytc_UgzAY8vNhZk-3OjwBUV4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"none","emotion":"indifference"},
{"id":"ytc_UgxxWx6cIo69q8nPH314AaABAg","responsibility":"company","reasoning":"mixed","policy":"unclear","emotion":"resignation"},
{"id":"ytc_UgwKriY1t3DCQT8Ya5p4AaABAg","responsibility":"unclear","reasoning":"unclear","policy":"unclear","emotion":"unclear"},
{"id":"ytc_UgxkUX2CIqoYfQcnI5V4AaABAg","responsibility":"developer","reasoning":"deontological","policy":"none","emotion":"approval"},
{"id":"ytc_Ugw1TJxC0flXLlbgIId4AaABAg","responsibility":"distributed","reasoning":"consequentialist","policy":"liability","emotion":"fear"},
{"id":"ytc_UgwcfO9bHyJcoYhaQMp4AaABAg","responsibility":"unclear","reasoning":"unclear","policy":"unclear","emotion":"mixed"},
{"id":"ytc_UgxExXwpT9eXmXXIt6R4AaABAg","responsibility":"developer","reasoning":"deontological","policy":"industry_self","emotion":"outrage"},
{"id":"ytc_UgwDU_RS007lj6ACr0J4AaABAg","responsibility":"developer","reasoning":"virtue","policy":"none","emotion":"approval"},
{"id":"ytc_Ugwu_R4LfV8TZxgtgkh4AaABAg","responsibility":"company","reasoning":"consequentialist","policy":"regulate","emotion":"outrage"}
]