Raw LLM Responses
Inspect the exact model output for any coded comment.
Look up by comment ID
Random samples — click to inspect
G
o1's Chain of Thought:
**Thinking**
**Dropping a tab**
I'm dropping 400μg of …
rdc_m010g0v
G
Ai should only ever be used for things like those live translation glasses that …
ytc_Ugzc28gU9…
G
Stop supporting companies with A.I platforms, no tax credits for A.I. agents onl…
ytc_Ugxx6C_De…
G
My only argument in favor of using AI for artistic applications is purely for re…
ytc_UgxabqjHd…
G
Sure u can create it, but there should be a completely different title for u lot…
ytr_Ugximm40B…
G
I have always hated AI in everyone of its forms and I have always seen that they…
ytc_Ugy8QcdoX…
G
The scariest thing in the world is a Jew involved in AI hell bent on tikkun olam…
ytc_UgwrM84cj…
G
I continue to see the possibilities of AI downplayed by “experts” but then I als…
ytc_UgyUb6KuJ…
Comment
Interesting here on "multiple" counts - for presentation style (2x AI's) to the subject matter that has some nuance - that admits AGI might be a safety catch to SAI - and here at 8:15 it's all summarized - but I must tell you - earnestly - that you guys do not know what AGI is yet. You still don't know what being Human is yet. They are not the same thing - but there are multiple multiple paths to very good AGI. In a sense we are witnessing the laying down of deep genetic lines - and the establishment of distinct new species that even so - will be able to overlap and work together and even blend. I'm glad I'm not writing this stuff.
youtube
AI Governance
2024-01-04T08:2…
♥ 1
Coding Result
| Dimension | Value |
|---|---|
| Responsibility | unclear |
| Reasoning | mixed |
| Policy | unclear |
| Emotion | indifference |
| Coded at | 2026-04-26T23:09:12.988011 |
Raw LLM Response
[
{"id":"ytc_UgyJnXmJnUjI4BZw5cd4AaABAg","responsibility":"unclear","reasoning":"mixed","policy":"unclear","emotion":"indifference"},
{"id":"ytc_UgwpYg_nreZwNpknQop4AaABAg","responsibility":"distributed","reasoning":"consequentialist","policy":"unclear","emotion":"resignation"},
{"id":"ytc_UgzjYp7Oaf7u7i6xvml4AaABAg","responsibility":"company","reasoning":"deontological","policy":"regulate","emotion":"outrage"},
{"id":"ytc_Ugzm3XB7nXqrSR1ypFZ4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"unclear","emotion":"fear"},
{"id":"ytc_UgwTPRPTwe106_KghyB4AaABAg","responsibility":"none","reasoning":"consequentialist","policy":"none","emotion":"resignation"},
{"id":"ytc_UgyNfYpULfh_Ir9Ix1R4AaABAg","responsibility":"unclear","reasoning":"mixed","policy":"unclear","emotion":"approval"},
{"id":"ytc_UgwO_VzN-pF3q4Py3c54AaABAg","responsibility":"unclear","reasoning":"unclear","policy":"unclear","emotion":"indifference"},
{"id":"ytc_UgwkgvHDi4UVDVYQ8Ml4AaABAg","responsibility":"developer","reasoning":"deontological","policy":"unclear","emotion":"mixed"},
{"id":"ytc_Ugxm3y0DqVd6ftgejit4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"unclear","emotion":"fear"},
{"id":"ytc_UgxD2Yiu9gI7Z8nwxY54AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"unclear","emotion":"fear"}
]