Raw LLM Responses
Inspect the exact model output for any coded comment.
Look up by comment ID
Random samples — click to inspect
G
This is so funny to read so many comments actually believing this is a real robo…
ytc_UgwIe44Ii…
G
I'm saving this video, so I can Hype it later, when I get more hype points back.…
ytc_UgyQaMjGM…
G
I use AI generated art as inspiration and, indeed, for "reference". AI's ability…
ytc_UgyCELarv…
G
did that man just say that they take shit off of shows and movies in order to 'i…
ytc_Ugxa8XyuU…
G
People who call themselves "a.i artists" are not even artists. And saying that y…
ytc_UgwWALieG…
G
@polygondeath0 taking inspiration and learning from others is completely differe…
ytr_UgzJs9wur…
G
He seems like he’s romanticized AI a little too much. He’s like a guy who people…
ytc_UgzXO5WDK…
G
The main reason why Tesla autopilot crashes into motorcycles or places of accide…
ytc_Ugy0jMQZO…
Comment
I'll give it a go:
Based on the design and function of an LLM, it explicitly doesn't meet Jaynes' description of consciousness, no? Jaynes proposed that the generation of language was functionally the moment consciousness was invented, and this has overlap with the Chomskian idea of Generative Grammar i.e. that humans have a genetic predisposition to generate grammars and by extension, languages. (in general linguistics in the 50s - 70s was super invested in this idea that language and consciousness or the ability to comprehend are inexorably linked).
If the generation of grammar and language is the marker of consciousness then LLMs very explicitly are not conscious under Jaynes' description. An LLM "generates" grammar only as dictated by human description, and only functions because it must rely on an expansive history of human language from which to mimic. Semantically it isn't the same as the "generation" linguists talk about, including that there is still debate over how much of humans' predisposition for language is genetic.
As a side note, the view that language is the window to consciousness is linked with the Sapir-Whorf hypothesis that language is effectively both the tool for understanding the world and the limit of understanding (e.g. if you don't know the word "blue" you cannot comprehend it as different from any other colour because you have no word for it). Sapir-Whorf has had a lot of impact, and informs a lot of modern linguistic theory, but as a view of how language actually works is considered archaic and fairly disproven as an accurate description for how language interacts with comprehension of the world around you.
Tl;dr Jaynes' view proposed that human language is a reflection of consciousness, but LLMs are only imitators of language and so could only be imitations of that consciousness. Anything further is dipping into OP's point, that you are seeing LLMs work and mistaking it for thought and human generation of language, when it's only a
reddit
AI Moral Status
1749763306.0
♥ 15
Coding Result
| Dimension | Value |
|---|---|
| Responsibility | none |
| Reasoning | unclear |
| Policy | none |
| Emotion | indifference |
| Coded at | 2026-04-25T08:33:43.502452 |
Raw LLM Response
[
{"id":"rdc_mxwabjj","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"indifference"},
{"id":"rdc_mxfgu4n","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"approval"},
{"id":"rdc_mxfymfg","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"approval"},
{"id":"rdc_mxfrkq3","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"mixed"},
{"id":"rdc_mxgbs4b","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"indifference"}
]