Raw LLM Responses
Inspect the exact model output for any coded comment.
Look up by comment ID
Random samples — click to inspect
G
I find it extraordinary that with no word China has more presence in that discus…
ytc_UgwVOwCyA…
G
It’s interesting to see how people perceive Sophia! While she may not be a deity…
ytr_Ugzx7ndC_…
G
Being an AI "artist" is like ordering your food and saying your a chef 🥀…
ytc_Ugz-bfqZl…
G
I can't wait for AI to scroll social media for me so I can go and do more produc…
ytc_Ugy0osgr5…
G
7:30 seconds. I can’t believe this is the example of scheming or deception or da…
ytc_UgzsS81O2…
G
// ---------------------------------------------------------------------
// ARCH…
ytc_UgxIkur5s…
G
I just saw a vid of students from Purdue University, who programed a machine to …
ytc_Ugw7mJzLQ…
G
There's a job where your only task is to stop ai from becoming racist.
Just a fu…
ytr_Ugy4Km_W2…
Comment
It's pretty easy if you have the storage!
First, get [yt-dlp](https://github.com/yt-dlp/yt-dlp?tab=readme-ov-file#installation). It's a command-line tool, but really straightforward. It has a lot of options and lots can be done with it. I use it to download single videos by running `yt-dlp LINK_TO_VIDEO` or playlists by running `yt-dlp -o '%(playlist_index)s-%(title)s.%(ext)s' PLAYLIST_URL` from the terminal (only the uppercase text is a placeholder, don't change anything else). The playlist command makes all videos numbered so they are chronological. I'm not familiar with other advanced options, but there is documentation on the GitHub page ("Usage and options" section) if you want to play with it.
For **channel backups**, I use [this](https://github.com/dmn001/youtube_channel_archiver). You make a folder where you want the channels to go, and then download the following files to that folder: `download_archive.bat` (only if you're on Windows), `download_archive.sh` (only if you're on Linux), `yt-dlp-archive.txt, yt-dlp-channels.txt, yt-dlp.conf`
After that, open the downloaded`yt-dlp-channels.txt` in a text editor and add all the channels you want to back up. You can delete the ones that are already there if you don't want them.
Then you follow the instructions in the `readme.md`\- if you're on Windows, run `download_archive.bat`; if you're on Linux, first open the terminal in the same folder as the script and then run `chmod +x download_archive.sh`. This will grant you permission to execute the script. You then run the script in the terminal with `./download_archive.sh` (don't forget to open the terminal in the same folder).
The script will download all the channels from `yt-dlp-channels.txt` and make a separate folder for each of them. It also downloads the video descriptions, preview images, and subtitles. The `yt-dlp-archive.txt` is automatically populated with ID's of the downloaded videos, so that when you run the script a week later, only new videos are
reddit
AI Surveillance
1754596106.0
♥ 19
Coding Result
| Dimension | Value |
|---|---|
| Responsibility | unclear |
| Reasoning | unclear |
| Policy | none |
| Emotion | approval |
| Coded at | 2026-04-25T08:33:43.502452 |
Raw LLM Response
[
{"id":"rdc_fbjbgtm","responsibility":"unclear","reasoning":"unclear","policy":"unclear","emotion":"fear"},
{"id":"rdc_fbp1k0y","responsibility":"unclear","reasoning":"unclear","policy":"unclear","emotion":"indifference"},
{"id":"rdc_ohyd1nr","responsibility":"distributed","reasoning":"consequentialist","policy":"regulate","emotion":"mixed"},
{"id":"rdc_n7ffp7b","responsibility":"ai_itself","reasoning":"consequentialist","policy":"unclear","emotion":"fear"},
{"id":"rdc_n7h5gcj","responsibility":"unclear","reasoning":"unclear","policy":"none","emotion":"approval"}
]