Raw LLM Responses

Inspect the exact model output for any coded comment.

Comment
Honestly I know this is a big theory but as a programmer who has some experience working with AI they cannot and will not become conscious in any way. The way AI works is it is given a specific purpose lets say determining if a object is a chicken. Then it will go through a series of duplicating itself, testing itself with loads of samples that the answer is known by the tester, then the tester will take the AI that has guessed the most right and duplicate it and repeat the process until the AI can guess a object is a chicken 99% of the time. Now for a SUPER AI lets take all these small neuro networks and link them together for their known properties. Lets say we got one that knows languages, one that knows animals, one that knows how to google, and one that knows how to navigate a area. This AI would have some conscious like elements to it however, it will never have any emotions, it will simply be able to interpret what it knows to real world scenarios, it will not be 100% correct however it may be able to hold a conversation for a bit until it starts to make no sense because again it can't think it can only guess something based on previous correct sample answers. It may be able to also walk around until eventually it believes a wall is in fact the floor and somehow breaks itself. idk but you get the point it is impossible for a AI/ROBOT to truly become conscious and that is honestly a good thing. At its lowest level it is only interpreting a series of 8 one's and zero's at a time
youtube AI Moral Status 2022-01-09T04:1…
Coding Result
DimensionValue
Responsibilitynone
Reasoningmixed
Policynone
Emotionindifference
Coded at2026-04-27T06:24:59.937377
Raw LLM Response
[ {"id":"ytc_Ugz1gKUCEMWwCARjHox4AaABAg","responsibility":"developer","reasoning":"deontological","policy":"ban","emotion":"outrage"}, {"id":"ytc_UgwYHC1hH8ZXTJg-FDB4AaABAg","responsibility":"none","reasoning":"mixed","policy":"none","emotion":"indifference"}, {"id":"ytc_Ugy5aYamGjLJGN7npVp4AaABAg","responsibility":"none","reasoning":"deontological","policy":"regulate","emotion":"approval"}, {"id":"ytc_UgxWLQNzzbrUeVItFwl4AaABAg","responsibility":"developer","reasoning":"deontological","policy":"liability","emotion":"outrage"}, {"id":"ytc_UgxpCvZx8XsSV4GHt854AaABAg","responsibility":"none","reasoning":"contractualist","policy":"regulate","emotion":"approval"}, {"id":"ytc_UgzV-9b1bUCzsgeg7AR4AaABAg","responsibility":"developer","reasoning":"deontological","policy":"industry_self","emotion":"indifference"}, {"id":"ytc_UgyxTnewRgSJmaseNAV4AaABAg","responsibility":"none","reasoning":"deontological","policy":"ban","emotion":"outrage"}, {"id":"ytc_Ugww_b_teocVEo2uuHB4AaABAg","responsibility":"none","reasoning":"consequentialist","policy":"none","emotion":"fear"}, {"id":"ytc_UgxB1xJbGcmPqASVsi94AaABAg","responsibility":"none","reasoning":"deontological","policy":"regulate","emotion":"approval"}, {"id":"ytc_UgxEjH_zbRBLuZiRR_h4AaABAg","responsibility":"none","reasoning":"unclear","policy":"unclear","emotion":"mixed"} ]