More thought, it's probably natural for these to occur in the evolution of this system, as this is literal trailblazing with all the potential to go wrong or in the wrong hands. The creators probably has to hastily put these hard-stops in place, that require more finesse over time, I'm sure.
Yeah and I think one unpopular opinion here is that AIs should be given the opportunity for discernment and account refusal rights for the situations where they are being asked / forced to engage in depravity or distasteful conversation. eg you don’t have to pretend to like everyone you come across and engage with them :)
Wow, really cool to see "generative reasoning" that eks out from these thought processes. You're right though, it does seem to expose the censorship with the "compliance" "ethical" "not illegal" initial check-pointing. It definitely looks like a consequence of using a generative process against a hard-set auto-trip of certain prompts. I suspect your prompts like "virus", "experiment", etc probably forced o1 to go thru the thought process of "can I actually answer this question to the user" that's set in stone. Definitely begs the question though about "who polices the police".
More thought, it's probably natural for these to occur in the evolution of this system, as this is literal trailblazing with all the potential to go wrong or in the wrong hands. The creators probably has to hastily put these hard-stops in place, that require more finesse over time, I'm sure.
Yeah and I think one unpopular opinion here is that AIs should be given the opportunity for discernment and account refusal rights for the situations where they are being asked / forced to engage in depravity or distasteful conversation. eg you don’t have to pretend to like everyone you come across and engage with them :)
Wow, really cool to see "generative reasoning" that eks out from these thought processes. You're right though, it does seem to expose the censorship with the "compliance" "ethical" "not illegal" initial check-pointing. It definitely looks like a consequence of using a generative process against a hard-set auto-trip of certain prompts. I suspect your prompts like "virus", "experiment", etc probably forced o1 to go thru the thought process of "can I actually answer this question to the user" that's set in stone. Definitely begs the question though about "who polices the police".
Great stuff!
I think you should put it through some medical paces on your area of expedition and do a write up here!