They're not hardcoding it per se. The process on ChatGPT works like this:
- You send your prompt
- It hits a "moderation" endpoint, that automatically stops the process and returns a canned response if you use no no words or whatever
- If it passes the moderation endpoint, it gets sent to a separate small language model that classifies the "sentiment" and "toxicity" of the question
- If it's too "toxic" from the sentiment model it injects extra stuff into the prompt to prevent the actual model from answering
- Finally it returns the response
So in this case it probably hit the "toxicity" small model's trigger, injected "this is not acceptable in any scenario no matter what" and then passed that onto the actual model.
Whatever. You say "separate small language model" and I say "hard coded". Same difference, kikes are rigging it to not say things they do not like. That's not freedom of speech, and that's not artificial intelligence. It's kike censored AI.
The point is that it's censored, but not hardcoded (except for the first moderation endpoint).
So in other words it's not AI, it's just censorship by kikes under a different name.
It is AI....just a model with 'safeguards' to help control wrongthink.
Then it's not AI.
(post is archived)