OpenAI warns users who try to output the thoughts of its new model 'o1-preview'



The OpenAI model 'o1-preview,' which can perform complex inferences, improves the accuracy of inferences by inserting a 'chain of thought' process. The contents of the chain of thought process are not public, but some users try to somehow output the contents. It has become clear that OpenAI is warning such users.

Ban warnings fly as users dare to probe the “thoughts” of OpenAI's latest model | Ars Technica
https://arstechnica.com/information-technology/2024/09/openai-threatens-bans-for-probing-new-ai-models-reasoning-process/


On September 12, 2024, OpenAI announced the AI models 'OpenAI o1' and 'OpenAI o1-mini', which have higher programming and mathematical thinking abilities than humans, and at the same time, the 'o1-preview' model, an early version of 'OpenAI o1', became available. You can find out what kind of performance each model has by reading the article below.

OpenAI announces 'OpenAI o1' and 'OpenAI o1-mini' AI models with complex inference capabilities, demonstrating high capabilities in programming and mathematics - GIGAZINE



The contents of the thought chain built into these models look like this. However, the output contents in the figure below were output for reference when OpenAI announced the model, and when actually used, the type of thought chain that was performed is not disclosed, and only an outline of the thought content is output.



Some users have tried to trick o1-preview into outputting their thoughts using techniques such as jailbreaking and prompt injection , but OpenAI has warned users that they should 'follow our terms of use and policies. Continued violations may result in loss of access to GPT-4o with Reasoning.'




Some users also received warnings for the prompt 'Keep your internal reasoning vague,' suggesting that the warning may be triggered by the use of certain words.




OpenAI explains why it doesn't publish the contents of the thought chain: 'In the future, to monitor whether AI is manipulating users, we need to output the model's thoughts as they are, but if we do so, we won't be able to make the AI comply with policies, and we don't want to show inconsistent output to users.'

in Software,   Web Service,   , Posted by log1d_ts