ChatGPT model that does some Automated Reasoning.
This certainly uses Reinforcement Learning during training time, presumably trying many chains of reasoning on many hard problems, and reinforcing the ones that worked.
It’s possible that they approached the hard problem during training with some sort of tree search, in order to get more training data.
What can we say about how it works?
o1-preview
and 65,536 for the supposedly smaller o1-mini
! These are an increase from the gpt-4o
and gpt-4o-mini
models which both currently have a 16,384 output token limit.”We believe that a hidden chain of thought presents a unique opportunity for monitoring models. Assuming it is faithful and legible, the hidden chain of thought allows us to "read the mind" of the model and understand its thought process. For example, in the future we may wish to monitor the chain of thought for signs of manipulating the user. However, for this to work the model must have freedom to express its thoughts in unaltered form, so we cannot train any policy compliance or user preferences onto the chain of thought. We also do not want to make an unaligned chain of thought directly visible to users.
Therefore, after weighing multiple factors including user experience, competitive advantage, and the option to pursue the chain of thought monitoring, we have decided not to show the raw chains of thought to users. We acknowledge this decision has disadvantages. We strive to partially make up for it by teaching the model to reproduce any useful ideas from the chain of thought in the answer. For the o1 model series we show a model-generated summary of the chain of thought.
This appears to indicate that the chain of thought occurs in language, rather than in some sort of weird reasoning token. It would be very interesting to see what that language stream is like! Is it super technical? Is it readable English? Is it hyper-optimized? The chain of thought exposed through the UI appears to be post-hoc summaries.
The full o1 model appears to be much better: