AIBullishOpenAI News ยท Dec 207/107
๐ง
Deliberative alignment: reasoning enables safer language models
OpenAI introduces deliberative alignment, a new safety strategy for their o1 models that directly teaches AI systems safety specifications and how to reason through them. This approach aims to make language models safer by incorporating reasoning capabilities into the alignment process.