Powering next generation applications with OpenAI Codex Codex is now powering 70 different applications across a variety of use cases through the OpenAI API.…
Measuring Goodhart’s law Goodhart’s law famously says: “When a measure becomes a target, it ceases to be a good measure.” Although originally from economics, it’s something we have to grapple with at OpenAI when figuring out how to optimize objectives that are difficult…
Lessons learned on language model safety and misuse We describe our latest thinking in the hope of helping other AI developers address safety and misuse of deployed models.…
Solving (some) formal math olympiad problems We built a neural theorem prover for Lean that learned to solve a variety of challenging high-school olympiad problems, including problems from the AMC12 and AIME competitions, as well as two problems adapted from the IMO.…
Aligning language models to follow instructions We’ve trained language models that are much better at following user intentions than GPT-3 while also making them more truthful and less toxic, using techniques developed through our alignment research. These InstructGPT models, which are trained with humans in the…