Evolution through large models
Log in to bookmark articles and create collections
AI-Powered Learning bringing you YOUR best news
Log in to bookmark articles and create collections
We trained “critique-writing” models to describe flaws in summaries. Human evaluators find flaws in summaries much more often when shown our model’s critiques. Larger models are better at self-critiquing, with scale improving critique-writing more than summary-writing. This shows promise for using A...
Log in to bookmark articles and create collections
Large neural networks are at the core of many recent advances in AI, but training them is a difficult engineering and research challenge which requires orchestrating a cluster of GPUs to perform a single synchronized calculation.
Log in to bookmark articles and create collections
Cohere, OpenAI, and AI21 Labs have developed a preliminary set of best practices applicable to any organization developing or deploying large language models.
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Codex is now powering 70 different applications across a variety of use cases through the OpenAI API.
Log in to bookmark articles and create collections
Early users have created over 3 million images to date and helped us improve our safety processes. We’re excited to begin adding up to 1,000 new users from our waitlist each week.
Log in to bookmark articles and create collections
We’re happy to announce several executive role changes that reflect our recent progress and will ensure continued momentum toward our next major milestones.
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Goodhart’s law famously says: “When a measure becomes a target, it ceases to be a good measure.” Although originally from economics, it’s something we have to grapple with at OpenAI when figuring out how to optimize objectives that are difficult or costly to measure.
Log in to bookmark articles and create collections
We’ve released new versions of GPT-3 and Codex which can edit or insert content into existing text, rather than just completing existing text.
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
We describe our latest thinking in the hope of helping other AI developers address safety and misuse of deployed models.
Log in to bookmark articles and create collections
Call for expressions of interest to study the economic impacts of large language models.
Log in to bookmark articles and create collections
We built a neural theorem prover for Lean that learned to solve a variety of challenging high-school olympiad problems, including problems from the AMC12 and AIME competitions, as well as two problems adapted from the IMO.
Log in to bookmark articles and create collections
We’ve trained language models that are much better at following user intentions than GPT-3 while also making them more truthful and less toxic, using techniques developed through our alignment research. These InstructGPT models, which are trained with humans in the loop, are now deployed as the defa...
Log in to bookmark articles and create collections
We are introducing embeddings, a new endpoint in the OpenAI API that makes it easy to perform natural language and code tasks like semantic search, clustering, topic modeling, and classification.
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
We’ve fine-tuned GPT-3 to more accurately answer open-ended questions using a text-based web browser.
Log in to bookmark articles and create collections
Fine-tune with a single command.
Log in to bookmark articles and create collections
As part of our effort to support and develop AI talent, we’re excited to announce the OpenAI Residency.
Log in to bookmark articles and create collections
Wider availability made possible by safety progress.
Log in to bookmark articles and create collections
We’ve trained a system that solves grade school math problems with nearly twice the accuracy of a fine-tuned GPT-3 model. It solves about 90% as many problems as real kids: a small sample of 9-12 year olds scored 60% on a test from our dataset, while our system scored 55% on those same problems.
Log in to bookmark articles and create collections
Scaling human oversight of AI systems for tasks that are difficult to evaluate.
Log in to bookmark articles and create collections
Today, we’re excited to announce the appointment of Helen Toner to our board of directors.
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
We’ve created an improved version of OpenAI Codex, our AI system that translates natural language to code, and we are releasing it through our API in private beta starting today.
Log in to bookmark articles and create collections
We’re releasing Triton 1.0, an open-source Python-like programming language which enables researchers with no CUDA experience to write highly efficient GPU code—most of the time on par with what an expert would be able to produce.
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Our latest research finds we can improve language model behavior with respect to specific behavioral values by fine-tuning on a small, curated dataset.
Log in to bookmark articles and create collections