Understanding the capabilities, limitations, and societal impact of large language models
Log in to bookmark articles and create collections
AI-Powered Learning bringing you YOUR best news
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Weβve scaled Kubernetes clusters to 7,500 nodes, producing a scalable infrastructure for large models likeΒ GPT-3,Β CLIP, andΒ DALLΒ·E, but also for rapid small-scale iterative research such asΒ Scaling Laws for Neural Language Models.
Log in to bookmark articles and create collections
Weβve trained a neural network called DALLΒ·E that creates images from text captions for a wide range of concepts expressible in naturalΒ language.
Log in to bookmark articles and create collections
Weβre introducing a neural network called CLIP which efficiently learns visual concepts from natural language supervision. CLIP can be applied to any visual classification benchmark by simply providing the names of the visual categories to be recognized, similar to the βzero-shotβ capabilities of GP...
Log in to bookmark articles and create collections
Itβs been a year of dramatic change and growth at OpenAI.
Log in to bookmark articles and create collections
OpenAI has agreed to license GPT-3 to Microsoft for their own products and services.
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Weβve applied reinforcement learning from human feedback to train language models that are better at summarization.
Log in to bookmark articles and create collections
Our third class of OpenAI Scholars presented their final projects at virtual Demo Day, showcasing their research results from over the past five months.
Log in to bookmark articles and create collections
Weβre excited to announce that OpenAI is co-organizing two NeurIPS 2020 competitions with AIcrowd, Carnegie Mellon University, and DeepMind, using Procgen Benchmark and MineRL.
Log in to bookmark articles and create collections
We find that, just as a large transformer model trained on language can generate coherent text, the same exact model trained on pixel sequences can generate coherent imageΒ completionsΒ andΒ samples. By establishing a correlation between sample quality and image classification accuracy, we show that ou...
Log in to bookmark articles and create collections
Weβre releasing an API for accessing new AI models developed by OpenAI.
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Weβre releasing an analysis showing that since 2012 the amount of compute needed to train a neural net to the same performance on ImageNetΒ classification has been decreasing by a factor of 2 every 16 months. Compared to 2012, it now takes 44 times less compute to train a neural network to the level...
Log in to bookmark articles and create collections
Weβre introducing Jukebox, a neural net that generates music, including rudimentary singing, as raw audio in a variety of genres and artist styles. Weβre releasing the model weights and code, along with a tool to explore the generatedΒ samples.
Log in to bookmark articles and create collections
Weβve contributed to a multi-stakeholder report byΒ 58 co-authorsΒ at 30 organizations, including theΒ Centre for the Future of Intelligence,Β Mila,Β Schwartz Reisman Institute for Technology and Society,Β Center for Advanced Study in the Behavioral Sciences, andΒ Center for Security and Emerging Technolog...
Log in to bookmark articles and create collections
Weβre introducingΒ OpenAI Microscope, a collection of visualizations of every significant layer and neuron of eight vision βmodel organismsβ which are often studied in interpretability. Microscope makes it easier to analyze the features that form inside these neural networks, and we hope it will help...
Log in to bookmark articles and create collections
We are standardizing OpenAIβs deep learning framework onΒ PyTorch.
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
We show that theΒ doubleΒ descentΒ phenomenonΒ occurs in CNNs, ResNets, and transformers: performance first improves, then gets worse, and then improves again with increasing model size, data size, or training time. This effect is often avoided through careful regularization. While this behavior appears...
Log in to bookmark articles and create collections
Weβre releasing Procgen Benchmark, 16 simple-to-useΒ procedurally-generatedΒ environments which provide a direct measure of how quickly a reinforcement learning agent learns generalizableΒ skills.
Log in to bookmark articles and create collections
Log in to bookmark articles and create collections
Weβre releasing Safety Gym, a suite of environments and tools for measuring progress towards reinforcement learning agents that respect safety constraints whileΒ training.
Log in to bookmark articles and create collections
As the final model release ofΒ GPT-2βsΒ staged release, weβre releasing the largest version (1.5B parameters) of GPT-2 along withΒ code and model weightsΒ to facilitate detection of outputs of GPT-2 models. While there have been larger language models released since August, weβve continued with our orig...
Log in to bookmark articles and create collections
Weβve trained a pair of neural networks to solve the Rubikβs Cube with a human-like robot hand. The neural networks are trained entirely in simulation, using the same reinforcement learning code asΒ OpenAI FiveΒ paired with a new technique called Automatic Domain Randomization (ADR). The system can ha...
Log in to bookmark articles and create collections
We are now accepting applications for our third class of OpenAI Scholars.
Log in to bookmark articles and create collections
Weβve fine-tuned the 774M parameter GPT-2 language model using human feedback for various tasks, successfully matching the preferences of the external human labelers, though those preferences did not always match our own. Specifically, for summarization tasks the labelers preferred sentences copied...
Log in to bookmark articles and create collections
Weβve observed agents discovering progressively more complex tool use while playing a simple game of hide-and-seek. Through training in our new simulated hide-and-seek environment, agents build a series of six distinct strategies and counterstrategies, some of which we did not know our environment s...
Log in to bookmark articles and create collections
Weβve developed a method to assess whether a neural network classifier can reliably defend against adversarial attacks not seen during training. Our method yields a new metric, UAR (Unforeseen Attack Robustness), which evaluates the robustness of a single model against an unanticipated attack, and h...
Log in to bookmark articles and create collections