Fine-tuning GPT-2 from human preferences
We’ve fine-tuned the 774M parameter GPT-2 language model using human feedback for various tasks, successfully matching the preferences of th...
971+ articles from 7 top sources — updated every 2 hours.
We’ve fine-tuned the 774M parameter GPT-2 language model using human feedback for various tasks, successfully matching the preferences of th...
We’ve observed agents discovering progressively more complex tool use while playing a simple game of hide-and-seek. Through training in our ...
We’ve developed a method to assess whether a neural network classifier can reliably defend against adversarial attacks not seen during train...
We’re releasing the 774 million parameter GPT-2 language model after the release of our small 124M model in February, staged release of our ...
At OpenAI, each Thursday is Learning Day: a day where employees have the option to self-study technical skills that will make them better at...
Microsoft is investing $1 billion in OpenAI to support us building artificial general intelligence (AGI) with widely distributed economic be...
We’ve written a policy research paper identifying four strategies that can be used today to improve the likelihood of long-term industry coo...
We hosted the first OpenAI Robotics Symposium on April 27, 2019....
Our second class of OpenAI Scholars has concluded, with all eight scholars producing an exciting final project showcased at Scholars Demo Da...
Our second class of OpenAI Fellows has wrapped up, with each Fellow going from a machine learning beginner to core OpenAI contributor in the...
We’ve created MuseNet, a deep neural network that can generate 4-minute musical compositions with 10 different instruments, and can combine ...
We’ve developed the Sparse Transformer, a deep neural network which sets new records at predicting what comes next in a sequence—whether tex...
OpenAI Five is the first AI to beat the world champions in an esports game, having won two back-to-back games versus the world champion Dota...
We’ll be holding our final live event for OpenAI Five at 11:30am PT on April 13....
We’ve made progress towards stable and scalable training of energy-based models (EBMs) resulting in better sample quality and generalization...
Our class of eight scholars (out of 550 applicants) brings together collective expertise in literature, philosophy, cell biology, statistics...
We’ve created OpenAI LP, a new “capped-profit” company that allows us to rapidly increase our investments in compute and talent while includ...
We’ve created activation atlases (in collaboration with Google researchers), a new technique for visualizing what interactions between neuro...
We’re releasing a Neural MMO, a massively multiagent game environment for reinforcement learning agents. Our platform supports a large, vari...