We’re proposing an AI safety technique which trains agents to debate topics with one another, using a human to judge who wins.
We’re releasing an experimental metalearning approach called Evolved Policy Gradients, a method that evolves the loss function of learnin...
We’re launching a transfer learning contest that measures a reinforcement learning algorithm’s ability to generalize from previous experi...
On March 3rd, we hosted our first hackathon with 100 members of the artificial intelligence community.
We’ve developed a simple meta-learning algorithm called Reptile which works by repeatedly sampling a task, performing stochastic gradient...
We’re providing 6–10 stipends and mentorship to individuals from underrepresented groups to study deep learning full-time for 3 months an...
We’re releasing eight simulated robotics environments and a Baselines implementation of Hindsight Experience Replay, all developed for ou...
Come to OpenAI’s office in San Francisco’s Mission District for talks and a hackathon on Saturday, March 3rd.
We’ve co-authored a paper that forecasts how malicious actors could misuse AI technology, and potential ways we can prevent and mitigate ...
We’re excited to welcome new donors to OpenAI.
We’ve designed a method that encourages AIs to teach each other with examples that also make sense to humans. Our approach automatically ...
We’ve built a system for automatically figuring out which object is meant by a word by having a neural network decide if the word belongs...
We’re releasing a new batch of seven unsolved problems which have come up in the course of our research at OpenAI.
We’re releasing highly-optimized GPU kernels for an underexplored class of neural network architectures: networks with block-sparse weigh...
We’ve developed a hierarchical reinforcement learning algorithm that learns high-level actions useful for solving a range of tasks, allow...