General Loading...

Procgen and MineRL Competitions

We’re excited to announce that OpenAI is co-organizing two NeurIPS 2020 competitions with AIcrowd, Carnegie Mellon University, and DeepMind, using Procgen Benchmark and MineRL.

Weiterlesen
General Loading...

Image GPT

We find that, just as a large transformer model trained on language can generate coherent text, the same exact model trained on pixel sequences can generate...

Weiterlesen
General Loading...

OpenAI API

We’re releasing an API for accessing new AI models developed by OpenAI.

Weiterlesen
General Loading...

AI and efficiency

We’re releasing an analysis showing that since 2012 the amount of compute needed to train a neural net to the same performance on ImageNet classification has been...

Weiterlesen
General Loading...

Jukebox

We’re introducing Jukebox, a neural net that generates music, including rudimentary singing, as raw audio in a variety of genres and artist styles. We’re releasing the...

Weiterlesen
General Loading...

Improving verifiability in AI development

We’ve contributed to a multi-stakeholder report by 58 co-authors at 30 organizations, including the Centre for the Future of Intelligence, Mila, Schwartz Reisman Institute for Technology and Society, Center for Advanced Study...

Weiterlesen
General Loading...

OpenAI Microscope

We’re introducing OpenAI Microscope, a collection of visualizations of every significant layer and neuron of eight vision “model organisms” which are often studied in interpretability. Microscope makes...

Weiterlesen
General Loading...

Deep double descent

We show that the double descent phenomenon occurs in CNNs, ResNets, and transformers: performance first improves, then gets worse, and then improves again with increasing model size, data size, or...

Weiterlesen
General Loading...

Procgen Benchmark

We’re releasing Procgen Benchmark, 16 simple-to-use procedurally-generated environments which provide a direct measure of how quickly a reinforcement learning agent learns generalizable skills.

Weiterlesen
General Loading...

Safety Gym

We’re releasing Safety Gym, a suite of environments and tools for measuring progress towards reinforcement learning agents that respect safety constraints while training.

Weiterlesen
General Loading...

GPT-2: 1.5B release

As the final model release of GPT-2’s staged release, we’re releasing the largest version (1.5B parameters) of GPT-2 along with code and model weights to facilitate detection of outputs of...

Weiterlesen
General Loading...

Fine-tuning GPT-2 from human preferences

We’ve fine-tuned the 774M parameter GPT-2 language model using human feedback for various tasks, successfully matching the preferences of the external human labelers, though those preferences...

Weiterlesen
General Loading...

Emergent tool use from multi-agent interaction

We’ve observed agents discovering progressively more complex tool use while playing a simple game of hide-and-seek. Through training in our new simulated hide-and-seek environment, agents build...

Weiterlesen
General Loading...

GPT-2: 6-month follow-up

We’re releasing the 774 million parameter GPT-2 language model after the release of our small 124M model in February, staged release of our medium 355M model in May, and subsequent...

Weiterlesen
General Loading...

Learning Day

At OpenAI, each Thursday is Learning Day: a day where employees have the option to self-study technical skills that will make them better at their job...

Weiterlesen
General Loading...

MuseNet

We’ve created MuseNet, a deep neural network that can generate 4-minute musical compositions with 10 different instruments, and can combine styles from country to Mozart to...

Weiterlesen
General Loading...

OpenAI Five Finals

We’ll be holding our final live event for OpenAI Five at 11:30am PT on April 13.

Weiterlesen
General Loading...

OpenAI Scholars 2019: Meet our Scholars

Our class of eight scholars (out of 550 applicants) brings together collective expertise in literature, philosophy, cell biology, statistics, economics, quantum physics, and business innovation.

Weiterlesen
General Loading...

OpenAI LP

We’ve created OpenAI LP, a new “capped-profit” company that allows us to rapidly increase our investments in compute and talent while including checks and balances to...

Weiterlesen
General Loading...

Introducing Activation Atlases

We’ve created activation atlases (in collaboration with Google researchers), a new technique for visualizing what interactions between neurons can represent. As AI systems are deployed in increasingly sensitive contexts, having...

Weiterlesen
General Loading...

AI safety needs social scientists

We’ve written a paper arguing that long-term AI safety research needs social scientists to ensure AI alignment algorithms succeed when actual humans are involved. Properly aligning...

Weiterlesen
General Loading...

Better language models and their implications

We’ve trained a large-scale unsupervised language model which generates coherent paragraphs of text, achieves state-of-the-art performance on many language modeling benchmarks, and performs rudimentary reading comprehension,...

Weiterlesen
General Loading...

How AI training scales

We’ve discovered that the gradient noise scale, a simple statistical metric, predicts the parallelizability of neural network training on a wide range of tasks. Since complex...

Weiterlesen
General Loading...

Spinning Up in Deep RL

We’re releasing Spinning Up in Deep RL, an educational resource designed to let anyone learn to become a skilled practitioner in deep reinforcement learning. Spinning Up...

Weiterlesen
General Loading...

Learning concepts with energy functions

We’ve developed an energy-based model that can quickly learn to identify and generate instances of concepts, such as near, above, between, closest, and furthest, expressed as sets of...

Weiterlesen
General Loading...

Reinforcement learning with prediction-based rewards

We’ve developed Random Network Distillation (RND), a prediction-based method for encouraging reinforcement learning agents to explore their environments through curiosity, which for the first time exceeds average...

Weiterlesen
General Loading...

OpenAI Scholars 2019: Applications open

We are now accepting applications for our second cohort of OpenAI Scholars, a program where we provide 6–10 stipends and mentorship to individuals from underrepresented groups...

Weiterlesen
General Loading...

The International 2018: Results

OpenAI Five lost two games against top Dota 2 players at The International in Vancouver this week, maintaining a good chance of winning for the first...

Weiterlesen
General Loading...

OpenAI Five Benchmark: Results

Yesterday, OpenAI Five won a best-of-three against a team of 99.95th percentile Dota players: Blitz, Cap, Fogged, Merlini, and MoonMeander—four of whom have played Dota professionally—in front of a live audience and 100,000...

Weiterlesen
General Loading...

Learning dexterity

We’ve trained a human-like robot hand to manipulate physical objects with unprecedented dexterity.

Weiterlesen
General Loading...

OpenAI Scholars 2018: Meet our Scholars

Our first class of OpenAI Scholars is underway, and you can now follow along as this group of experienced software developers becomes machine learning practitioners.

Weiterlesen
General Loading...

Glow: Better reversible generative models

We introduce Glow, a reversible generative model which uses invertible 1×1 convolutions. It extends previous work on reversible generative models and simplifies the architecture. Our model can generate realistic high...

Weiterlesen
General Loading...

OpenAI Five

Our team of five neural networks, OpenAI Five, has started to defeat amateur human teams at Dota 2.

Weiterlesen
General Loading...

Retro Contest: Results

The first run of our Retro Contest—exploring the development of algorithms that can generalize from previous experience—is now complete.

Weiterlesen
General Loading...

OpenAI Fellows Fall 2018

We’re now accepting applications for the next cohort of OpenAI Fellows, a program which offers a compensated 6-month apprenticeship in AI research at OpenAI.

Weiterlesen
General Loading...

Gym Retro

We’re releasing the full version of Gym Retro, a platform for reinforcement learning research on games. This brings our publicly-released game count from around 70 Atari games...

Weiterlesen
General Loading...

AI and compute

We’re releasing an analysis showing that since 2012, the amount of compute used in the largest AI training runs has been increasing exponentially with a 3.4-month...

Weiterlesen
General Loading...

AI safety via debate

We’re proposing an AI safety technique which trains agents to debate topics with one another, using a human to judge who wins.

Weiterlesen
General Loading...

Evolved Policy Gradients

We’re releasing an experimental metalearning approach called Evolved Policy Gradients, a method that evolves the loss function of learning agents, which can enable fast training on...

Weiterlesen
General Loading...

Retro Contest

We’re launching a transfer learning contest that measures a reinforcement learning algorithm’s ability to generalize from previous experience.

Weiterlesen
General Loading...

Reptile: A scalable meta-learning algorithm

We’ve developed a simple meta-learning algorithm called Reptile which works by repeatedly sampling a task, performing stochastic gradient descent on it, and updating the initial parameters...

Weiterlesen
General Loading...

OpenAI Scholars

We’re providing 6–10 stipends and mentorship to individuals from underrepresented groups to study deep learning full-time for 3 months and open-source a project.

Weiterlesen
General Loading...

Ingredients for robotics research

We’re releasing eight simulated robotics environments and a Baselines implementation of Hindsight Experience Replay, all developed for our research over the past year. We’ve used these...

Weiterlesen
General Loading...

OpenAI hackathon

Come to OpenAI’s office in San Francisco’s Mission District for talks and a hackathon on Saturday, March 3rd.

Weiterlesen
General Loading...

Preparing for malicious uses of AI

We’ve co-authored a paper that forecasts how malicious actors could misuse AI technology, and potential ways we can prevent and mitigate these threats. This paper is...

Weiterlesen
General Loading...

Requests for Research 2.0

We’re releasing a new batch of seven unsolved problems which have come up in the course of our research at OpenAI.

Weiterlesen
General Loading...

Block-sparse GPU kernels

We’re releasing highly-optimized GPU kernels for an underexplored class of neural network architectures: networks with block-sparse weights. Depending on the chosen sparsity, these kernels can run...

Weiterlesen
General Loading...

Learning a hierarchy

We’ve developed a hierarchical reinforcement learning algorithm that learns high-level actions useful for solving a range of tasks, allowing fast solving of tasks requiring thousands of...

Weiterlesen
General Loading...

Generalizing from simulation

Our latest robotics techniques allow robot controllers, trained entirely in simulation and deployed on physical robots, to react to unplanned changes in the environment as they...

Weiterlesen
General Loading...

Meta-learning for wrestling

We show that for the task of simulated robot wrestling, a meta-learning agent can learn to quickly defeat a stronger non-meta-learning agent, and also show that...

Weiterlesen
General Loading...

Competitive self-play

We’ve found that self-play allows simulated AIs to discover physical skills like tackling, ducking, faking, kicking, catching, and diving for the ball, without explicitly designing an...

Weiterlesen
General Loading...

Learning to model other minds

We’re releasing an algorithm which accounts for the fact that other agents are learning too, and discovers self-interested yet collaborative strategies like tit-for-tat in the iterated...

Weiterlesen
General Loading...

OpenAI Baselines: ACKTR & A2C

We’re releasing two new OpenAI Baselines implementations: ACKTR and A2C. A2C is a synchronous, deterministic variant of Asynchronous Advantage Actor Critic (A3C) which we’ve found gives...

Weiterlesen
General Loading...

More on Dota 2

Our Dota 2 result shows that self-play can catapult the performance of machine learning systems from far below human level to superhuman, given sufficient compute. In...

Weiterlesen
General Loading...

Dota 2

We’ve created a bot which beats the world’s top professionals at 1v1 matches of Dota 2 under standard tournament rules. The bot learned the game from...

Weiterlesen
General Loading...

Gathering human feedback

RL-Teacher is an open-source implementation of our interface to train AIs via occasional human feedback rather than hand-crafted reward functions. The underlying technique was developed as...

Weiterlesen
General Loading...

Better exploration with parameter noise

We’ve found that adding adaptive noise to the parameters of reinforcement learning algorithms frequently boosts performance. This exploration method is simple to implement and very rarely...

Weiterlesen
General Loading...

Proximal Policy Optimization

We’re releasing a new class of reinforcement learning algorithms, Proximal Policy Optimization (PPO), which perform comparably or better than state-of-the-art approaches while being much simpler to...

Weiterlesen
General Loading...

Robust adversarial inputs

We’ve created images that reliably fool neural network classifiers when viewed from varied scales and perspectives. This challenges a claim from last week that self-driving cars...

Weiterlesen
General Loading...

Faster physics in Python

We’re open-sourcing a high-performance Python library for robotic simulation using the MuJoCo engine, developed over our past year of robotics research.

Weiterlesen
General Loading...

Learning from human preferences

One step towards building safe AI systems is to remove the need for humans to write goal functions, since using a simple proxy for a complex...

Weiterlesen
General Loading...

OpenAI Baselines: DQN

We’re open-sourcing OpenAI Baselines, our internal effort to reproduce reinforcement learning algorithms with performance on par with published results. We’ll release the algorithms over upcoming months;...

Weiterlesen
General Loading...

Robots that learn

We’ve created a robotics system, trained entirely in simulation and deployed on a physical robot, which can learn a new task after seeing it done once.

Weiterlesen
General Loading...

Roboschool

We are releasing Roboschool: open-source software for robot simulation, integrated with OpenAI Gym.

Weiterlesen
General Loading...

Unsupervised sentiment neuron

We’ve developed an unsupervised system which learns an excellent representation of sentiment, despite being trained only to predict the next character in the text of Amazon...

Weiterlesen
General Loading...

Distill

We’re excited to support today’s launch of Distill, a new kind of journal aimed at excellent communication of machine learning results (novel or existing).

Weiterlesen
General Loading...

Team update

The OpenAI team is now 45 people. Together, we’re pushing the frontier of AI capabilities—whether by validating novel ideas, creating new software systems, or deploying machine...

Weiterlesen
General Loading...

Faulty reward functions in the wild

Reinforcement learning algorithms can break in surprising, counterintuitive ways. In this post we’ll explore one failure mode, which is where you misspecify your reward function.

Weiterlesen
General Loading...

Universe

We’re releasing Universe, a software platform for measuring and training an AI’s general intelligence across the world’s supply of games, websites and other applications.

Weiterlesen
General Loading...

OpenAI and Microsoft

We’re working with Microsoft to start running most of our large-scale experiments on Azure.

Weiterlesen
General Loading...

Infrastructure for deep learning

Deep learning is an empirical science, and the quality of a group’s infrastructure is a multiplier on progress. Fortunately, today’s open-source ecosystem makes it possible for...

Weiterlesen
General Loading...

Machine Learning Unconference

The latest information about the Unconference is now available at the Unconference wiki, which will be periodically updated with more information for attendees.

Weiterlesen
General Loading...

Team update

We’ve hired more great people to help us achieve our goals. Welcome, everyone!

Weiterlesen
General Loading...

Special projects

Impactful scientific work requires working on the right problems—problems which are not just interesting, but whose solutions matter.

Weiterlesen
General Loading...

Concrete AI safety problems

We (along with researchers from Berkeley and Stanford) are co-authors on today’s paper led by Google Brain researchers, Concrete Problems in AI Safety. The paper explores many...

Weiterlesen
General Loading...

OpenAI technical goals

OpenAI’s mission is to build safe AI, and ensure AI’s benefits are as widely and evenly distributed as possible.

Weiterlesen
General Loading...

Generative models

This post describes four projects that share a common theme of enhancing or using generative models, a branch of unsupervised learning techniques in machine learning. In...

Weiterlesen
General Loading...

Team update

We’d like to welcome the latest set of team members to OpenAI (and we’re still hiring!)

Weiterlesen
General Loading...

OpenAI Gym Beta

We’re releasing the public beta of OpenAI Gym, a toolkit for developing and comparing reinforcement learning (RL) algorithms. It consists of a growing suite of environments...

Weiterlesen
General Loading...

Team++

We’ve had some fantastic people join over the past few months (and we’re still hiring). Welcome, everyone!

Weiterlesen
General Loading...

Introducing OpenAI

OpenAI is a non-profit artificial intelligence research company. Our goal is to advance digital intelligence in the way that is most likely to benefit humanity as...

Weiterlesen