OpenAI licenses GPT-3 technology to Microsoft
OpenAI has agreed to license GPT-3 to Microsoft for their own products and services.
WeiterlesenOpenAI has agreed to license GPT-3 to Microsoft for their own products and services.
WeiterlesenWe’ve applied reinforcement learning from human feedback to train language models that are better at summarization.
WeiterlesenOur third class of OpenAI Scholars presented their final projects at virtual Demo Day, showcasing their research results from over the past five months.
WeiterlesenWe’re excited to announce that OpenAI is co-organizing two NeurIPS 2020 competitions with AIcrowd, Carnegie Mellon University, and DeepMind, using Procgen Benchmark and MineRL.
WeiterlesenWe find that, just as a large transformer model trained on language can generate coherent text, the same exact model trained on pixel sequences can generate...
WeiterlesenWe’re releasing an analysis showing that since 2012 the amount of compute needed to train a neural net to the same performance on ImageNet classification has been...
WeiterlesenWe’re introducing Jukebox, a neural net that generates music, including rudimentary singing, as raw audio in a variety of genres and artist styles. We’re releasing the...
WeiterlesenWe’ve contributed to a multi-stakeholder report by 58 co-authors at 30 organizations, including the Centre for the Future of Intelligence, Mila, Schwartz Reisman Institute for Technology and Society, Center for Advanced Study...
WeiterlesenWe’re introducing OpenAI Microscope, a collection of visualizations of every significant layer and neuron of eight vision “model organisms” which are often studied in interpretability. Microscope makes...
WeiterlesenWe are standardizing OpenAI’s deep learning framework on PyTorch.
WeiterlesenWe show that the double descent phenomenon occurs in CNNs, ResNets, and transformers: performance first improves, then gets worse, and then improves again with increasing model size, data size, or...
WeiterlesenWe’re releasing Procgen Benchmark, 16 simple-to-use procedurally-generated environments which provide a direct measure of how quickly a reinforcement learning agent learns generalizable skills.
WeiterlesenWe’re releasing Safety Gym, a suite of environments and tools for measuring progress towards reinforcement learning agents that respect safety constraints while training.
WeiterlesenAs the final model release of GPT-2’s staged release, we’re releasing the largest version (1.5B parameters) of GPT-2 along with code and model weights to facilitate detection of outputs of...
WeiterlesenWe’ve trained a pair of neural networks to solve the Rubik’s Cube with a human-like robot hand. The neural networks are trained entirely in simulation, using...
WeiterlesenWe are now accepting applications for our third class of OpenAI Scholars.
WeiterlesenWe’ve fine-tuned the 774M parameter GPT-2 language model using human feedback for various tasks, successfully matching the preferences of the external human labelers, though those preferences...
WeiterlesenWe’ve observed agents discovering progressively more complex tool use while playing a simple game of hide-and-seek. Through training in our new simulated hide-and-seek environment, agents build...
WeiterlesenWe’ve developed a method to assess whether a neural network classifier can reliably defend against adversarial attacks not seen during training. Our method yields a new...
WeiterlesenWe’re releasing the 774 million parameter GPT-2 language model after the release of our small 124M model in February, staged release of our medium 355M model in May, and subsequent...
WeiterlesenAt OpenAI, each Thursday is Learning Day: a day where employees have the option to self-study technical skills that will make them better at their job...
WeiterlesenMicrosoft is investing $1 billion in OpenAI to support us building artificial general intelligence (AGI) with widely distributed economic benefits. We’re partnering to develop a hardware...
WeiterlesenWe’ve written a policy research paper identifying four strategies that can be used today to improve the likelihood of long-term industry cooperation on safety norms in...
WeiterlesenWe hosted the first OpenAI Robotics Symposium on April 27, 2019.
WeiterlesenOur second class of OpenAI Scholars has concluded, with all eight scholars producing an exciting final project showcased at Scholars Demo Day at OpenAI.
WeiterlesenOur second class of OpenAI Fellows has wrapped up, with each Fellow going from a machine learning beginner to core OpenAI contributor in the course of...
WeiterlesenWe’ve created MuseNet, a deep neural network that can generate 4-minute musical compositions with 10 different instruments, and can combine styles from country to Mozart to...
WeiterlesenWe’ve developed the Sparse Transformer, a deep neural network which sets new records at predicting what comes next in a sequence—whether text, images, or sound. It...
WeiterlesenOpenAI Five is the first AI to beat the world champions in an esports game, having won two back-to-back games versus the world champion Dota 2...
WeiterlesenWe’ll be holding our final live event for OpenAI Five at 11:30am PT on April 13.
WeiterlesenWe’ve made progress towards stable and scalable training of energy-based models (EBMs) resulting in better sample quality and generalization ability than existing models. Generation in EBMs spends more...
WeiterlesenOur class of eight scholars (out of 550 applicants) brings together collective expertise in literature, philosophy, cell biology, statistics, economics, quantum physics, and business innovation.
WeiterlesenWe’ve created OpenAI LP, a new “capped-profit” company that allows us to rapidly increase our investments in compute and talent while including checks and balances to...
WeiterlesenWe’ve created activation atlases (in collaboration with Google researchers), a new technique for visualizing what interactions between neurons can represent. As AI systems are deployed in increasingly sensitive contexts, having...
WeiterlesenWe’re releasing a Neural MMO, a massively multiagent game environment for reinforcement learning agents. Our platform supports a large, variable number of agents within a persistent and...
WeiterlesenOn February 2, we held our first Spinning Up Workshop as part of our new education initiative at OpenAI.
WeiterlesenWe’ve written a paper arguing that long-term AI safety research needs social scientists to ensure AI alignment algorithms succeed when actual humans are involved. Properly aligning...
WeiterlesenWe’ve trained a large-scale unsupervised language model which generates coherent paragraphs of text, achieves state-of-the-art performance on many language modeling benchmarks, and performs rudimentary reading comprehension,...
WeiterlesenOur first cohort of OpenAI Fellows has concluded, with each Fellow going from a machine learning beginner to core OpenAI contributor in the course of a...
WeiterlesenWe’ve discovered that the gradient noise scale, a simple statistical metric, predicts the parallelizability of neural network training on a wide range of tasks. Since complex...
WeiterlesenWe’re releasing CoinRun, a training environment which provides a metric for an agent’s ability to transfer its experience to novel situations and has already helped clarify...
WeiterlesenWe’re releasing Spinning Up in Deep RL, an educational resource designed to let anyone learn to become a skilled practitioner in deep reinforcement learning. Spinning Up...
WeiterlesenWe’ve developed an energy-based model that can quickly learn to identify and generate instances of concepts, such as near, above, between, closest, and furthest, expressed as sets of...
WeiterlesenWe’ve developed Random Network Distillation (RND), a prediction-based method for encouraging reinforcement learning agents to explore their environments through curiosity, which for the first time exceeds average...
WeiterlesenWe’re proposing an AI safety technique called iterated amplification that lets us specify complicated behaviors and goals that are beyond human scale, by demonstrating how to...
WeiterlesenWe are now accepting applications for our second cohort of OpenAI Scholars, a program where we provide 6–10 stipends and mentorship to individuals from underrepresented groups...
WeiterlesenWe are now accepting applications for OpenAI Fellows and Interns for 2019.
WeiterlesenOur first cohort of OpenAI Scholars has now completed the program.
WeiterlesenOpenAI Five lost two games against top Dota 2 players at The International in Vancouver this week, maintaining a good chance of winning for the first...
WeiterlesenYesterday, OpenAI Five won a best-of-three against a team of 99.95th percentile Dota players: Blitz, Cap, Fogged, Merlini, and MoonMeander—four of whom have played Dota professionally—in front of a live audience and 100,000...
WeiterlesenWe’ve trained a human-like robot hand to manipulate physical objects with unprecedented dexterity.
WeiterlesenOur first class of OpenAI Scholars is underway, and you can now follow along as this group of experienced software developers becomes machine learning practitioners.
WeiterlesenWe introduce Glow, a reversible generative model which uses invertible 1×1 convolutions. It extends previous work on reversible generative models and simplifies the architecture. Our model can generate realistic high...
WeiterlesenWe’ve trained an agent to achieve a high score of 74,500 on Montezuma’s Revenge from a single human demonstration, better than any previously published result. Our algorithm is...
WeiterlesenOur team of five neural networks, OpenAI Five, has started to defeat amateur human teams at Dota 2.
WeiterlesenThe first run of our Retro Contest—exploring the development of algorithms that can generalize from previous experience—is now complete.
WeiterlesenWe’ve obtained state-of-the-art results on a suite of diverse language tasks with a scalable, task-agnostic system, which we’re also releasing. Our approach is a combination of...
WeiterlesenWe’re now accepting applications for the next cohort of OpenAI Fellows, a program which offers a compensated 6-month apprenticeship in AI research at OpenAI.
WeiterlesenWe’re releasing the full version of Gym Retro, a platform for reinforcement learning research on games. This brings our publicly-released game count from around 70 Atari games...
WeiterlesenWe’re releasing an analysis showing that since 2012, the amount of compute used in the largest AI training runs has been increasing exponentially with a 3.4-month...
WeiterlesenWe’re proposing an AI safety technique which trains agents to debate topics with one another, using a human to judge who wins.
WeiterlesenWe’re releasing an experimental metalearning approach called Evolved Policy Gradients, a method that evolves the loss function of learning agents, which can enable fast training on...
WeiterlesenWe’re launching a transfer learning contest that measures a reinforcement learning algorithm’s ability to generalize from previous experience.
WeiterlesenOn March 3rd, we hosted our first hackathon with 100 members of the artificial intelligence community.
WeiterlesenWe’ve developed a simple meta-learning algorithm called Reptile which works by repeatedly sampling a task, performing stochastic gradient descent on it, and updating the initial parameters...
WeiterlesenWe’re providing 6–10 stipends and mentorship to individuals from underrepresented groups to study deep learning full-time for 3 months and open-source a project.
WeiterlesenWe’re releasing eight simulated robotics environments and a Baselines implementation of Hindsight Experience Replay, all developed for our research over the past year. We’ve used these...
WeiterlesenCome to OpenAI’s office in San Francisco’s Mission District for talks and a hackathon on Saturday, March 3rd.
WeiterlesenWe’ve co-authored a paper that forecasts how malicious actors could misuse AI technology, and potential ways we can prevent and mitigate these threats. This paper is...
WeiterlesenWe’ve designed a method that encourages AIs to teach each other with examples that also make sense to humans. Our approach automatically selects the most informative...
WeiterlesenWe’ve built a system for automatically figuring out which object is meant by a word by having a neural network decide if the word belongs to...
WeiterlesenWe’re releasing a new batch of seven unsolved problems which have come up in the course of our research at OpenAI.
WeiterlesenWe’re releasing highly-optimized GPU kernels for an underexplored class of neural network architectures: networks with block-sparse weights. Depending on the chosen sparsity, these kernels can run...
WeiterlesenWe’ve developed a hierarchical reinforcement learning algorithm that learns high-level actions useful for solving a range of tasks, allowing fast solving of tasks requiring thousands of...
WeiterlesenOur latest robotics techniques allow robot controllers, trained entirely in simulation and deployed on physical robots, to react to unplanned changes in the environment as they...
WeiterlesenWe show that for the task of simulated robot wrestling, a meta-learning agent can learn to quickly defeat a stronger non-meta-learning agent, and also show that...
WeiterlesenWe’ve found that self-play allows simulated AIs to discover physical skills like tackling, ducking, faking, kicking, catching, and diving for the ball, without explicitly designing an...
WeiterlesenWe’re releasing an algorithm which accounts for the fact that other agents are learning too, and discovers self-interested yet collaborative strategies like tit-for-tat in the iterated...
WeiterlesenWe’re releasing two new OpenAI Baselines implementations: ACKTR and A2C. A2C is a synchronous, deterministic variant of Asynchronous Advantage Actor Critic (A3C) which we’ve found gives...
WeiterlesenOur Dota 2 result shows that self-play can catapult the performance of machine learning systems from far below human level to superhuman, given sufficient compute. In...
WeiterlesenWe’ve created a bot which beats the world’s top professionals at 1v1 matches of Dota 2 under standard tournament rules. The bot learned the game from...
WeiterlesenRL-Teacher is an open-source implementation of our interface to train AIs via occasional human feedback rather than hand-crafted reward functions. The underlying technique was developed as...
WeiterlesenWe’ve found that adding adaptive noise to the parameters of reinforcement learning algorithms frequently boosts performance. This exploration method is simple to implement and very rarely...
WeiterlesenWe’re releasing a new class of reinforcement learning algorithms, Proximal Policy Optimization (PPO), which perform comparably or better than state-of-the-art approaches while being much simpler to...
WeiterlesenWe’ve created images that reliably fool neural network classifiers when viewed from varied scales and perspectives. This challenges a claim from last week that self-driving cars...
WeiterlesenWe’re open-sourcing a high-performance Python library for robotic simulation using the MuJoCo engine, developed over our past year of robotics research.
WeiterlesenOne step towards building safe AI systems is to remove the need for humans to write goal functions, since using a simple proxy for a complex...
WeiterlesenMultiagent environments where agents compete for resources are stepping stones on the path to AGI. Multiagent environments have two useful properties: first, there is a natural...
WeiterlesenWe’re open-sourcing OpenAI Baselines, our internal effort to reproduce reinforcement learning algorithms with performance on par with published results. We’ll release the algorithms over upcoming months;...
WeiterlesenWe’ve created a robotics system, trained entirely in simulation and deployed on a physical robot, which can learn a new task after seeing it done once.
WeiterlesenWe are releasing Roboschool: open-source software for robot simulation, integrated with OpenAI Gym.
WeiterlesenWe’ve developed an unsupervised system which learns an excellent representation of sentiment, despite being trained only to predict the next character in the text of Amazon...
WeiterlesenWe’ve created the world’s first Spam-detecting AI trained entirely in simulation and deployed on a physical robot.
WeiterlesenWe’ve discovered that evolution strategies (ES), an optimization technique that’s been known for decades, rivals the performance of standard reinforcement learning (RL) techniques on modern RL...
WeiterlesenWe’re excited to support today’s launch of Distill, a new kind of journal aimed at excellent communication of machine learning results (novel or existing).
WeiterlesenIn this post we’ll outline new OpenAI research in which agents develop their own language.
WeiterlesenAdversarial examples are inputs to machine learning models that an attacker has intentionally designed to cause the model to make a mistake; they’re like optical illusions...
WeiterlesenThe OpenAI team is now 45 people. Together, we’re pushing the frontier of AI capabilities—whether by validating novel ideas, creating new software systems, or deploying machine...
WeiterlesenReinforcement learning algorithms can break in surprising, counterintuitive ways. In this post we’ll explore one failure mode, which is where you misspecify your reward function.
WeiterlesenWe’re releasing Universe, a software platform for measuring and training an AI’s general intelligence across the world’s supply of games, websites and other applications.
WeiterlesenWe’re working with Microsoft to start running most of our large-scale experiments on Azure.
WeiterlesenLast week we hosted over a hundred and fifty AI practitioners in our offices for our first self-organizing conference on machine learning.
WeiterlesenDeep learning is an empirical science, and the quality of a group’s infrastructure is a multiplier on progress. Fortunately, today’s open-source ecosystem makes it possible for...
WeiterlesenThe latest information about the Unconference is now available at the Unconference wiki, which will be periodically updated with more information for attendees.
WeiterlesenWe’ve hired more great people to help us achieve our goals. Welcome, everyone!
WeiterlesenImpactful scientific work requires working on the right problems—problems which are not just interesting, but whose solutions matter.
WeiterlesenWe (along with researchers from Berkeley and Stanford) are co-authors on today’s paper led by Google Brain researchers, Concrete Problems in AI Safety. The paper explores many...
WeiterlesenOpenAI’s mission is to build safe AI, and ensure AI’s benefits are as widely and evenly distributed as possible.
WeiterlesenThis post describes four projects that share a common theme of enhancing or using generative models, a branch of unsupervised learning techniques in machine learning. In...
WeiterlesenWe’d like to welcome the latest set of team members to OpenAI (and we’re still hiring!)
WeiterlesenWe’re releasing the public beta of OpenAI Gym, a toolkit for developing and comparing reinforcement learning (RL) algorithms. It consists of a growing suite of environments...
WeiterlesenWe’ve had some fantastic people join over the past few months (and we’re still hiring). Welcome, everyone!
WeiterlesenOpenAI is a non-profit artificial intelligence research company. Our goal is to advance digital intelligence in the way that is most likely to benefit humanity as...
Weiterlesen