-
-
Learning Day
-
Microsoft invests in and partners with OpenAI to support us building beneficial AGI
-
Chine : une startup utilise l’IA pour retrouver votre animal de compagnie, mais pas seulement
-
Why responsible AI development needs cooperation on safety
-
Londres : la reconnaissance faciale a un taux d’échec de 81%
-
Une marque créée par une IA met en vente de vrais t-shirts
-
OpenAI Robotics Symposium 2019
-
OpenAI Scholars 2019: Final projects
-
OpenAI Fellows Fall 2018: Final projects
-
Transfer of adversarial robustness between perturbation types
-
MuseNet
-
Generative modeling with sparse transformers
-
OpenAI Five defeats Dota 2 world champions
-
OpenAI Five Finals
-
Implicit generation and generalization methods for energy-based models
-
OpenAI Scholars 2019: Meet our Scholars
-
OpenAI LP
-
Introducing Activation Atlases
-
Neural MMO: A massively multiagent game environment
-
Spinning Up in Deep RL: Workshop review
-
AI safety needs social scientists
-
Better language models and their implications
-
Better language models and their implications
-
Computational limitations in robust classification and win-win results
-
OpenAI Fellows Summer 2018: Final projects
-
How AI training scales
-
Quantifying generalization in reinforcement learning
-
Spinning Up in Deep RL
-
Learning concepts with energy functions
-
Plan online, learn offline: Efficient learning and exploration via model-based control
-
Reinforcement learning with prediction-based rewards
-
Learning complex goals with iterated amplification
-
OpenAI Scholars 2019: Applications open
-
OpenAI Fellows Winter 2019 & Interns Summer 2019
-
FFJORD: Free-form continuous dynamics for scalable reversible generative models
-
FFJORD: Free-form continuous dynamics for scalable reversible generative models
-
OpenAI Scholars 2018: Final projects
-
OpenAI Scholars 2018: Final projects
-
The International 2018: Results
-
Large-scale study of curiosity-driven learning
-
OpenAI Five Benchmark: Results
-
Learning dexterity
-
Variational option discovery algorithms
-
OpenAI Scholars 2018: Meet our Scholars
-
OpenAI Five Benchmark
-
Glow: Better reversible generative models
-
Learning Montezuma’s Revenge from a single demonstration
-
OpenAI Five
-
Retro Contest: Results
-
Learning policy representations in multiagent systems
-
Improving language understanding with unsupervised learning
-
GamePad: A learning environment for theorem proving
-
OpenAI Fellows Fall 2018
-
Gym Retro
-
AI and compute
-
AI safety via debate
-
Evolved Policy Gradients
-
Gotta Learn Fast: A new benchmark for generalization in RL
-
How to make AI that’s good for people
-
Retro Contest
-
Variance reduction for policy gradient with action-dependent factorized baselines
-
Report from the OpenAI hackathon
-
Improving GANs using optimal transport
-
On first-order meta-learning algorithms
-
Reptile: A scalable meta-learning algorithm
-
OpenAI Scholars
-
Some considerations on learning to explore via meta-reinforcement learning
-
Ingredients for robotics research
-
Multi-Goal Reinforcement Learning: Challenging robotics environments and request for research
-
OpenAI hackathon
-
Preparing for malicious uses of AI
-
OpenAI supporters
-
Interpretable machine learning through teaching
-
Interpretable machine learning through teaching
-
Discovering types for entity disambiguation
-
Requests for Research 2.0
-
Scaling Kubernetes to 2,500 nodes
-
Block-sparse GPU kernels
-
Learning sparse neural networks through L₀ regularization
-
Interpretable and pedagogical examples
-
Learning a hierarchy
-
Generalizing from simulation
-
Asymmetric actor critic for image-based robot learning
-
Sim-to-real transfer of robotic control with dynamics randomization
-
Domain randomization and generative models for robotic grasping
-
Competitive self-play
-
Meta-learning for wrestling
-
Nonlinear computation in deep linear networks
-
Learning to model other minds
-
Learning with opponent-learning awareness
-
Learning with opponent-learning awareness
-
OpenAI Baselines: ACKTR & A2C
-
More on Dota 2
-
Dota 2
-
Gathering human feedback
-
Better exploration with parameter noise
-
Proximal Policy Optimization
-
Robust adversarial inputs
-
Hindsight Experience Replay
-
Teacher–student curriculum learning