Category
Openai IA
Toute l’actualité IA d’Openai
-
-
OpenAI Scholars 2019: Final projects
-
OpenAI Fellows Fall 2018: Final projects
-
Transfer of adversarial robustness between perturbation types
-
MuseNet
-
Generative modeling with sparse transformers
-
OpenAI Five defeats Dota 2 world champions
-
OpenAI Five Finals
-
Implicit generation and generalization methods for energy-based models
-
OpenAI Scholars 2019: Meet our Scholars
-
OpenAI LP
-
Introducing Activation Atlases
-
Neural MMO: A massively multiagent game environment
-
Spinning Up in Deep RL: Workshop review
-
AI safety needs social scientists
-
Better language models and their implications
-
Better language models and their implications
-
Computational limitations in robust classification and win-win results
-
OpenAI Fellows Summer 2018: Final projects
-
How AI training scales
-
Quantifying generalization in reinforcement learning
-
Spinning Up in Deep RL
-
Learning concepts with energy functions
-
Plan online, learn offline: Efficient learning and exploration via model-based control
-
Reinforcement learning with prediction-based rewards
-
Learning complex goals with iterated amplification
-
OpenAI Scholars 2019: Applications open
-
OpenAI Fellows Winter 2019 & Interns Summer 2019
-
FFJORD: Free-form continuous dynamics for scalable reversible generative models
-
FFJORD: Free-form continuous dynamics for scalable reversible generative models
-
OpenAI Scholars 2018: Final projects
-
OpenAI Scholars 2018: Final projects
-
The International 2018: Results
-
Large-scale study of curiosity-driven learning
-
OpenAI Five Benchmark: Results
-
Learning dexterity
-
Variational option discovery algorithms
-
OpenAI Scholars 2018: Meet our Scholars
-
OpenAI Five Benchmark
-
Glow: Better reversible generative models
-
Learning Montezuma’s Revenge from a single demonstration
-
OpenAI Five
-
Retro Contest: Results
-
Learning policy representations in multiagent systems
-
Improving language understanding with unsupervised learning
-
GamePad: A learning environment for theorem proving
-
OpenAI Fellows Fall 2018
-
Gym Retro
-
AI and compute
-
AI safety via debate
-
Evolved Policy Gradients
-
Gotta Learn Fast: A new benchmark for generalization in RL
-
Retro Contest
-
Variance reduction for policy gradient with action-dependent factorized baselines
-
Improving GANs using optimal transport
-
Report from the OpenAI hackathon
-
On first-order meta-learning algorithms
-
Reptile: A scalable meta-learning algorithm
-
OpenAI Scholars
-
Some considerations on learning to explore via meta-reinforcement learning
-
Multi-Goal Reinforcement Learning: Challenging robotics environments and request for research
-
Ingredients for robotics research
-
OpenAI hackathon
-
OpenAI supporters
-
Preparing for malicious uses of AI
-
Interpretable machine learning through teaching
-
Interpretable machine learning through teaching
-
Discovering types for entity disambiguation
-
Requests for Research 2.0
-
Scaling Kubernetes to 2,500 nodes
-
Block-sparse GPU kernels
-
Learning sparse neural networks through L₀ regularization
-
Interpretable and pedagogical examples
-
Learning a hierarchy
-
Generalizing from simulation
-
Sim-to-real transfer of robotic control with dynamics randomization
-
Asymmetric actor critic for image-based robot learning
-
Domain randomization and generative models for robotic grasping
-
Meta-learning for wrestling
-
Competitive self-play
-
Nonlinear computation in deep linear networks
-
Learning to model other minds
-
Learning with opponent-learning awareness
-
Learning with opponent-learning awareness
-
OpenAI Baselines: ACKTR & A2C
-
More on Dota 2
-
Dota 2
-
Gathering human feedback
-
Better exploration with parameter noise
-
Proximal Policy Optimization
-
Robust adversarial inputs
-
Hindsight Experience Replay
-
Teacher–student curriculum learning
-
Faster physics in Python
-
Learning from human preferences
-
Learning to cooperate, compete, and communicate
-
UCB exploration via Q-ensembles
-
OpenAI Baselines: DQN
-
Robots that learn
-
Roboschool
-
Equivalence between policy gradients and soft Q-learning