Category
Research
-
-
Efficient training of language models to fill in the middle
-
DALL·E 2 pre-training mitigations
-
Learning to play Minecraft with Video PreTraining
-
Evolution through large models
-
Techniques for training large neural networks
-
Teaching models to express their uncertainty in words
-
Hierarchical text-conditional image generation with CLIP latents
-
A research agenda for assessing the economic impacts of code generation models
-
Solving (some) formal math olympiad problems
-
Text and code embeddings by contrastive pre-training
-
WebGPT: Improving the factual accuracy of language models through web browsing
-
Solving math word problems
-
TruthfulQA: Measuring how models mimic human falsehoods
-
Introducing Triton: Open-source GPU programming for neural networks
-
Evaluating large language models trained on code
-
Multimodal neurons in artificial neural networks
-
Understanding the capabilities, limitations, and societal impact of large language models
-
Understanding the capabilities, limitations, and societal impact of large language models
-
Scaling Kubernetes to 7,500 nodes
-
CLIP: Connecting text and images
-
DALL·E: Creating images from text
-
DALL·E: Creating images from text
-
Generative language modeling for automated theorem proving
-
Image GPT
-
Language models are few-shot learners
-
AI and efficiency
-
Jukebox
-
Improving verifiability in AI development
-
Improving verifiability in AI development
-
OpenAI Microscope
-
Scaling laws for neural language models
-
Scaling laws for neural language models
-
Dota 2 with large scale deep reinforcement learning
-
Deep double descent
-
Procgen Benchmark
-
GPT-2: 1.5B release
-
Solving Rubik’s Cube with a robot hand
-
Emergent tool use from multi-agent interaction
-
GPT-2: 6-month follow-up
-
MuseNet
-
Generative modeling with sparse transformers
-
OpenAI Five defeats Dota 2 world champions
-
Implicit generation and generalization methods for energy-based models
-
Neural MMO: A massively multiagent game environment
-
Better language models and their implications
-
Better language models and their implications
-
Computational limitations in robust classification and win-win results
-
How AI training scales
-
Quantifying generalization in reinforcement learning
-
Spinning Up in Deep RL
-
Learning concepts with energy functions
-
Plan online, learn offline: Efficient learning and exploration via model-based control
-
Reinforcement learning with prediction-based rewards
-
FFJORD: Free-form continuous dynamics for scalable reversible generative models
-
FFJORD: Free-form continuous dynamics for scalable reversible generative models
-
The International 2018: Results
-
Large-scale study of curiosity-driven learning
-
OpenAI Five Benchmark: Results
-
Learning dexterity
-
Variational option discovery algorithms
-
Glow: Better reversible generative models
-
Learning Montezuma’s Revenge from a single demonstration
-
OpenAI Five
-
Retro Contest: Results
-
Learning policy representations in multiagent systems
-
GamePad: A learning environment for theorem proving
-
Gym Retro
-
AI and compute
-
Evolved Policy Gradients
-
Gotta Learn Fast: A new benchmark for generalization in RL
-
Retro Contest
-
Variance reduction for policy gradient with action-dependent factorized baselines
-
Improving GANs using optimal transport
-
On first-order meta-learning algorithms
-
Reptile: A scalable meta-learning algorithm
-
Some considerations on learning to explore via meta-reinforcement learning
-
Multi-Goal Reinforcement Learning: Challenging robotics environments and request for research
-
Ingredients for robotics research
-
Interpretable machine learning through teaching
-
Interpretable machine learning through teaching
-
Discovering types for entity disambiguation
-
Requests for Research 2.0
-
Scaling Kubernetes to 2,500 nodes
-
Block-sparse GPU kernels
-
Learning sparse neural networks through L₀ regularization
-
Interpretable and pedagogical examples
-
Learning a hierarchy
-
Generalizing from simulation
-
Sim-to-real transfer of robotic control with dynamics randomization
-
Asymmetric actor critic for image-based robot learning
-
Domain randomization and generative models for robotic grasping
-
Meta-learning for wrestling
-
Competitive self-play
-
Nonlinear computation in deep linear networks
-
Learning to model other minds
-
Learning with opponent-learning awareness
-
Learning with opponent-learning awareness
-
OpenAI Baselines: ACKTR & A2C
-
More on Dota 2
-
Dota 2