Category
Research
-
-
Improving Model Safety Behavior with Rule-Based Rewards
-
GPT-4o mini: advancing cost-efficient intelligence
-
Prover-Verifier Games improve legibility of language model outputs
-
OpenAI and Los Alamos National Laboratory announce research partnership
-
A Holistic Approach to Undesired Content Detection in the Real World
-
A Holistic Approach to Undesired Content Detection in the Real World
-
Consistency Models
-
Improved Techniques for Training Consistency Models
-
Extracting Concepts from GPT-4
-
Hello GPT-4o
-
Understanding the source of what we see and hear online
-
The Instruction Hierarchy: Training LLMs to Prioritize Privileged Instructions
-
Video generation models as world simulators
-
Building an early warning system for LLM-aided biological threat creation
-
Improving mathematical reasoning with process supervision
-
Democratic inputs to AI
-
GPTs are GPTs: An early look at the labor market impact potential of large language models
-
GPT-4
-
Point-E: A system for generating 3D point clouds from complex prompts
-
Scaling laws for reward model overoptimization
-
Introducing Whisper
-
Efficient training of language models to fill in the middle
-
DALL·E 2 pre-training mitigations
-
Learning to play Minecraft with Video PreTraining
-
Evolution through large models
-
Techniques for training large neural networks
-
Teaching models to express their uncertainty in words
-
Hierarchical text-conditional image generation with CLIP latents
-
A research agenda for assessing the economic impacts of code generation models
-
Solving (some) formal math olympiad problems
-
Text and code embeddings by contrastive pre-training
-
WebGPT: Improving the factual accuracy of language models through web browsing
-
Solving math word problems
-
TruthfulQA: Measuring how models mimic human falsehoods
-
Introducing Triton: Open-source GPU programming for neural networks
-
Evaluating large language models trained on code
-
Multimodal neurons in artificial neural networks
-
Understanding the capabilities, limitations, and societal impact of large language models
-
Understanding the capabilities, limitations, and societal impact of large language models
-
Scaling Kubernetes to 7,500 nodes
-
CLIP: Connecting text and images
-
DALL·E: Creating images from text
-
DALL·E: Creating images from text
-
Generative language modeling for automated theorem proving
-
Image GPT
-
Language models are few-shot learners
-
AI and efficiency
-
Jukebox
-
Improving verifiability in AI development
-
Improving verifiability in AI development
-
OpenAI Microscope
-
Scaling laws for neural language models
-
Scaling laws for neural language models
-
Dota 2 with large scale deep reinforcement learning
-
Deep double descent
-
Procgen Benchmark
-
GPT-2: 1.5B release
-
Solving Rubik’s Cube with a robot hand
-
Emergent tool use from multi-agent interaction
-
GPT-2: 6-month follow-up
-
MuseNet
-
Generative modeling with sparse transformers
-
OpenAI Five defeats Dota 2 world champions
-
Implicit generation and generalization methods for energy-based models
-
Neural MMO: A massively multiagent game environment
-
Better language models and their implications
-
Better language models and their implications
-
Computational limitations in robust classification and win-win results
-
How AI training scales
-
Quantifying generalization in reinforcement learning
-
Spinning Up in Deep RL
-
Learning concepts with energy functions
-
Plan online, learn offline: Efficient learning and exploration via model-based control
-
Reinforcement learning with prediction-based rewards
-
FFJORD: Free-form continuous dynamics for scalable reversible generative models
-
FFJORD: Free-form continuous dynamics for scalable reversible generative models
-
The International 2018: Results
-
Large-scale study of curiosity-driven learning
-
OpenAI Five Benchmark: Results
-
Learning dexterity
-
Variational option discovery algorithms
-
Glow: Better reversible generative models
-
Learning Montezuma’s Revenge from a single demonstration
-
OpenAI Five
-
Retro Contest: Results
-
Learning policy representations in multiagent systems
-
GamePad: A learning environment for theorem proving
-
Gym Retro
-
AI and compute
-
Evolved Policy Gradients
-
Gotta Learn Fast: A new benchmark for generalization in RL
-
Retro Contest
-
Variance reduction for policy gradient with action-dependent factorized baselines
-
Improving GANs using optimal transport
-
On first-order meta-learning algorithms
-
Reptile: A scalable meta-learning algorithm
-
Some considerations on learning to explore via meta-reinforcement learning
-
Multi-Goal Reinforcement Learning: Challenging robotics environments and request for research
-
Ingredients for robotics research
-
Interpretable machine learning through teaching