Category
Research
-
-
AI system predicts protein fragments that can bind to or inhibit a target
-
We’re launching a new AI system for scientists.
-
Like human brains, large language models reason about diverse data in a general way
-
AI model deciphers the code in proteins that tells them where to go
-
Physicists Tap James Web Space Telescope to Track New Asteroids and City-Killer Rock
-
Can deep learning transform heart failure prevention?
-
Introducing deep research
-
OpenAI o3-mini
-
OpenAI o3-mini System Card
-
Computer-Using Agent
-
Trading inference-time compute for adversarial robustness
-
OpenAI o1 System Card
-
Advancing red teaming with people and AI
-
Introducing SimpleQA
-
Simplifying, stabilizing, and scaling continuous-time consistency models
-
Evaluating fairness in ChatGPT
-
MLE-bench: Evaluating Machine Learning Agents on Machine Learning Engineering
-
Learning to reason with LLMs
-
OpenAI o1-mini
-
OpenAI o1 System Card External Testers Acknowledgements
-
Introducing SWE-bench Verified
-
GPT-4o System Card External Testers Acknowledgements
-
Improving Model Safety Behavior with Rule-Based Rewards
-
GPT-4o mini: advancing cost-efficient intelligence
-
Prover-Verifier Games improve legibility of language model outputs
-
OpenAI and Los Alamos National Laboratory announce research partnership
-
A Holistic Approach to Undesired Content Detection in the Real World
-
A Holistic Approach to Undesired Content Detection in the Real World
-
Consistency Models
-
Improved Techniques for Training Consistency Models
-
Extracting Concepts from GPT-4
-
Hello GPT-4o
-
Understanding the source of what we see and hear online
-
The Instruction Hierarchy: Training LLMs to Prioritize Privileged Instructions
-
Video generation models as world simulators
-
Building an early warning system for LLM-aided biological threat creation
-
Improving mathematical reasoning with process supervision
-
Democratic inputs to AI
-
GPTs are GPTs: An early look at the labor market impact potential of large language models
-
GPT-4
-
Point-E: A system for generating 3D point clouds from complex prompts
-
Scaling laws for reward model overoptimization
-
Introducing Whisper
-
Efficient training of language models to fill in the middle
-
DALL·E 2 pre-training mitigations
-
Learning to play Minecraft with Video PreTraining
-
Evolution through large models
-
Techniques for training large neural networks
-
Teaching models to express their uncertainty in words
-
Hierarchical text-conditional image generation with CLIP latents
-
A research agenda for assessing the economic impacts of code generation models
-
Solving (some) formal math olympiad problems
-
Text and code embeddings by contrastive pre-training
-
WebGPT: Improving the factual accuracy of language models through web browsing
-
Solving math word problems
-
TruthfulQA: Measuring how models mimic human falsehoods
-
Introducing Triton: Open-source GPU programming for neural networks
-
Evaluating large language models trained on code
-
Multimodal neurons in artificial neural networks
-
Understanding the capabilities, limitations, and societal impact of large language models
-
Understanding the capabilities, limitations, and societal impact of large language models
-
Scaling Kubernetes to 7,500 nodes
-
CLIP: Connecting text and images
-
DALL·E: Creating images from text
-
DALL·E: Creating images from text
-
Generative language modeling for automated theorem proving
-
Image GPT
-
Language models are few-shot learners
-
AI and efficiency
-
Jukebox
-
Improving verifiability in AI development
-
Improving verifiability in AI development
-
OpenAI Microscope
-
Scaling laws for neural language models
-
Scaling laws for neural language models
-
Dota 2 with large scale deep reinforcement learning
-
Deep double descent
-
Procgen Benchmark
-
GPT-2: 1.5B release
-
Solving Rubik’s Cube with a robot hand
-
Emergent tool use from multi-agent interaction
-
GPT-2: 6-month follow-up
-
MuseNet
-
Generative modeling with sparse transformers
-
OpenAI Five defeats Dota 2 world champions
-
Implicit generation and generalization methods for energy-based models
-
Neural MMO: A massively multiagent game environment
-
Better language models and their implications
-
Better language models and their implications
-
Computational limitations in robust classification and win-win results
-
How AI training scales
-
Quantifying generalization in reinforcement learning
-
Spinning Up in Deep RL
-
Learning concepts with energy functions
-
Plan online, learn offline: Efficient learning and exploration via model-based control
-
Reinforcement learning with prediction-based rewards
-
FFJORD: Free-form continuous dynamics for scalable reversible generative models
-
FFJORD: Free-form continuous dynamics for scalable reversible generative models