Products and applications of OpenAI
The American artificial intelligence organization OpenAI has released a variety of products and applications since its founding in 2013.
Reinforcement learning
At its beginning, OpenAI's research included many projects focused on reinforcement learning. OpenAI has been viewed as an important competitor to DeepMind.Gym
Announced in 2016, Gym was an open-source Python library designed to facilitate the development of reinforcement learning algorithms. It aimed to standardize how environments are defined in AI research, making published research more easily reproducible while providing users with a simple interface for interacting with these environments. In 2022, new developments of Gym have been moved to the library Gymnasium.Gym Retro
Released in 2018, Gym Retro is a platform for reinforcement learning research on video games using RL algorithms and study generalization. Prior RL research focused mainly on optimizing agents to solve single tasks. Gym Retro gives the ability to generalize between games with similar concepts but different appearances.RoboSumo
Released in 2017, RoboSumo is a virtual world where humanoid metalearning robot agents initially lack knowledge of how to even walk, but are given the goals of learning to move and to push the opposing agent out of the ring. Through this adversarial learning process, the agents learn how to adapt to changing conditions. When an agent is then removed from this virtual environment and placed in a new virtual environment with high winds, the agent braces to remain upright, suggesting it had learned how to balance in a generalized way. OpenAI's Igor Mordatch argued that competition between agents could create an intelligence "arms race" that could increase an agent's ability to function even outside the context of the competition.OpenAI Five
OpenAI Five is a team of five OpenAI-curated bots used in the competitive five-on-five video game Dota 2, that learn to play against human players at a high skill level entirely through trial-and-error algorithms. Before becoming a team of five, the first public demonstration occurred at The International 2017, the annual premiere championship tournament for the game, where Dendi, a professional Ukrainian player, lost against a bot in a live one-on-one matchup. After the match, CTO Greg Brockman explained that the bot had learned by playing against itself for two weeks of real time, and that the learning software was a step in the direction of creating software that can handle complex tasks like a surgeon. The system uses a form of reinforcement learning, as the bots learn over time by playing against themselves hundreds of times a day for months, and are rewarded for actions such as killing an enemy and taking map objectives.By June 2018, the ability of the bots expanded to play together as a full team of five, and they were able to defeat teams of amateur and semi-professional players. At The International 2018, OpenAI Five played in two exhibition matches against professional players, but ended up losing both games. In April 2019, OpenAI Five defeated OG, the reigning world champions of the game at the time, 2:0 in a live exhibition match in San Francisco. The bots' final public appearance came later that month, where they played in 42,729 total games in a four-day open online competition, winning 99.4% of those games.
OpenAI Five's mechanisms in Dota 2's bot player show the challenges of AI systems in multiplayer online battle arena games and how OpenAI Five has demonstrated the use of deep reinforcement learning agents to achieve superhuman competence in Dota 2 matches.
Dactyl
Developed in 2018, Dactyl uses machine learning to train a Shadow Hand, a human-like robot hand, to manipulate physical objects. It learns entirely in simulation using the same RL algorithms and training code as OpenAI Five. OpenAI tackled the object orientation problem by using domain randomization, a simulation approach which exposes the learner to a variety of experiences rather than trying to fit to reality. The setup for Dactyl, aside from having motion tracking cameras, also has RGB cameras to allow the robot to manipulate an arbitrary object by seeing it. In 2018, OpenAI showed that the system was able to manipulate a cube and an octagonal prism.In 2019, OpenAI demonstrated that Dactyl could solve a Rubik's Cube. The robot was able to solve the puzzle 60% of the time. Objects like the Rubik's Cube introduce complex physics that is harder to model. OpenAI did this by improving the robustness of Dactyl to perturbations by using Automatic Domain Randomization, a simulation approach of generating progressively more difficult environments. ADR differs from manual domain randomization by not needing a human to specify randomization ranges.
API
In June 2020, OpenAI announced a multi-purpose API which it said was "for accessing new AI models developed by OpenAI" to let developers call on it for "any English language AI task".AgentKit
On October 6, 2025, Sam Altman announced OpenAI's new AgentKit at the 2025 Dev Day opening keynote. AgentKit is a new integrated suite of tools for building, deploying and optimizing AI agents.According to OpenAI, AgentKit builds upon its Responses API released in March, offering a more streamlined approach to agent creation. Several early adopters report significant time savings and efficiency gains when using the new agentic tools.
| Capability | Description |
| Agent Builder | A visual canvas for creating and versioning multi-agent workflows |
| ChatKit | A toolkit for embedding customizable chat-based agent experiences in your product |
| Connector Registry | Central admin panel for managing data sources across OpenAI products |
| Enhanced Evals | New Evals capabilities, including datasets, trace grading, automated prompt optimization and third-party model support |
| Guardrails | Custom guardrail configurations |
| Reinforcement Fine-Tuning | Ability to customize OpenAI's reasoning models, including custom tool calls and custom graders |
Text generation
The company has popularized generative pretrained transformers.| Model | Architecture | Parameter count | Training data | Release date | Training cost |
| GPT-1 | 12-level, 12-headed Transformer decoder, followed by linear-softmax | 117 million | BookCorpus: 4.5 GB of text, from 7,000 unpublished books of various genres. | 30 days on 8 P600 graphics cards, or 1 petaFLOPS-day | |
| GPT-2 | GPT-1, but with modified normalization | 1.5 billion | WebText: 40 GB of text, 8 million documents, from 45 million webpages upvoted on Reddit. | and | "tens of petaFLOPS-days", or 1.5 × 1021 FLOPS |
| GPT-3 | GPT-2, but with modifications to allow larger scaling | 175 billion | 499 billion tokens consisting of CommonCrawl, WebText, English Wikipedia, and two books corpora | 3640 petaFLOPS-days, or 3.1 × 1023 FLOPS | |
| GPT-3.5 | Undisclosed | 175 billion | Undisclosed | March 15, 2022 | Undisclosed |
| GPT-4 | Also trained with both text prediction and RLHF; accepts both text and images as input. Further details are not public. | Undisclosed. Estimated 1.7 trillion. | Undisclosed | Undisclosed. Estimated 2.1 × 1025 FLOPS. | |
| GPT-4o | |||||
| GPT-4.5 | |||||
| GPT-4.1 | |||||
| GPT-5 | August 7, 2025 |
OpenAI's original GPT model ("GPT-1")
The original paper on generative pre-training of a transformer-based language model was written by Alec Radford and his colleagues, and published as a preprint on OpenAI's website on June 11, 2018. It showed how a generative model of language could acquire world knowledge and process long-range dependencies by pre-training on a diverse corpus with long stretches of contiguous text.GPT-2
Generative Pre-trained Transformer 2 is an unsupervised transformer language model and the successor to OpenAI's original GPT model. GPT-2 was announced in February 2019, with only limited demonstrative versions initially released to the public. The full version of GPT-2 was not immediately released due to concerns about potential misuse, including applications for writing fake news. Some experts expressed skepticism that GPT-2 posed a significant threat.In response to GPT-2, the Allen Institute for Artificial Intelligence responded with a tool to detect "neural fake news". Other researchers, such as Jeremy Howard, warned of "the technology to totally fill Twitter, email, and the web up with reasonable-sounding, context-appropriate prose, which would drown out all other speech and be impossible to filter". In November 2019, OpenAI released the complete version of the GPT-2 language model. Several websites host interactive demonstrations of different instances of GPT-2 and other transformer models.
GPT-2's authors argue that unsupervised language models are general-purpose learners, illustrated by GPT-2 achieving state-of-the-art accuracy and perplexity on 7 of 8 zero-shot tasks.
The corpus it was trained on, called WebText, contains slightly 40 gigabytes of text from URLs shared in Reddit submissions with at least 3 upvotes. It avoids certain issues encoding vocabulary with word tokens by using byte pair encoding. This permits representing any string of characters by encoding both individual characters and multiple-character tokens.