OpenAI: Launching and scaling up experiments, made simple
An artificial intelligence research lab, OpenAI needed infrastructure for deep learning that would allow experiments to be run either in the cloud or in its own data center, and to easily scale. Portability, speed, and cost were the main drivers.
OpenAI began running Kubernetes on top of AWS in 2016, and in early 2017 migrated to Azure. The lab runs key experiments in fields including robotics and gaming both in Azure and in its own data centers, depending on which cluster has free capacity. “We use Kubernetes mainly as a batch scheduling system and rely on our autoscaler to dynamically scale up and down our cluster,” says Christopher Berner, Head of Infrastructure. “This lets us significantly reduce costs for idle nodes, while still providing low latency and rapid iteration.”
“Because Kubernetes provides a consistent API, we can move our research experiments very easily between clusters,” says Berner. Being able to use its own data centers when appropriate is “lowering costs and providing us access to hardware that we wouldn’t necessarily have access to in the cloud,” he adds. “As long as the utilization is high, the costs are much lower there.” Experiments can get up and running in two or three days, and scaled out in a week or two; before, that would have taken a couple of months.
By the numbers
By 10x or 50x easily
Launching and scaling experiments
Went from 2-3 months to 2-3 days
Can grow Kubernetes clusters to more than 2,500 nodes
From experiments in robotics to old-school video game play research, OpenAI’s work in artificial intelligence technology is meant to be shared.
With a mission to ensure powerful AI systems are safe, OpenAI cares deeply about open source—both benefiting from it and contributing safety technology into it. “The research that we do, we want to spread it as widely as possible so everyone can benefit,” says OpenAI’s Head of Infrastructure Christopher Berner. The lab’s philosophy—as well as its particular needs—lent itself to embracing an open source, cloud native strategy for its deep learning infrastructure.
OpenAI started running Kubernetes on top of AWS in 2016, and a year later, migrated the Kubernetes clusters to Azure. “We probably use Kubernetes differently from a lot of people,” says Berner. “We use it for batch scheduling and as a workload manager for the cluster. It’s a way of coordinating a large number of containers that are all connected together. We rely on our autoscaler to dynamically scale up and down our cluster. This lets us significantly reduce costs for idle nodes, while still providing low latency and rapid iteration.”
In the past year, Berner has overseen the launch of several Kubernetes clusters in OpenAI’s own data centers. “We run them in a hybrid model where the control planes—the Kubernetes API servers, etcd and everything—are all in Azure, and then all of the Kubernetes nodes are in our own data center,” says Berner. “The cloud is really convenient for managing etcd and all of the masters, and having backups and spinning up new nodes if anything breaks. This model allows us to take advantage of lower costs and have the availability of more specialized hardware in our own data center.”
Different teams at OpenAI currently run a couple dozen projects. While the largest-scale workloads manage bare cloud VMs directly, most of OpenAI’s experiments take advantage of Kubernetes’ benefits, including portability. “Because Kubernetes provides a consistent API, we can move our research experiments very easily between clusters,” says Berner. The on-prem clusters are generally “used for workloads where you need lots of GPUs, something like training an ImageNet model. Anything that’s CPU heavy, that’s run in the cloud. But we also have a number of teams that run their experiments both in Azure and in our own data centers, just depending on which cluster has free capacity, and that’s hugely valuable.”
Berner has made the Kubernetes clusters available to all OpenAI teams to use if it’s a good fit. “I’ve worked a lot with our games team, which at the moment is doing research on classic console games,” he says. “They had been running a bunch of their experiments on our dev servers, and they had been trying out Google cloud, managing their own VMs. We got them to try out our first on-prem Kubernetes cluster, and that was really successful. They’ve now moved over completely to it, and it has allowed them to scale up their experiments by 10x, and do that without needing to invest significant engineering time to figure out how to manage more machines. A lot of people are now following the same path.”
“Research teams can now take advantage of the frameworks we’ve built on top of Kubernetes, which make it easy to launch experiments, scale them by 10x or 50x, and take little effort to manage.”— CHRISTOPHER BERNER, HEAD OF INFRASTRUCTURE AT OPENAI
That path has been simplified by frameworks and tools that two of OpenAI’s teams have developed to handle interaction with Kubernetes. “You can just write some Python code, fill out a bit of configuration with exactly how many machines you need and which types, and then it will prepare all of those specifications and send it to the Kube cluster so that it gets launched there,” says Berner. “And it also provides a bit of extra monitoring and better tooling that’s designed specifically for these machine learning projects.”
The impact that Kubernetes has had at OpenAI is impressive. With Kubernetes, the frameworks and tooling, including the autoscaler, in place, launching experiments takes far less time. “One of our researchers who is working on a new distributed training system has been able to get his experiment running in two or three days,” says Berner. “In a week or two he scaled it out to hundreds of GPUs. Previously, that would have easily been a couple of months of work.”
Plus, the flexibility they now have to use their on-prem Kubernetes cluster when appropriate is “lowering costs and providing us access to hardware that we wouldn’t necessarily have access to in the cloud,” he says. “As long as the utilization is high, the costs are much lower in our data center. To an extent, you can also customize your hardware to exactly what you need.”
“Because Kubernetes provides a consistent API, we can move our research experiments very easily between clusters.”— CHRISTOPHER BERNER, HEAD OF INFRASTRUCTURE AT OPENAI
OpenAI is also benefiting from other technologies in the CNCF cloud native ecosystem. gRPC is used by many of its systems for communications between different services, and Prometheus is in place “as a debugging tool if things go wrong,” says Berner. “We actually haven’t had any real problems in our Kubernetes clusters recently, so I don’t think anyone has looked at our Prometheus monitoring in a while. If something breaks, it will be there.”
One of the things Berner continues to focus on is Kubernetes’ ability to scale, which is essential to deep learning experiments. OpenAI has been able to push one of its Kubernetes clusters on Azure up to more than 2,500 nodes. “I think we’ll probably hit the 5,000-machine number that Kubernetes has been tested at before too long,” says Berner, adding, “We’re definitely hiring if you’re excited about working on these things!”