Credit: Shutterstock Every decade or so, a new IT infrastructure model barges onto the scene and changes the way organizations use technology. In the 1990s, client/server architectures put computing resources in the back room and doled them out as needed. In the 2000s, virtual machines (VMs) created the ability to emulate one computer’s resources on another. In the 2010s, cloud hit big, helping companies become more agile and cost focused. Now that we’ve entered a new decade, what model will dominate the conversation? Based on current trends and expert forecasts, it’s clear that the 2020s will be defined by containers and microservices. What exactly are containers? Getting containers right is critical to success in any transformational journey. Given the impact the technology is expected to have over the next decade, it’s worth looking more closely at what containers are and why they’re becoming so popular. SUBSCRIBE TO OUR NEWSLETTER From our editors straight to your inbox Get started by entering your email address below. Please enter a valid email address Subscribe Anybody involved in enterprise IT has at least a passing knowledge of containers. Like their physical counterparts, these virtual operating system configurations pack items away for future use. They contain all the executables an IT team needs to run everything from a small microservice (like a single HTTP endpoint) to a much larger application (like a payroll program). Each container has its own binary code, libraries, and configuration files – but it doesn’t contain any operating system images. That makes containers lighter and easier to transport than applications in traditional hardware or VM environments. 3 key benefits of containers Containers offer a wide variety of benefits – most notably are speed, choice, and the ability to optimize based on the situation. • The need for speed Speed, of course, is critical in today’s IT world. Moving software through all the various stages of development improves efficiency, increases productivity, and allows more time for testing and quality control. Fast processes enable firms to get to market faster and update more frequently. That’s the name of the game. Using containers, your teams can speed up delivery in two ways. First, because VMs contain entire operating systems, they take longer to boot up each time they’re used. Containers don’t need to boot up; the operating system is already there. Second, teams using containers can release software in smaller segments than they can in legacy waterfall processes. Containers eliminate those pieces of software that stand between the application’s execution and the actual hardware that performs the task at hand. This construct provides you with purpose-built hardware, which serves the app alone. For example, let’s say you have an AI app and want to use a graphics processing unit (GPU). Ideally, you want to use that GPU as effectively as possible. The more software between the GPU and the orchestration function, the less effectively it will work. Stripping away unnecessary software gives you a higher density of containers per computer and better utilization for that system, thus increasing the processing speed for that particular use case. • The importance of platform choice No one likes vendor lock in. Developers and IT operators alike want to pick the best platform to run their apps, which means they need the ability to migrate apps from platform to platform with little friction. Containers enable choice. At their heart, a container holds all the software needed to run the app – in the container. That’s why they call them containers! This technology is important to a multi-cloud or hybrid cloud strategy because it gives you more choices. And choice matters when you are dealing with issues such as data privacy and data residency. Also, if you must run an application close to an edge device, you need to have choice about where you can place it. Having a container strategy gives you choice. Kubernetes appears to be the accepted platform that’s being adopted across the industry today; it enables people to run any app that supports the Kubernetes APIs – that gives users choice and flexibility. Three years from now, you may want to run your apps on another cloud or on-prem. If you selected an open source standard like Kubernetes, the likelihood you’ll be able to move is very high. Open standards, supported through open source comminutes, give you that flexibility. Developing a solution container strategy is the first step. • The value of optimization A decent amount of debate concerns whether to use containers with VMs or in place of VMs. In my opinion, running containers separately enables organizations to take advantage of all the lightweight features and optimize their environments for success. You can pack more applications on a host computer using containers in place of VMs. Conversely, running containers inside a VM is like trying to attach a horse to the front of a car. Why take something this is optimized for speed and then go backward in technology to run it inside a VM? A container strategy should bridge both the public and the private areas. Containers enable you to make a step forward in your goal of collapsing the technology stack and eliminating unneeded weight that typically comes with VMs. Now is the time The technology driving transformations in the 2020s is clearly containers. They offer a host of benefits for organizations seeking speed, choice, and flexibility. As this decade begins, now is the time to start building out what your new ecosystem looks like – both in the data center and in the cloud. A vital piece of your core infrastructure should include containers. Original article first published on Hewlett Packard Enterprise’s digital publication “Enterprise.nxt”. Reproduced with permission. ____________________________________ About Robert Christiansen Robert Christiansen is a key executive in the CTO Office of Hybrid IT at Hewlett Packard Enterprise, setting the strategy and evangelizing the company’s vision. Hybrid IT is a $25B group, and central to HPE’s core technologies. In this role, Robert spends his time with key global clients and partners, deepening the relationships and aligning the joint technology efforts to improve the way people live and work. Robert is a contributing writer for CIO, Forbes, TechTarget, and numerous industry magazines and is a major contributor to The Doppler, the cloud industry’s thought-leadership publication. He is also a keynote speaker at numerous technology and HPE led events, clearly articulating technology shifts while having a great time doing it! Related content brandpost How ML Ops Can Help Scale Your AI and ML Models Machine learning operations, or ML Ops, can help enterprises improve governance and regulatory compliance, automation, and production model quality. By Richard Hatheway Apr 07, 2022 7 mins Machine Learning IT Leadership brandpost Edge Computing is Thriving in the Cloud Era Todayu2019s edge technology is not just bolstering profits, but also helping reduce risk and improve products, services, and customer experience. By Denis Vilfort, Al Madden Apr 06, 2022 11 mins Edge Computing Artificial Intelligence IT Leadership brandpost 5 Types of Costly Data Waste and How to Avoid Them Poor choices in data infrastructure and data habits can lead to data waste u2013 but a comprehensive data strategy can help resolve the problem. By Ellen Friedman Mar 29, 2022 11 mins Data Center Management Data Architecture IT Leadership brandpost 2022 is the Year of the Edge By Matthew Hausmann Feb 28, 2022 9 mins Data Science Edge Computing IT Leadership Podcasts Videos Resources Events SUBSCRIBE TO OUR NEWSLETTER From our editors straight to your inbox Get started by entering your email address below. Please enter a valid email address Subscribe