How integrated, open infrastructure meets tomorrow's computing demands

How integrated, open infrastructure meets tomorrow's computing demands

The beauty of open infrastructure is its flexibility: solutions can be added according to each organization's particular needs.

How integrated, open infrastructure meets tomorrow's computing demands
Image by : 

opensource.com

x

Get the newsletter

Join the 85,000 open source advocates who receive our giveaway alerts and article roundups.

Cloud infrastructure is quickly becoming an integral business component for nearly every company.

By virtualizing physical compute, networking, and storage resources, the cloud model of computing makes it possible to transform data center resources—previously limited, expensive and difficult to provision for users in a timely way—into flexible, elastic and easily consumable resources. Applications developers love cloud because the infrastructure resources are available on demand and are programmatically accessible via APIs, which helps them innovate more quickly; they no longer have to worry about building or requisitioning the infrastructure to test and deploy their applications—they just consume it! Operators love cloud because it enables them to deliver more powerful services faster, more cost-efficiently, more securely and with high degrees of automation. CIOs love cloud because it leads to better utilization of resources and a reduction in total cost of ownership for infrastructure. CEOs love cloud because it makes their organizations more agile—able to innovate faster and respond more quickly to set themselves apart in a competitive marketplace.

Needless to say, the cloud is popular for good reason, and enterprising companies are taking the lead in meeting the market demand with public cloud services. Three companies, in particular, have set themselves apart: Amazon Web Services, Microsoft Azure, and Google Cloud Platform together account for half of the global public cloud market. Increasingly, however, voices are being raised against the inevitability of a future in which all of the world's infrastructure needs will be met by "The Big 3."

In fact, many large companies opt to deploy their own infrastructure rather than pay for the high margins of those Internet giants (who use those margins to finance their aggressive extension to new areas). Instead of essentially funding their own long-term extinction by paying for public cloud services, companies like Walmart, Best Buy, eBay, Paypal, AT&T, or Comcast have chosen a different path for their infrastructure needs.

And they are not alone. Others have made the choice to resist the hegemony of The Big 3 by offering their own public cloud alternative in the form of local and interoperable clouds. This is the case with OVH in France, Deutsche Telekom in Germany, City Network in Sweden, and VEXXHOST in Canada.

Likewise, in the area of scientific research, we find organizations where high-performance computing power is essential and every dollar must be used to the best advantage, and these institutions are opting to build their own infrastructure. CERN deployed an impressive private cloud (300,000 CPU cores, 800 Tb of RAM) to cost-effectively analyze the results of the Large Hadron Collider and other experiments. It's now collaborating with other research institutions to build the compute and storage infrastructure for the Square Kilometer Array, an Earth-sized telescope expected to produce petabytes of data every day.

In China, where everything is oversized and there is an inherent mistrust of American suppliers, the choice to build your own infrastructure for both enterprise and government use is a natural one. We already find such infrastructure supporting trains for China Railway, the payment system for China UnionPay, and telephone services for millions of China Mobile customers.

To build their own infrastructure, these organizations are making a wise and strategic investment in the alternative to The Big 3: open infrastructure—combining the best programmable, flexible and interoperable open source software to build their own infrastructure. OpenStack provides the basis for this infrastructure, on which it is possible to deploy further cloud-native open source solutions like Kubernetes. Although developed by adjacent communities, these two solutions are being tested in collaboration with the various actors of the open infrastructure community to ensure compatibility.

Open infrastructure makes good business sense. Leveraging openly developed, free, and open source software, organizations can:

  • Avoid vendor lock-in;
  • "See under the hood" and fully understand how it all works and how to optimize the system;
  • Develop innovative solutions to meet your own unique circumstances; and
  • Participate directly in the improvement of the software so that it will serve you even better in the future.

The beauty of open infrastructure is its flexibility: solutions can be added according to each organization's particular needs. Finally, the use of open solution standards makes the components of open infrastructure interoperable. The same APIs can be used on different deployments, allowing peer organizations to federate resources among different cloud environments, even to public clouds that have made the same choice of open infrastructure or use the same APIs.

The era of open infrastructure is just beginning. Already, artificial intelligence (AI) is booming, with AI applications infiltrating more and more industries, leading to a rapidly increasing demand for compute-intensive infrastructure resources. Tomorrow, mobile applications (such as augmented reality applications) will require even more processing capacity and far less network latency, forcing organizations to push processing capacity as close as possible to the user who consumes it, i.e., to “the edge.” Thus we are witnessing the advent of edge computing, a bursting out of the traditional data center model to a more decentralized, distributed and efficient model. Only open infrastructure will meet the challenges of standardization and interoperability that this revolution will bring.

Learn more about open infrastructure

Open infrastructure proponents from around the world will be gathering May 21-24 at the OpenStack Open Infrastructure Summit in Vancouver to share case studies and best practices and work together to solve integration challenges. The agenda includes notable speakers from ARM, AT&T, China Mobile, China Unicom, Google, Heptio, Hyper HQ, Oath Inc., Progressive Insurance, Target, Verizon, Walmart, and many others.

The summit is focused on helping users compose, integrate, and operate open infrastructure technologies to solve real problems at scale. In addition to sessions on OpenStack, there will be featured sessions on the newest project at the Foundation, Kata Containers, and a strong focus on other open source technologies relevant to infrastructure operators, including Kubernetes, Docker, Ansible, Ceph, Istio, Envoy, Spinnaker, Tungsten Fabric (formerly OpenContrail), ONAP, OPNFV and many more.

The full event agenda is organized by use cases, including AI and machine learning; edge computing; NFV; CI/CD; container infrastructure; and public, private and multi-cloud strategies.

It's going to be an inspiring and productive four days. I hope to see you there!

About the author

Thierry Carrez - Thierry Carrez has been the Release Manager for the OpenStack project since its inception, coordinating the effort and facilitating collaboration between contributors. He is the elected chair of the OpenStack Technical Committee, which is in charge of the technical direction of the project. A Python Software Foundation fellow, he was previously the Technical lead for Ubuntu Server at Canonical, and an operational manager for the Gentoo Linux Security Team.