今日推荐开源项目：《野生计科教材 Computer Science》GitHub 地址：https://github.com/ossu/computer-science
- 有期末考试（Final project）（可以使用任何语言）（为了让你学以致用，把知识运用到解决现实问题上）
今日推荐英文原文：《How integrated, open infrastructure meets tomorrow’s computing demands》
How integrated, open infrastructure meets tomorrow’s computing demands
Cloud infrastructure is quickly becoming an integral business component for nearly every company.
By virtualizing physical compute, networking, and storage resources, the cloud model of computing makes it possible to transform data center resources—previously limited, expensive and difficult to provision for users in a timely way—into flexible, elastic and easily consumable resources. Applications developers love cloud because the infrastructure resources are available on demand and are programmatically accessible via APIs, which helps them innovate more quickly; they no longer have to worry about building or requisitioning the infrastructure to test and deploy their applications—they just consume it! Operators love cloud because it enables them to deliver more powerful services faster, more cost-efficiently, more securely and with high degrees of automation. CIOs love cloud because it leads to better utilization of resources and a reduction in total cost of ownership for infrastructure. CEOs love cloud because it makes their organizations more agile—able to innovate faster and respond more quickly to set themselves apart in a competitive marketplace.
Needless to say, the cloud is popular for good reason, and enterprising companies are taking the lead in meeting the market demand with public cloud services. Three companies, in particular, have set themselves apart: Amazon Web Services, Microsoft Azure, and Google Cloud Platform together account for half of the global public cloud market. Increasingly, however, voices are being raised against the inevitability of a future in which all of the world’s infrastructure needs will be met by “The Big 3.”In fact, many large companies opt to deploy their own infrastructure rather than pay for the high margins of those Internet giants (who use those margins to finance their aggressive extension to new areas). Instead of essentially funding their own long-term extinction by paying for public cloud services, companies like Walmart, Best Buy, eBay, Paypal, AT&T, or Comcast have chosen a different path for their infrastructure needs.
And they are not alone. Others have made the choice to resist the hegemony of The Big 3 by offering their own public cloud alternative in the form of local and interoperable clouds. This is the case with OVH in France, Deutsche Telekom in Germany, City Network in Sweden, and VEXXHOST in Canada.
Likewise, in the area of scientific research, we find organizations where high-performance computing power is essential and every dollar must be used to the best advantage, and these institutions are opting to build their own infrastructure. CERN deployed an impressive private cloud (300,000 CPU cores, 800 Tb of RAM) to cost-effectively analyze the results of the Large Hadron Collider and other experiments. It’s now collaborating with other research institutions to build the compute and storage infrastructure for the Square Kilometer Array, an Earth-sized telescope expected to produce petabytes of data every day.
In China, where everything is oversized and there is an inherent mistrust of American suppliers, the choice to build your own infrastructure for both enterprise and government use is a natural one. We already find such infrastructure supporting trains for China Railway, the payment system for China UnionPay, and telephone services for millions of China Mobile customers.
To build their own infrastructure, these organizations are making a wise and strategic investment in the alternative to The Big 3: open infrastructure—combining the best programmable, flexible and interoperable open source software to build their own infrastructure. OpenStack provides the basis for this infrastructure, on which it is possible to deploy further cloud-native open source solutions like Kubernetes. Although developed by adjacent communities, these two solutions are being tested in collaboration with the various actors of the open infrastructure community to ensure compatibility.
Open infrastructure makes good business sense. Leveraging openly developed, free, and open source software, organizations can:
- Avoid vendor lock-in;
- “See under the hood” and fully understand how it all works and how to optimize the system;
- Develop innovative solutions to meet your own unique circumstances; and
- Participate directly in the improvement of the software so that it will serve you even better in the future.
The beauty of open infrastructure is its flexibility: solutions can be added according to each organization’s particular needs. Finally, the use of open solution standards makes the components of open infrastructure interoperable. The same APIs can be used on different deployments, allowing peer organizations to federate resources among different cloud environments, even to public clouds that have made the same choice of open infrastructure or use the same APIs.
The era of open infrastructure is just beginning. Already, artificial intelligence (AI) is booming, with AI applications infiltrating more and more industries, leading to a rapidly increasing demand for compute-intensive infrastructure resources. Tomorrow, mobile applications (such as augmented reality applications) will require even more processing capacity and far less network latency, forcing organizations to push processing capacity as close as possible to the user who consumes it, i.e., to “the edge.” Thus we are witnessing the advent of edge computing, a bursting out of the traditional data center model to a more decentralized, distributed and efficient model. Only open infrastructure will meet the challenges of standardization and interoperability that this revolution will bring.