Welcome!

Linux Containers Authors: Liz McMillan, Yeshim Deniz, Elizabeth White, Pat Romanski, Zakia Bouachraoui

Related Topics: SDN Journal, Java IoT, Microservices Expo, Linux Containers, Containers Expo Blog, @CloudExpo

SDN Journal: Blog Post

Different Shades of Invisible

We need to get to a network that is invisible enough to allow you to create coarse provisioning policies

We love analogies. No matter what the topic, analogies are a great way to explain something in a different context to make a specific point with a frame of reference that may be more familiar to those we are making a point to. There is one that seems to come back over and over again in our industry, the one that compares the network to the power grid, network connections to power plugs.  I had not heard it for a while but at Interop last week, I heard it used twice in booth demonstrations as part of plug and play pitches. And I really do not like that analogy.

The comparison to power comes from the angle of it just being there and available. Plug something in and it turns on. Its universal and it’s simple, anyone can use it. It’s highly standardized, you can buy bits and pieces that work together in any hardware store. It’s reliable. You just get a bill each month, pay it, and in many places you can even pick your provider different from the owner of the cable that actually plugs into your house. It’s all good, it’s mostly invisible.

Except that all of these come with their own challenges. The overall power grid reliability is well below 99.99%. Across the US, the average outage minutes per year is hovering in the 2 hour range, in the Northeast almost double that. There are only very people I know that do not have a backup generator, just in case, and get to use it at least once a year, ignoring the smaller outages where we just live in the dark for a few minutes. It is very standardized, with plenty of standard variations from country to country, thank goodness for $30 converters. And the cost, well, it’s rising faster than inflation (almost twice as fast in fact).

Now, this article is not meant to be a complaining session about power. When power works and you forget about the bill, it is very much invisible, which is the only piece of the analogy I do like, maybe.

Networks should be invisible, but not in the same sense power should be invisible. When comparing networks to power it always comes down to “it’s just there, plentiful, accessible, and just works”. And (not surprising), I think the network can do much better than “just be there”. Networks need to migrate away from an entity that is architected, designed, implemented, managed and debugged separately from the applications it serves. Networks have to become an integral part of the application infrastructure.

Now, any two (or more) distinctly different entities need to be glued together. No different for applications, servers and networks. Applications are glued to servers by means of the OS. As a consumer of an application, or even an administrator, the amount of work you have to do to put these two together is (usually) fairy minimal. The creator of the OS has provided tons of hooks and tools to give the application what it needs, the application provides to and asks what it wants from the OS.

The glue between the server and the network is much harsher, much less refined. There is an expectation of matching configuration, very little is actually transacted between the two to get them to work together nicely. Yes there is DHCP and perhaps you can even count LLDP in that, but the amount of alignment of information between the two so they can work together seamlessly without human intervention is minimal. And if something is not quite working the way it should, you need heavily certified network engineers and tons of homegrown tools and scripts to figure out why. That is by no means a knock on network certifications (that is for @cloudtoad to comment on), but it should not be that hard. The exchange is extremely focused on data plane traffic for transportation, very little provisioning, debug and other feedback is exchanged and reacted to.

We need to get to a network that is invisible enough to allow you to create coarse provisioning policies created from application templates. There should be no need for me to configure individual ports and protocols on ports. There should be no need for me to configure how network components are tied together and provide optimized connectivity based on the needs of its customers. The network needs to provide debugging tools that interact with applications in feedback loops. It needs to provide enough details for orchestration systems to adjust their placement of apps. Or in Plexxi’s case, take in information from orchestration systems to change the L1, L2 and L3 behavior of the network to the requested demand. And that orchestration system could be a specialized Hadoop controller or a far more generic cloud orchestrator.

The network needs to cease to exist as a separate entity, it needs to become an integrated part of the application infrastructure that uses its services. That’s the kind of invisible we need to achieve.

[Today's Fun Fact: An Iranian man paid $500 to a wizard to become invisible. The man then walked into a bank and grabbed money from people's hands believing he was invisible. He was promptly arrested and convicted. Conclusion: pay more or pick a better wizard next time.]

The post Different shades of invisible appeared first on Plexxi.

Read the original blog entry...

More Stories By Marten Terpstra

Marten Terpstra is a Product Management Director at Plexxi Inc. Marten has extensive knowledge of the architecture, design, deployment and management of enterprise and carrier networks.

IoT & Smart Cities Stories
The deluge of IoT sensor data collected from connected devices and the powerful AI required to make that data actionable are giving rise to a hybrid ecosystem in which cloud, on-prem and edge processes become interweaved. Attendees will learn how emerging composable infrastructure solutions deliver the adaptive architecture needed to manage this new data reality. Machine learning algorithms can better anticipate data storms and automate resources to support surges, including fully scalable GPU-c...
Machine learning has taken residence at our cities' cores and now we can finally have "smart cities." Cities are a collection of buildings made to provide the structure and safety necessary for people to function, create and survive. Buildings are a pool of ever-changing performance data from large automated systems such as heating and cooling to the people that live and work within them. Through machine learning, buildings can optimize performance, reduce costs, and improve occupant comfort by ...
The explosion of new web/cloud/IoT-based applications and the data they generate are transforming our world right before our eyes. In this rush to adopt these new technologies, organizations are often ignoring fundamental questions concerning who owns the data and failing to ask for permission to conduct invasive surveillance of their customers. Organizations that are not transparent about how their systems gather data telemetry without offering shared data ownership risk product rejection, regu...
René Bostic is the Technical VP of the IBM Cloud Unit in North America. Enjoying her career with IBM during the modern millennial technological era, she is an expert in cloud computing, DevOps and emerging cloud technologies such as Blockchain. Her strengths and core competencies include a proven record of accomplishments in consensus building at all levels to assess, plan, and implement enterprise and cloud computing solutions. René is a member of the Society of Women Engineers (SWE) and a m...
Poor data quality and analytics drive down business value. In fact, Gartner estimated that the average financial impact of poor data quality on organizations is $9.7 million per year. But bad data is much more than a cost center. By eroding trust in information, analytics and the business decisions based on these, it is a serious impediment to digital transformation.
Digital Transformation: Preparing Cloud & IoT Security for the Age of Artificial Intelligence. As automation and artificial intelligence (AI) power solution development and delivery, many businesses need to build backend cloud capabilities. Well-poised organizations, marketing smart devices with AI and BlockChain capabilities prepare to refine compliance and regulatory capabilities in 2018. Volumes of health, financial, technical and privacy data, along with tightening compliance requirements by...
Predicting the future has never been more challenging - not because of the lack of data but because of the flood of ungoverned and risk laden information. Microsoft states that 2.5 exabytes of data are created every day. Expectations and reliance on data are being pushed to the limits, as demands around hybrid options continue to grow.
Digital Transformation and Disruption, Amazon Style - What You Can Learn. Chris Kocher is a co-founder of Grey Heron, a management and strategic marketing consulting firm. He has 25+ years in both strategic and hands-on operating experience helping executives and investors build revenues and shareholder value. He has consulted with over 130 companies on innovating with new business models, product strategies and monetization. Chris has held management positions at HP and Symantec in addition to ...
Enterprises have taken advantage of IoT to achieve important revenue and cost advantages. What is less apparent is how incumbent enterprises operating at scale have, following success with IoT, built analytic, operations management and software development capabilities - ranging from autonomous vehicles to manageable robotics installations. They have embraced these capabilities as if they were Silicon Valley startups.
As IoT continues to increase momentum, so does the associated risk. Secure Device Lifecycle Management (DLM) is ranked as one of the most important technology areas of IoT. Driving this trend is the realization that secure support for IoT devices provides companies the ability to deliver high-quality, reliable, secure offerings faster, create new revenue streams, and reduce support costs, all while building a competitive advantage in their markets. In this session, we will use customer use cases...