|By Lori MacVittie||
|September 15, 2009 05:45 AM EDT||
Tower of Babel by Pieter Bruegel the Elder. One of the topics surrounding cloud computing that continues to rear its ugly head is the problem of portability across clouds. Avoiding vendor lock-in has been problematic since the day the first line of proprietary code was written and cloud computing does nothing to address this. If anything, cloud makes this worse because one of its premises is that users (that’s you, IT staff) need not concern themselves with the underlying infrastructure. It’s a service, right, so you just use it and don’t worry about it.
Let’s assume for a moment that you can easily move applications from data center to cloud to cloud. Plenty of folks are working on that, but very few of them address the “rest of the story”: the metastructure.
Metastructure contains the metadata that describes the network, application network, and security infrastructure providing all those “don’t worry about” services cloud providers offer. Load balancing, firewalls, IPS, IDS, application acceleration, secure remote access. If you’ve spent time with your cloud provider tweaking those services – or configuring them yourself – then moving to a new cloud provider is not only a huge investment in time, it’s actually going to be painful because you’re essentially going to have to recreate every metastructure configuration again.
Yes, you’ve done this inside your own data center for years. Every forklift replacement or upgrade of infrastructure has come with its own load of baggage in the configuration arena. Switching out vendor equipment – especially core components – can be extremely painful, especially when configurations need to essentially be “translated” between them. But cloud makes this worse because technically speaking you don’t even have access to the existing configurations. You can’t see them, you can’t have them, and you can’t run them through whatever “upgrade” or “migration” script your new vendor offers to ease the process.
Are you depressed yet?
There’s been some talk of including metastructure data with the virtual machine, but the problem with this is that it almost always requires that the meta data be wrapped up using a proprietary API, such as is provided by VMware. That’s okay if you restrict yourself to only cloud providers that use the same virtualization technology, but not okay if you want to be able to make a move from one technology to another. It also assumes that the metadata is specific to the infrastructure, which is even more unlikely when moving between cloud providers.
HOW ABOUT A CLOUD-BASED CMDB (Configuration Management Database)?
There are several ongoing efforts to address this very scenario because it is so painful. Most of them would, if adopted, require vendors to implement support for a specific standard so that configurations can be managed and exchanged in that standard format. That makes sense, that’s how we’ve always handled translation of data between disparate systems that don’t speak the same language. In the application world we call the process of mapping one format to another “integration” and you can easily evoke a look of terror on a co-worker’s face just by saying the word within their range of hearing. Go ahead, try it. Just make sure they aren’t carrying anything heavy that can be easily thrown at you when you do.
CMDB (Configuration Management Database) technology is another method of addressing the problem of, well, managing configurations. These solutions store configuration of a wide variety of infrastructure solutions – from routers and switches to web and application servers to application delivery controllers. They do a great job of managing configuration and can even “push” configuration out to devices if so desired. But the configurations stored and managed in a CMDB are product-specific, not generic, so they can’t adequately today address the problem of portability.
You can probably see where this is going: a cross between CMDB and a nice, industry-wide standard would probably do the trick, wouldn’t it? And if it was public (in the sense that any application or service is public on the network – that is, accessible via the Internet to any cloud provider or customer site) then cloud providers and organizations alike could take advantage of that configuration management mechanism and use it to their advantage. Portability becomes possible rather than fantasy.
Cloud providers and organizations alike are likely to stop right there. Sharing configuration of infrastructure and core components is just asking for trouble. If ever such a cloud-based CMDB were compromised, well…let’s just say it would be A Very Bad Thing.
But what if the actual metadata, the configuration information, were stored either in the enterprise or the cloud provider (or both), and merely pushed and pulled via a public mechanism on-demand? Configuration isn’t changed all that often and if an organization is moving between clouds they certainly know when they’re doing it. If there was some mechanism through which metastructure could be published and to which infrastructure could subscribe then when changes were made or providers changed that metastructure data could be easily grabbed from the public cloud-CMDB system (cloud catalog, anyone?) and interpreted into product-specific configuration by the products themselves.
Think of it like SOA clients pulling WSDL (Web Services Description Language) from a UDDI (Universal Description, Discovery, and Integration) server. The SOA client pulls the WSDL, which describes the service(s), configures itself appropriately, and then is able to make use of those services. The intent of introducing UDDI was a service-catalog that could be polled on-demand to provide the latest information about the service and describe it in an abstract, vendor-neutral way such that any client could access any service, regardless of implementation language or environment. Sounds a lot like what we want for infrastructure portability, doesn’t it?
That’s where PubSubHubub comes in. While this draft standard for a publish-subscribe system is generally being leveraged by software developers to enable faster sharing of information across the Internet, it is also a fine example of a system that could be used by infrastructure 2.0 solutions to share metastructure. Consider the existence of a public PubSubhubub Hub, like Google’s public PubSubHubub Hub, and how it might be leveraged to share metastructure between clouds or the organization and the cloud.
Note that XMPP is used today by at least one cloud provider to enable distributed cloud management in a nature very similar to that of PubSubhubhub.
In any case, the specific implementation of the configuration “hub” is relatively unimportant; what’s important is that (a) customers can publish a vendor-neutral metastructure to an isolated channel that communicates their specific infrastructure needs and (b) providers can subscribe, at will, to customer topics and retrieve metastructure in a way that allows their infrastructure to in turn configuration itself (or be configured by the provider’s system, as is required by the provider’s implementation).
Early on it would be necessary for the cloud provider to provide the “translation” and configuration services simply because even if a metastructure standard existed today (and it doesn’t) it would take months and possibly years before all the possible infrastructure vendors were able to update their systems to interpret the standard. If the provider implements a configuration “gateway”, however, he can immediately take advantage of such a standard and use existing skills and knowledge gained from its automation and orchestration of its cloud to configure the infrastructure appropriately based on the metastructure. This has the added advantage of “hiding” the infrastructure implementation from the outside world, which for some providers is a very important thing to do.
SOME CONFIGURATIONS ARE INHERENTLY VENDOR SPECIFIC
That’s okay for two reasons: first, we ensure that the metadata description is XML-based, because it’s extensible. If we build into the standard a way to extend it naturally such as is provided with XML the interpreters (configuration “gateways”) can either (a) translate if it can or (b) ignore.
Consider the use of OVF (Open Virtualization Format) to further describe what is called a Virtual Machine Contract (VMC):
For each virtual system, the associated metadata is described in a set of specific sections. The VirtualHardwareSection describes the virtual hardware required including the amount of memory, number of CPUs, information about network interfaces, etc. The OperatingSystemSection describes the guest operating system that will run in the virtual system. The ProductSection provides basic information such as the name and vendor of the appliance and can also specify a set of properties that can be used to customize the appliance.
While VMC is very basic at this point, it’s a good start at providing the foundation for building out a more complete, standards-based description of the metastructure necessary to configure an infrastructure to deploy a specific application in a virtual machine format. Using this as the basis for metadata exchange – when fully described – via a public hub could alleviate most of the issues with sharing infrastructure metadata (metastructure) across clouds in a generally vendor non-specific manner. In other words, portability of both the virtual machine and the specific infrastructure configurations necessary to optimally execute and deliver the application to the end user in the most fast and secure manner possible.
We’re nowhere near this point, by the way. VMC needs to be fleshed out as far as standard metadata goes for infrastructure (perhaps a good chore for the SRI Infrastructure 2.0 Working Group) and vendors would need to adopt and extend out the ProductSection of VMC for product specific configuration that isn’t included in the base format. And PubSubHubub would need to be proven to be a secure method of exchanging the metastructure across clouds. What is likely is that as we move forward trying to extend the plateau of collaboration down the stack toward the core infrastructure is that a new set of tools, products, solutions, and services will emerge to fill the unavoidable gaps in the standards, e.g. a service-based cloud configuration hub offering translation of proprietary metastructure data to some other proprietary metastructure data.
Perhaps there’s a better way overall, and OVF/VMC and PubSubHubub will simply remain in our memories as the catalyst and template for a different set of standards providing portability across clouds. But there is a way to provide this level of portability and collaboration across clouds, across the infrastructure and the application. The need – and perhaps more importantly the belief that it’s necessary to address the need – is growing.
UPDATE: Christofer Hoff pointed out that vCloud has been submitted to the DMTF for standardization, technically making it "open" rather than "proprietary." It is still only implemented by VMware technologies, so for the time being it might as well be proprietary, but this may change in the future.
Related blogs & articles:
- Virtual Machine Contracts for Data Center and Cloud Computing Environments [PDF]
- Draft: PubSubHubub Core 0.2
- DMTF Virtual Management Initiative (OVF Specification)
- Who owns application delivery meta-data in the cloud?
- More on the Meta-data Menagerie
- Governance: Service Catalogs and the Cloud
- OVF: A few layers short of a full stack
- Interoperability between clouds requires more than just VM portability
In addition to all the benefits, IoT is also bringing new kind of customer experience challenges - cars that unlock themselves, thermostats turning houses into saunas and baby video monitors broadcasting over the internet. This list can only increase because while IoT services should be intuitive and simple to use, the delivery ecosystem is a myriad of potential problems as IoT explodes complexity. So finding a performance issue is like finding the proverbial needle in the haystack.
Jul. 28, 2016 05:45 AM EDT Reads: 2,314
The 19th International Cloud Expo has announced that its Call for Papers is open. Cloud Expo, to be held November 1-3, 2016, at the Santa Clara Convention Center in Santa Clara, CA, brings together Cloud Computing, Big Data, Internet of Things, DevOps, Digital Transformation, Microservices and WebRTC to one location. With cloud computing driving a higher percentage of enterprise IT budgets every year, it becomes increasingly important to plant your flag in this fast-expanding business opportuni...
Jul. 28, 2016 03:15 AM EDT Reads: 2,637
Large scale deployments present unique planning challenges, system commissioning hurdles between IT and OT and demand careful system hand-off orchestration. In his session at @ThingsExpo, Jeff Smith, Senior Director and a founding member of Incenergy, will discuss some of the key tactics to ensure delivery success based on his experience of the last two years deploying Industrial IoT systems across four continents.
Jul. 28, 2016 03:00 AM EDT Reads: 1,621
The Internet of Things will challenge the status quo of how IT and development organizations operate. Or will it? Certainly the fog layer of IoT requires special insights about data ontology, security and transactional integrity. But the developmental challenges are the same: People, Process and Platform. In his session at @ThingsExpo, Craig Sproule, CEO of Metavine, demonstrated how to move beyond today's coding paradigm and shared the must-have mindsets for removing complexity from the develo...
Jul. 28, 2016 02:30 AM EDT Reads: 1,544
SYS-CON Events announced today that MangoApps will exhibit at the 19th International Cloud Expo, which will take place on November 1–3, 2016, at the Santa Clara Convention Center in Santa Clara, CA. MangoApps provides modern company intranets and team collaboration software, allowing workers to stay connected and productive from anywhere in the world and from any device.
Jul. 28, 2016 02:15 AM EDT Reads: 1,403
IoT is rapidly changing the way enterprises are using data to improve business decision-making. In order to derive business value, organizations must unlock insights from the data gathered and then act on these. In their session at @ThingsExpo, Eric Hoffman, Vice President at EastBanc Technologies, and Peter Shashkin, Head of Development Department at EastBanc Technologies, discussed how one organization leveraged IoT, cloud technology and data analysis to improve customer experiences and effi...
Jul. 28, 2016 02:00 AM EDT Reads: 2,057
The IETF draft standard for M2M certificates is a security solution specifically designed for the demanding needs of IoT/M2M applications. In his session at @ThingsExpo, Brian Romansky, VP of Strategic Technology at TrustPoint Innovation, explained how M2M certificates can efficiently enable confidentiality, integrity, and authenticity on highly constrained devices.
Jul. 28, 2016 01:45 AM EDT Reads: 1,110
In today's uber-connected, consumer-centric, cloud-enabled, insights-driven, multi-device, global world, the focus of solutions has shifted from the product that is sold to the person who is buying the product or service. Enterprises have rebranded their business around the consumers of their products. The buyer is the person and the focus is not on the offering. The person is connected through multiple devices, wearables, at home, on the road, and in multiple locations, sometimes simultaneously...
Jul. 28, 2016 01:15 AM EDT Reads: 857
“delaPlex Software provides software outsourcing services. We have a hybrid model where we have onshore developers and project managers that we can place anywhere in the U.S. or in Europe,” explained Manish Sachdeva, CEO at delaPlex Software, in this SYS-CON.tv interview at @ThingsExpo, held June 7-9, 2016, at the Javits Center in New York City, NY.
Jul. 28, 2016 01:00 AM EDT Reads: 1,628
"We've discovered that after shows 80% if leads that people get, 80% of the conversations end up on the show floor, meaning people forget about it, people forget who they talk to, people forget that there are actual business opportunities to be had here so we try to help out and keep the conversations going," explained Jeff Mesnik, Founder and President of ContentMX, in this SYS-CON.tv interview at 18th Cloud Expo, held June 7-9, 2016, at the Javits Center in New York City, NY.
Jul. 27, 2016 10:30 PM EDT Reads: 1,412
Internet of @ThingsExpo, taking place November 1-3, 2016, at the Santa Clara Convention Center in Santa Clara, CA, is co-located with the 19th International Cloud Expo and will feature technical sessions from a rock star conference faculty and the leading industry players in the world and ThingsExpo Silicon Valley Call for Papers is now open.
Jul. 27, 2016 10:00 PM EDT Reads: 2,676
The IoT is changing the way enterprises conduct business. In his session at @ThingsExpo, Eric Hoffman, Vice President at EastBanc Technologies, discussed how businesses can gain an edge over competitors by empowering consumers to take control through IoT. He cited examples such as a Washington, D.C.-based sports club that leveraged IoT and the cloud to develop a comprehensive booking system. He also highlighted how IoT can revitalize and restore outdated business models, making them profitable ...
Jul. 27, 2016 09:45 PM EDT Reads: 2,012
"delaPlex is a software development company. We do team-based outsourcing development," explained Mark Rivers, COO and Co-founder of delaPlex Software, in this SYS-CON.tv interview at 18th Cloud Expo, held June 7-9, 2016, at the Javits Center in New York City, NY.
Jul. 27, 2016 08:00 PM EDT Reads: 2,040
We all know the latest numbers: Gartner, Inc. forecasts that 6.4 billion connected things will be in use worldwide in 2016, up 30 percent from last year, and will reach 20.8 billion by 2020. We're rapidly approaching a data production of 40 zettabytes a day – more than we can every physically store, and exabytes and yottabytes are just around the corner. For many that’s a good sign, as data has been proven to equal money – IF it’s ingested, integrated, and analyzed fast enough. Without real-ti...
Jul. 27, 2016 07:30 PM EDT Reads: 1,120
I wanted to gather all of my Internet of Things (IOT) blogs into a single blog (that I could later use with my University of San Francisco (USF) Big Data “MBA” course). However as I started to pull these blogs together, I realized that my IOT discussion lacked a vision; it lacked an end point towards which an organization could drive their IOT envisioning, proof of value, app dev, data engineering and data science efforts. And I think that the IOT end point is really quite simple…
Jul. 27, 2016 06:45 PM EDT Reads: 1,194
"There's a growing demand from users for things to be faster. When you think about all the transactions or interactions users will have with your product and everything that is between those transactions and interactions - what drives us at Catchpoint Systems is the idea to measure that and to analyze it," explained Leo Vasiliou, Director of Web Performance Engineering at Catchpoint Systems, in this SYS-CON.tv interview at 18th Cloud Expo, held June 7-9, 2016, at the Javits Center in New York Ci...
Jul. 27, 2016 06:45 PM EDT Reads: 2,075
Big Data, cloud, analytics, contextual information, wearable tech, sensors, mobility, and WebRTC: together, these advances have created a perfect storm of technologies that are disrupting and transforming classic communications models and ecosystems. In his session at @ThingsExpo, Erik Perotti, Senior Manager of New Ventures on Plantronics’ Innovation team, provided an overview of this technological shift, including associated business and consumer communications impacts, and opportunities it ...
Jul. 27, 2016 04:30 PM EDT Reads: 229
A critical component of any IoT project is what to do with all the data being generated. This data needs to be captured, processed, structured, and stored in a way to facilitate different kinds of queries. Traditional data warehouse and analytical systems are mature technologies that can be used to handle certain kinds of queries, but they are not always well suited to many problems, particularly when there is a need for real-time insights.
Jul. 27, 2016 04:30 PM EDT Reads: 1,886
You think you know what’s in your data. But do you? Most organizations are now aware of the business intelligence represented by their data. Data science stands to take this to a level you never thought of – literally. The techniques of data science, when used with the capabilities of Big Data technologies, can make connections you had not yet imagined, helping you discover new insights and ask new questions of your data. In his session at @ThingsExpo, Sarbjit Sarkaria, data science team lead ...
Jul. 27, 2016 04:15 PM EDT Reads: 1,155
Is your aging software platform suffering from technical debt while the market changes and demands new solutions at a faster clip? It’s a bold move, but you might consider walking away from your core platform and starting fresh. ReadyTalk did exactly that. In his General Session at 19th Cloud Expo, Michael Chambliss, Head of Engineering at ReadyTalk, will discuss why and how ReadyTalk diverted from healthy revenue and over a decade of audio conferencing product development to start an innovati...
Jul. 27, 2016 04:00 PM EDT Reads: 1,077