Welcome!

Linux Containers Authors: Liz McMillan, XebiaLabs Blog, Akhil Sahai, Elizabeth White, Pat Romanski

Related Topics: Linux Containers

Linux Containers: Article

Making Sure Migration Is an Option

What pitfalls await your open source move

As time passes, the base of Linux users is growing in the data center, desktop, and even embedded electronic devices. Numbers from industry analysts point out that Linux server shipments have shown double-digit growth every quarter for over two years. In many cases these servers are being used for expansion or new projects. Inevitably they will be put into service to replace systems that once ran Unix or Windows.

In these cases there is usually an event like a hardware upgrade due to obsolescence or capacity concerns, software maintenance renewal, or other occasion that drives the migration. Rarely do we see a complete replacement of all legacy systems; it's commonly done piecemeal where one part of the infrastructure has been migrated in lieu of renewing an investment in other systems. This means migration to Linux. It also often means adoption of open standards that for the most part is a prime consideration for the open source community. Open source without open standards bears less advantage than a fully open system.

Linux migration is usually a matter of expansion, adding Linux into an increasingly diverse environment. Desktop PCs may be predominantly Windows, while file servers and application servers once hosted on Windows or even Novell NetWare might soon be hosted on Linux. In this case there are two hurdles that you need to overcome, especially if there is a need for communication between systems. The first is the obvious hurdle of moving from one system to another. This is a short-term problem. It's usually very disruptive and requires a considerable amount of planning and staging, though it's not especially unique as you face many of the same problems moving from one version of Windows to another or from one brand of Unix to another. The second problem is longer term and involves interoperability with existing systems. To lessen the burden in this area you should be planning well before a migration. The consideration I believe is most important is whether your systems lock you in and make it unlikely you can easily change vendors should you become unhappy with and want to investigate others. This applies to your data and network services. For example, could data stored in DB/2, Sybase, or Oracle be stored in MySQL or vice versa? Does one system have features you can't live without? Could documents originally authored in Microsoft Office be read in OpenOffice.org? If a new version of Windows becomes available, does it still allow you to access your Samba file system hosted on Linux? These are all questions you should be asking regardless of platform. For example, using Apache on a FreeBSD server is a fine choice, but if you decide that Red Hat offers a good value, can you move from one platform to another? The answer is likely yes. Does the same hold true for Web applications developed on Microsoft's Internet Information Server?

My advice is to adhere to open standards and portable file formats that are more easily migrated later on. Even if you decide to stay with one vendor, it's much better for you to select which solutions to use rather than your vendor making that choice for you. I suggest looking at things that are going on today in your enterprise. On the desktop you are likely using Microsoft Office. Their next-generation file formats are XML and intuitively should allow for easier collaboration between Microsoft and OpenOffice.org users. However, watch closely to make sure that they really do facilitate the sharing of files and that the hype is not the result of a clever PR campaign. Another thing to be wary of is the potential of a new feature in Microsoft's yet-to-be-released Vista operating system: encrypted file systems. Since the product has not gone live, it's hard to understand how this will affect cross-platform enterprises. My understanding is that this feature would likely include a level of encryption that protects the data on the hard drive (for example, the one on your Linux laptop). The idea being that if your laptop were stolen, it could not be booted under another operating system and the data would not be accessible. At first glance this sounds like a valuable feature. However, the question is: Would this prevent you from legitimately accessing that data from another operating system (like I do on my dual boot Windows/Linux laptop). In fact I often help Microsoft users migrate their data from an out-of-service Windows PC using a bootable Linux CD and a network. Will I still have this option with future products? Since this new technology prevents the bad guys from getting my data, I wonder if it also prevents me from accessing my data in a way that I choose? Is the encryption technology open source and does it allows me to authenticate my data from systems other than Windows? Does that make it possible for me to share files between systems that aren't licensed to use this new cryptology technology? I'm unsure of the answer.

I also worry about the inclusion of Digital Rights Management (DRM) technology in hardware. For example, DVI connections that are present on many modern graphics cards are very similar to the HDMI (High Definition Multimedia Interface) cables that are being used to combine audio and video into one cable for home entertainment equipment. Eventually, computers will use this same interface. Now here's a bit of trivia: the HDMI standard includes an element called HDCP (High Definition Copy Protection, developed by Intel) that does little to add value to my personal experience (I'm sure the recording and movie industries will offer some statistic about how reduced piracy keeps prices lower for me). However, it could mean that if I choose to use a "standard" graphics card with a standard PCI-E interface, I also must make sure that they adhere to less obvious standards buried within my hardware should I want to watch a DVD or HDTV broadcast. Does HDCP add value to me personally? Does it help me get more enjoyment out of my system? Should I be concerned about what's going on within the widgetry of my system? I would think so.

My point in mentioning these things is not to cause you any undue worry or to preach doom and gloom. My hope is to make you aware that while you continue to adopt mainstream technologies, you may also unknowingly be adopting features that lock you into a product's technology. Some of these features will have benefits that will be useful to you; just make sure you are getting what you bargained for. Also, what happens if these technologies add an additional point of failure? For example, the reason I know so much about HDMI is because I recently bought a plasma TV and while running cables from my HD receiver to my A/V receiver to my new TV, I found out that the receiver or the set top box didn't properly implement the standard. The result was that the copy protection software inhibited my ability to legally use my equipment. HDCP never came up in the sales process nor were the installers of my system aware of the potential problem. Take the same situation in a different content: What happens if data stored on your Windows server becomes unavailable to your Linux servers because of some obscure DRM scheme? Does it shut down your operation? Does it add unnecessary complexity and inconvenience? These are the questions I would be asking before I made my next investment in new technologies. The freedom to migrate is one that I believe to be more important than the actual act of migrating. You see, I have made my decisions and continue to make them but do so on my terms, not that of vendors that are conspiring to lock me in. In the future I don't know which conventions might be widely adopted that would prevent me from using legally purchase products in a reasonable way (of course, the consumer's version of reasonable way and the vendor's are bound to be different). What I do know is that I need to be vigilant and watch for these gotchas, and take steps to avoid them. I would advise you to do the same.

More Stories By Mark R. Hinkle

Mark Hinkle is the Senior Director, Open Soure Solutions at Citrix. He also is along-time open source expert and advocate. He is a co-founder of both the Open Source Management Consortium and the Desktop Linux Consortium. He has served as Editor-in-Chief for both LinuxWorld Magazine and Enterprise Open Source Magazine. Hinkle is also the author of the book, "Windows to Linux Business Desktop Migration" (Thomson, 2006). His blog on open source, technology, and new media can be found at http://www.socializedsoftware.com.

Comments (0)

Share your thoughts on this story.

Add your comment
You must be signed in to add a comment. Sign-in | Register

In accordance with our Comment Policy, we encourage comments that are on topic, relevant and to-the-point. We will remove comments that include profanity, personal attacks, racial slurs, threats of violence, or other inappropriate material that violates our Terms and Conditions, and will block users who make repeated violations. We ask all readers to expect diversity of opinion and to treat one another with dignity and respect.


@ThingsExpo Stories
There will be new vendors providing applications, middleware, and connected devices to support the thriving IoT ecosystem. This essentially means that electronic device manufacturers will also be in the software business. Many will be new to building embedded software or robust software. This creates an increased importance on software quality, particularly within the Industrial Internet of Things where business-critical applications are becoming dependent on products controlled by software. Qua...
SYS-CON Events has announced today that Roger Strukhoff has been named conference chair of Cloud Expo and @ThingsExpo 2016 Silicon Valley. The 19th Cloud Expo and 6th @ThingsExpo will take place on November 1-3, 2016, at the Santa Clara Convention Center in Santa Clara, CA. "The Internet of Things brings trillions of dollars of opportunity to developers and enterprise IT, no matter how you measure it," stated Roger Strukhoff. "More importantly, it leverages the power of devices and the Interne...
Large scale deployments present unique planning challenges, system commissioning hurdles between IT and OT and demand careful system hand-off orchestration. In his session at @ThingsExpo, Jeff Smith, Senior Director and a founding member of Incenergy, will discuss some of the key tactics to ensure delivery success based on his experience of the last two years deploying Industrial IoT systems across four continents.
CenturyLink has announced that application server solutions from GENBAND are now available as part of CenturyLink’s Networx contracts. The General Services Administration (GSA)’s Networx program includes the largest telecommunications contract vehicles ever awarded by the federal government. CenturyLink recently secured an extension through spring 2020 of its offerings available to federal government agencies via GSA’s Networx Universal and Enterprise contracts. GENBAND’s EXPERiUS™ Application...
The Internet of Things will challenge the status quo of how IT and development organizations operate. Or will it? Certainly the fog layer of IoT requires special insights about data ontology, security and transactional integrity. But the developmental challenges are the same: People, Process and Platform. In his session at @ThingsExpo, Craig Sproule, CEO of Metavine, demonstrated how to move beyond today's coding paradigm and shared the must-have mindsets for removing complexity from the develo...
SYS-CON Events announced today that MangoApps will exhibit at the 19th International Cloud Expo, which will take place on November 1–3, 2016, at the Santa Clara Convention Center in Santa Clara, CA. MangoApps provides modern company intranets and team collaboration software, allowing workers to stay connected and productive from anywhere in the world and from any device.
The IETF draft standard for M2M certificates is a security solution specifically designed for the demanding needs of IoT/M2M applications. In his session at @ThingsExpo, Brian Romansky, VP of Strategic Technology at TrustPoint Innovation, explained how M2M certificates can efficiently enable confidentiality, integrity, and authenticity on highly constrained devices.
The 19th International Cloud Expo has announced that its Call for Papers is open. Cloud Expo, to be held November 1-3, 2016, at the Santa Clara Convention Center in Santa Clara, CA, brings together Cloud Computing, Big Data, Internet of Things, DevOps, Digital Transformation, Microservices and WebRTC to one location. With cloud computing driving a higher percentage of enterprise IT budgets every year, it becomes increasingly important to plant your flag in this fast-expanding business opportuni...
In today's uber-connected, consumer-centric, cloud-enabled, insights-driven, multi-device, global world, the focus of solutions has shifted from the product that is sold to the person who is buying the product or service. Enterprises have rebranded their business around the consumers of their products. The buyer is the person and the focus is not on the offering. The person is connected through multiple devices, wearables, at home, on the road, and in multiple locations, sometimes simultaneously...
“delaPlex Software provides software outsourcing services. We have a hybrid model where we have onshore developers and project managers that we can place anywhere in the U.S. or in Europe,” explained Manish Sachdeva, CEO at delaPlex Software, in this SYS-CON.tv interview at @ThingsExpo, held June 7-9, 2016, at the Javits Center in New York City, NY.
"We've discovered that after shows 80% if leads that people get, 80% of the conversations end up on the show floor, meaning people forget about it, people forget who they talk to, people forget that there are actual business opportunities to be had here so we try to help out and keep the conversations going," explained Jeff Mesnik, Founder and President of ContentMX, in this SYS-CON.tv interview at 18th Cloud Expo, held June 7-9, 2016, at the Javits Center in New York City, NY.
From wearable activity trackers to fantasy e-sports, data and technology are transforming the way athletes train for the game and fans engage with their teams. In his session at @ThingsExpo, will present key data findings from leading sports organizations San Francisco 49ers, Orlando Magic NBA team. By utilizing data analytics these sports orgs have recognized new revenue streams, doubled its fan base and streamlined costs at its stadiums. John Paul is the CEO and Founder of VenueNext. Prior ...
Internet of @ThingsExpo, taking place November 1-3, 2016, at the Santa Clara Convention Center in Santa Clara, CA, is co-located with the 19th International Cloud Expo and will feature technical sessions from a rock star conference faculty and the leading industry players in the world and ThingsExpo Silicon Valley Call for Papers is now open.
The IoT is changing the way enterprises conduct business. In his session at @ThingsExpo, Eric Hoffman, Vice President at EastBanc Technologies, discussed how businesses can gain an edge over competitors by empowering consumers to take control through IoT. He cited examples such as a Washington, D.C.-based sports club that leveraged IoT and the cloud to develop a comprehensive booking system. He also highlighted how IoT can revitalize and restore outdated business models, making them profitable ...
With 15% of enterprises adopting a hybrid IT strategy, you need to set a plan to integrate hybrid cloud throughout your infrastructure. In his session at 18th Cloud Expo, Steven Dreher, Director of Solutions Architecture at Green House Data, discussed how to plan for shifting resource requirements, overcome challenges, and implement hybrid IT alongside your existing data center assets. Highlights included anticipating workload, cost and resource calculations, integrating services on both sides...
Big Data engines are powering a lot of service businesses right now. Data is collected from users from wearable technologies, web behaviors, purchase behavior as well as several arbitrary data points we’d never think of. The demand for faster and bigger engines to crunch and serve up the data to services is growing exponentially. You see a LOT of correlation between “Cloud” and “Big Data” but on Big Data and “Hybrid,” where hybrid hosting is the sanest approach to the Big Data Infrastructure pro...
"We are a well-established player in the application life cycle management market and we also have a very strong version control product," stated Flint Brenton, CEO of CollabNet,, in this SYS-CON.tv interview at 18th Cloud Expo, held June 7-9, 2016, at the Javits Center in New York City, NY.
We all know the latest numbers: Gartner, Inc. forecasts that 6.4 billion connected things will be in use worldwide in 2016, up 30 percent from last year, and will reach 20.8 billion by 2020. We're rapidly approaching a data production of 40 zettabytes a day – more than we can every physically store, and exabytes and yottabytes are just around the corner. For many that’s a good sign, as data has been proven to equal money – IF it’s ingested, integrated, and analyzed fast enough. Without real-ti...
I wanted to gather all of my Internet of Things (IOT) blogs into a single blog (that I could later use with my University of San Francisco (USF) Big Data “MBA” course). However as I started to pull these blogs together, I realized that my IOT discussion lacked a vision; it lacked an end point towards which an organization could drive their IOT envisioning, proof of value, app dev, data engineering and data science efforts. And I think that the IOT end point is really quite simple…
A critical component of any IoT project is what to do with all the data being generated. This data needs to be captured, processed, structured, and stored in a way to facilitate different kinds of queries. Traditional data warehouse and analytical systems are mature technologies that can be used to handle certain kinds of queries, but they are not always well suited to many problems, particularly when there is a need for real-time insights.