Welcome!

Linux Containers Authors: Elizabeth White, AppDynamics Blog, Christopher Keene, Janakiram MSV, Greg O'Connor

Related Topics: Linux Containers

Linux Containers: Article

Is Linux Enterprise Ready?

Making the move

No doubt this topic has been debated to death; however, as I have a different perspective on this issue, I reckon it's worth writing down.

Over the past few weeks I've been involved with one of our local customers who, after a lot of consideration, has decided to make the jump to Linux. This was no quick decision, mind you, and was more than a "I'm tired of paying Microsoft for licenses" thing.

Why the Move?
Linux made its way into the organization when I chose to use it as a desktop system while I was still consulting for the customer as a DBA and J2EE developer. (Yeah, I know, a weird combo, but I've never liked scripting languages that much, so I chose Java/J2EE for my DBA tools.) I got pretty uptight when Eclipse and Windows (the company standard) decided to crash or hang on me every few hours, and I moved back to Linux.

As a DBA, you tend to get involved with all sorts of issues, mainly because in the case of a reasonably large or busy application, the database is normally the first thing that takes the blame in the case of a performance dip. On one of my investigations, I found that the database had trouble sending data back to the client applications (network waits). The networking guys were just laughing at me and said that the database shouldn't send so much data back. (!?!)

There were some variables involved: all of the servers (application, Web, database, mail) were hosted at a different site, and all traffic (including Web traffic) was being routed to (through) the offsite location (a local ISP). My theory was that some people were misusing the Web, as my investigation pointed out that HTTP traffic was extremely high.

This was really the first case in which I could implement Linux with a direct business benefit. After numerous consultations with the client (a Windows-only type), I decided to take an older PC that was sitting around in the storeroom, slap some SCSI drives into it, and install Mandrake Linux on it. My reasoning here was that Mandrake is a pretty friendly O/S for a Windows-skilled "LANnie" to pick up. I then went for Squid proxy and installed a Web reporting tool (squint) onto the "proxy server," as it was called. This allowed us to report, per user, the amount of time spent on Web sites, the amount of data downloaded, site details, etc. We could basically pinpoint exactly who was surfing, for how long, and what sites they were viewing. We had to change some of the client browser settings to point to the proxy (you change firewall rules to allow only HTTP traffic from the proxy server, and point all client browsers to the proxy).

We gathered statistics for two or three days, and our first report proved that my hunch was correct. Some guy in the admin department was using up a lot of our much-needed bandwidth by downloading, well, porn; some other people were using the Web for audio streaming, and others were downloading MP3s and games, etc. Now, in South Africa, bandwidth is expensive and slow. We only have one provider of leased or other telco lines (changing in 2006), and 3G isn't what it should be (yet).

We blocked some sites; the client issued some final warnings; and, by the next day, the system was flying again. I started using our "proxy server" for more things, to see how much we could get out of a simple PC (about 128MB RAM, 40GB disk space, 1 GHz Pentium, 3 CPU). We implemented CVS, an open source version control tool. We gave users in the operations department a home directory to back up documents. We set up some print queues.

The CIO was pretty happy with what we managed to squeeze out of the PC. The key thing to realize here is that Linux could significantly benefit the business by doing small things very well, at a low cost. The question was: Could it take over critical operations in the enterprise system?

To me, the best place for Linux today is with the most "invisible" part of the business: the data. A database should do one thing very well: store data and provide easy and efficient access to it. It doesn't need fancy GUIs. It doesn't need wizards, graphs, reporting, and other things associated with client applications. The database is a storage engine (with a few twists). Linux on the desktop hasn't been successful so far for many reasons, which I'll address in my next article. But for database, application, Web, and mail servers? If configured correctly (on any operating system), they tend to run in lights-off mode most of the time, or they should.

One of the issues in the environment was that you had to reboot the Windows servers pretty regular, especially the database server. The database engine uses a lot of resources and was pushing the box to the limit. I felt that a Linux O/S would be a better database server than Windows could be; you have more flexibility in tuning Linux, and I perceive the Linux O/S to be more stable than Windows, after years of working with both environments, especially for a RDBMS.

While we were contemplating the shift, the Windows O/S did its best to help us make a decision. One night, I received a call at 3 a.m. from the network admin and was told that they couldn't boot any of their servers. A virus had managed to corrupt the ntoskernel.dll file (or something like that), and the O/S had to be recovered. (At least backups were complete....) Something went wrong on the recovery. By the time I arrived on site, I was told that the O/S had to be trashed and we would have to revert to backup. We lost about four days, due to wait time for hardware and O/S configuration. After that, the writing was on the wall - we were going Linux, wherever we could. As a matter of fact, we already had two Linux servers in the rack: our integration server and a server that was responsible for client communications (generated PDF documents and mailed it out).

Even before this happened I presented a greater Linux strategy to the customer. Here is a high level:

1.  Move the database servers to Linux.
This is the lowest risk, because the users aren't affected at all, except maybe we expected more uptime and better scalability. In effect, we didn't anticipate too much of a performance boost - moving to Linux on the same 32-bit hardware wouldn't make too much of an outright performance change, but we were expecting a small improvement.

2.  Move the Web server (IIS) to Apache or Tomcat.
Most Web servers in the world run Apache, and it gets rid of having to pay licenses for a commodity. Another thing to mention is that the customer's enterprise application runs a J2EE Webapp, and it was felt that we should standardize the corporate Website to something like JSP, which could be supported by more than one person and can run on multiple environments.

3.  Move the application server to Linux.
This should've been easy, but it wasn't. The early application developers used the PowerBuilder DataWindow in their J2EE app, and we weren't convinced that the move would be seamless. So we left this until last.

4.  Convert all remaining client/server apps to thin client, browser-based apps.
A browser-based app would mean that the end users could use any OS and browser they felt comfortable with. Also, it puts the business in a position to test out Desktop Linux, and do this at their own pace. Why would they want to? The most significant savings to be made out of a corporate Linux shift is at the desktop level for application users. Power users may still want to run Windows, but for the person who comes to work in the morning, switches on his PC, and fires up his e-mail client and the application he requires to do his work, he could use any operating system - Mac, Linux, Windows, Solaris.

Even better, you probably don't need the "enterprise" version of Linux at the desktop level, meaning that the O/S won't cost you a cent. Now, calculate this for an organization with 500 users. And remember to add up Office and any other Windows license, etc.

5.  Desktop Linux, where it makes sense.
More of the above. There are some good articles on the Web from various authors who point out that most Windows fans are really Office fans. Microsoft Outlook is the de facto standard for organizations because of the integrated collaboration. However, the largest percentage of employees in a standard-sized organization probably use about 15% of Office. It makes sense for these users to try out OpenOffice. The tactic here was to install OpenOffice on Windows, swap the mail client to something like Thunderbird, and do proper UAT to see how that goes.

6.  Mail servers.
Depending on the business and how the organization uses the Outlook Calendaring (if they use Outlook at all), this could be an easy or difficult shift. In this case, about 30% of the users in the organization uses Outlook with calendaring, so it's not practical yet. How do we do this? In this case, it doesn't really matter. Windows and Linux can co-exist pretty easily in the environment, and I would never advocate a "rip and replace" strategy. The best strategy we can think of now is to go for a CRM (the client needs and wants to implement CRM) that integrates collaboration. First choices for now: SugarCRM and possibly Compiere.

The customer was ready for phases one, two, and three. When we started strategizing the Linux shift, an interesting question came up, and it's one that comes up quite a lot now: While we're doing this move, how about investigating 64-bit architecture? Surely this will also make a massive difference? Our initial test showed that we would get a 10-15% performance increase by using our same hardware, but that's fairly insignificant. Sooner or later we would run into hardware limitations. The Linux shift would extend the use of the current hardware to about eight months, and this seemed to be a short-sighted strategy.

The customer asked what was needed for a "significant" performance improvement at the database level, and how can we ensure that our hardware lasts us for the next five years? The key thing about a database is that it's only as fast as the amount of I/O requests it can process. Generally, disk writes and reads are very expensive and slow I/O operations. To offset this, you throw RAM at the problem and increase the database cache so that it doesn't have to do as many direct disk reads and writes. There's a lot more to this, but that's the basic rule. This is especially true if you are sure that the database engine has been properly configured to use the machine resources efficiently, and that all of the queries thrown at the server are optimized.

More Stories By Rudi Leibbrandt

Rudi Leibbrandt works at Sybase South Africa managing the Sybase development toolset.

Comments (0)

Share your thoughts on this story.

Add your comment
You must be signed in to add a comment. Sign-in | Register

In accordance with our Comment Policy, we encourage comments that are on topic, relevant and to-the-point. We will remove comments that include profanity, personal attacks, racial slurs, threats of violence, or other inappropriate material that violates our Terms and Conditions, and will block users who make repeated violations. We ask all readers to expect diversity of opinion and to treat one another with dignity and respect.


@ThingsExpo Stories
With so much going on in this space you could be forgiven for thinking you were always working with yesterday’s technologies. So much change, so quickly. What do you do if you have to build a solution from the ground up that is expected to live in the field for at least 5-10 years? This is the challenge we faced when we looked to refresh our existing 10-year-old custom hardware stack to measure the fullness of trash cans and compactors.
The emerging Internet of Everything creates tremendous new opportunities for customer engagement and business model innovation. However, enterprises must overcome a number of critical challenges to bring these new solutions to market. In his session at @ThingsExpo, Michael Martin, CTO/CIO at nfrastructure, outlined these key challenges and recommended approaches for overcoming them to achieve speed and agility in the design, development and implementation of Internet of Everything solutions wi...
Cloud computing is being adopted in one form or another by 94% of enterprises today. Tens of billions of new devices are being connected to The Internet of Things. And Big Data is driving this bus. An exponential increase is expected in the amount of information being processed, managed, analyzed, and acted upon by enterprise IT. This amazing is not part of some distant future - it is happening today. One report shows a 650% increase in enterprise data by 2020. Other estimates are even higher....
Today we can collect lots and lots of performance data. We build beautiful dashboards and even have fancy query languages to access and transform the data. Still performance data is a secret language only a couple of people understand. The more business becomes digital the more stakeholders are interested in this data including how it relates to business. Some of these people have never used a monitoring tool before. They have a question on their mind like “How is my application doing” but no id...
The 19th International Cloud Expo has announced that its Call for Papers is open. Cloud Expo, to be held November 1-3, 2016, at the Santa Clara Convention Center in Santa Clara, CA, brings together Cloud Computing, Big Data, Internet of Things, DevOps, Digital Transformation, Microservices and WebRTC to one location. With cloud computing driving a higher percentage of enterprise IT budgets every year, it becomes increasingly important to plant your flag in this fast-expanding business opportuni...
Identity is in everything and customers are looking to their providers to ensure the security of their identities, transactions and data. With the increased reliance on cloud-based services, service providers must build security and trust into their offerings, adding value to customers and improving the user experience. Making identity, security and privacy easy for customers provides a unique advantage over the competition.
19th Cloud Expo, taking place November 1-3, 2016, at the Santa Clara Convention Center in Santa Clara, CA, will feature technical sessions from a rock star conference faculty and the leading industry players in the world. Cloud computing is now being embraced by a majority of enterprises of all sizes. Yesterday's debate about public vs. private has transformed into the reality of hybrid cloud: a recent survey shows that 74% of enterprises have a hybrid cloud strategy. Meanwhile, 94% of enterpri...
Smart Cities are here to stay, but for their promise to be delivered, the data they produce must not be put in new siloes. In his session at @ThingsExpo, Mathias Herberts, Co-founder and CTO of Cityzen Data, will deep dive into best practices that will ensure a successful smart city journey.
SYS-CON Events announced today that 910Telecom will exhibit at the 19th International Cloud Expo, which will take place on November 1–3, 2016, at the Santa Clara Convention Center in Santa Clara, CA. Housed in the classic Denver Gas & Electric Building, 910 15th St., 910Telecom is a carrier-neutral telecom hotel located in the heart of Denver. Adjacent to CenturyLink, AT&T, and Denver Main, 910Telecom offers connectivity to all major carriers, Internet service providers, Internet backbones and ...
There is growing need for data-driven applications and the need for digital platforms to build these apps. In his session at 19th Cloud Expo, Muddu Sudhakar, VP and GM of Security & IoT at Splunk, will cover different PaaS solutions and Big Data platforms that are available to build applications. In addition, AI and machine learning are creating new requirements that developers need in the building of next-gen apps. The next-generation digital platforms have some of the past platform needs a...
SYS-CON Events announced today Telecom Reseller has been named “Media Sponsor” of SYS-CON's 19th International Cloud Expo, which will take place on November 1–3, 2016, at the Santa Clara Convention Center in Santa Clara, CA. Telecom Reseller reports on Unified Communications, UCaaS, BPaaS for enterprise and SMBs. They report extensively on both customer premises based solutions such as IP-PBX as well as cloud based and hosted platforms.
I wanted to gather all of my Internet of Things (IOT) blogs into a single blog (that I could later use with my University of San Francisco (USF) Big Data “MBA” course). However as I started to pull these blogs together, I realized that my IOT discussion lacked a vision; it lacked an end point towards which an organization could drive their IOT envisioning, proof of value, app dev, data engineering and data science efforts. And I think that the IOT end point is really quite simple…
Internet of @ThingsExpo, taking place November 1-3, 2016, at the Santa Clara Convention Center in Santa Clara, CA, is co-located with 19th Cloud Expo and will feature technical sessions from a rock star conference faculty and the leading industry players in the world. The Internet of Things (IoT) is the most profound change in personal and enterprise IT since the creation of the Worldwide Web more than 20 years ago. All major researchers estimate there will be tens of billions devices - comp...
DevOps at Cloud Expo, taking place Nov 1-3, 2016, at the Santa Clara Convention Center in Santa Clara, CA, is co-located with 19th Cloud Expo and will feature technical sessions from a rock star conference faculty and the leading industry players in the world. The widespread success of cloud computing is driving the DevOps revolution in enterprise IT. Now as never before, development teams must communicate and collaborate in a dynamic, 24/7/365 environment. There is no time to wait for long dev...
Data is the fuel that drives the machine learning algorithmic engines and ultimately provides the business value. In his session at Cloud Expo, Ed Featherston, a director and senior enterprise architect at Collaborative Consulting, will discuss the key considerations around quality, volume, timeliness, and pedigree that must be dealt with in order to properly fuel that engine.
Pulzze Systems was happy to participate in such a premier event and thankful to be receiving the winning investment and global network support from G-Startup Worldwide. It is an exciting time for Pulzze to showcase the effectiveness of innovative technologies and enable them to make the world smarter and better. The reputable contest is held to identify promising startups around the globe that are assured to change the world through their innovative products and disruptive technologies. There w...
Personalization has long been the holy grail of marketing. Simply stated, communicate the most relevant offer to the right person and you will increase sales. To achieve this, you must understand the individual. Consequently, digital marketers developed many ways to gather and leverage customer information to deliver targeted experiences. In his session at @ThingsExpo, Lou Casal, Founder and Principal Consultant at Practicala, discussed how the Internet of Things (IoT) has accelerated our abil...
Is the ongoing quest for agility in the data center forcing you to evaluate how to be a part of infrastructure automation efforts? As organizations evolve toward bimodal IT operations, they are embracing new service delivery models and leveraging virtualization to increase infrastructure agility. Therefore, the network must evolve in parallel to become equally agile. Read this essential piece of Gartner research for recommendations on achieving greater agility.
SYS-CON Events announced today that Venafi, the Immune System for the Internet™ and the leading provider of Next Generation Trust Protection, will exhibit at @DevOpsSummit at 19th International Cloud Expo, which will take place on November 1–3, 2016, at the Santa Clara Convention Center in Santa Clara, CA. Venafi is the Immune System for the Internet™ that protects the foundation of all cybersecurity – cryptographic keys and digital certificates – so they can’t be misused by bad guys in attacks...
For basic one-to-one voice or video calling solutions, WebRTC has proven to be a very powerful technology. Although WebRTC’s core functionality is to provide secure, real-time p2p media streaming, leveraging native platform features and server-side components brings up new communication capabilities for web and native mobile applications, allowing for advanced multi-user use cases such as video broadcasting, conferencing, and media recording.