Linux Containers Authors: Zakia Bouachraoui, Elizabeth White, Yeshim Deniz, Liz McMillan, Pat Romanski

Related Topics: @DevOpsSummit, Linux Containers, Containers Expo Blog

@DevOpsSummit: Blog Feed Post

How Testing Fits into DevOps, Because It’s Here to Stay By @JustinRohrman | @DevOpsSummit #DevOps

When DevOps first appeared on the scene, no one really knew what it meant

How Testing Fits into DevOps, Because It's Here to Stay
by Justin Rohrman

When DevOps first appeared on the scene, no one really knew what it meant. Books were defining the term in completely different way; conference speakers were sending out conflicting messages about tools that you absolutely must use (or not) to do "real" DevOps. I distinctly remember seeing a job advertisement or two that were hiring a DevOps person to "dev all the ops."

We all know better now.

Or, at least, some of us know (a little) better now. Having some time to experiment taught us that DevOps is a lot like "agile". It describes a set of methods and tools that help programmers deliver software faster. This isn't something one person does, it is part of daily life as a programmer.  There is still one big question in all of this. How do testers fit into development groups? How can testers continue to make a difference when software companies are driving further and further into technical practices?

Dev Ops

Teaching Developers to Test
Anyone can test software. Tell them to "play with it", or give them a spec. They'll even find a bug or two. If it's big and obvious, for example, login in broken, they'll probably find that too.

Even programmers can do this sort of testing. Programmer testing tends to be verification that what they thought they built works - it tends to miss out on the difference between the customer need and what the programmer understood, verifying only what the customer understood.  Because what they do is write low-level code, a programmer might create a code-test ‘test' before writing production code to see that some value is set, then write that production code, and then finally run the code and test in concert. This usually comes in the form of TDD, or BDD, or unit testing. These tools are a nice way to help a person check their work, but it often isn't enough. Anything surprising, that the programmer didn't guess ahead of time might be a problem is still a mystery. One way to go beyond this is to have the programmer test someone else's code.

Jesse Alford, who is on the technical staff at Pivotal Labs, did a talk at CAST2015 about how he spends his time at Pivotal teaching programmers to test software. Through a combination of pairing with programmers and then talking about the work, playing games with software testing themes built in (like Zen Do, and the infamous dice game), the programmers learn more about what skilled software testing looks like, and Jesse learns more about writing good code.

Pivotal has created a stronger team through this process of programmer / tester pairing, and teaching exercises.

What Can't Be Done in CI
Honestly, I think Continuous Integration is pretty cool. Why wouldn't you want to get some base feedback on the software for every single build? Why not build every few minutes to every few hours at the slowest?

Yet there are other quality questions that CI just can't answer, where delivering faster might not help.

Stability and Reliability
This is the question of how well your product runs over time. A quick 1-hour exercise of the software won't find a memory leak.  If something goes wrong, and eventually it will (this is software we're talking about) how does the product recover and can I continue using it without intervention from some sort of administrator?

Continuous Integration runs on the short term. Get the latest code from the repository, mash it up into something testable, run the automated checks, and spin things back down. These environments don't exist long enough to get a meaningful feel for how stable or reliable the product will be.

Most of us understand the idea behind performance testing - run the software with a great number of simultaneous users for an extended series of time and see if it slows down. Yet how to do performance testing, and what to make of the results, is an interesting combination of technical skill, mathematics, and social science. Single user performance testing can be as easy as sitting at a computer with a stopwatch to see how long a page takes to load or form to submit. More complicated versions include running a series of HTTP requests, measuring various aspects of the call, comparing that to previous measurements in different environments, and then trying to decide if the results matter.

The important thing to note here is that you are the most important thing in the equation. The performance tester needs to observe differences, then decide whether a 25 millisecond difference in one HTTP call between versions is important enough to do something about, or if the fact that one button click triggers 30 HTTP POSTS should be reported. The context is important, and the tester has it. The CI system never will.

The term ‘Usability' encompasses a great number of factors, including utility (can it do the job), usability (does it work for me) and identity (do I think of it as compatible with my sense of self.). Figuring out if software is usable, and how to improve it, is ‘soft' but it is still science. Ideas like affordance of devices and interviews, which both come from the ‘soft' science of anthropology, can help use answer these questions and improve our product. There is also the intuition of the user, which is even harder to understand and measure. This intuition can manifest when customers user the software and rub their foreheads trying to figure out what to do next, or become frustrated and ask someone for help, or even giving up all together. (When customers abandon a request for vacation, or submitting reimbursement, that system has problems.)

In some cases, usability studies and design are carefully handled ahead of time and then forgotten. More than once, I've worked on a product that was immaculately designed, but after performing one task many times, found it very tedious. That feeling of tediousness is a hint that something is going on, and it is not good.

No software is an island. Even very small programs like games that run on your phone or tablet have to play well with their environment and the other software running there. Business systems integrate with user accounts and often send and receive data with other pieces of software. Healthcare software is constantly sending patient information for health records and insurance information for billing.

Often the fastest way to learn if your software is sending the right medical billing codes in the right format for a patient to get insurance coverage is to create the scenario in your product and then send the output to your test system. That might be a 3rd party test system - when millions of dollars in involved, don't worry, they'll have one you can use. You might be able to do this every build, but by the time you built the file and the scripts, you could already have discovered what was broken and started fixing it.

How Continuous
Where Continuous Integrations will get every new line of code into a build and checked against the unit tests, Continuous Delivery (CD) takes it to the next level. CD takes the latest build and automatically deploys that to an environment along with whatever database and frameworks go along with that build. Some companies have pushed this concept to its logical conclusion and are pushing new code to production on every commit, something they call Continuous Deployment. These terms are used so interchangeably, and are so confusing, that I prefer "Continuous Delivery (to where)" - for example, Continuous Delivery to a staging server or continuous delivery to production. CD to production ("true" Continuous Deployment) takes a variety of engineering practices designed to enable partial features, turning features on and off, sending new features in "dark", database changes that run simultaneously so you can cut-back if needed, and more.

The first step is usually CD to staging, and then only if all the automated checks run green. Deploying every build automatically to a staging gives the benefits of fast visibility, but also protects users from big, unanticipated, black swan problems that will ruin their day. Deploying continuously to staging has the added benefit of allowing testers to control their own test environments.

One other strategy I've had work well is deploying automatically to test after getting a green light from suites of tests that cover multiple layers of the product - unit, service, and UI. Although these are just checks and usually won't show unexpected problems, they will show that certain aspects of the software still function the way you think they do. Having a second test environment to control and compare against the latest is a nice touch, too.

Not Everything Is Functional
One important aspect of DevOps is defining when a feature or code change is officially done. When a company releases quickly, and often, that definition can be as light as a green light from all automated checks run on a given build. This method treats software as a simple set of functions. I can enter a value in this text field and select one of these radio buttons then click a button and get a value out. With some higher (than unit) types of automated checking, we can string these functions together to get something a little more complicated.

That isn't enough though, and it certainly doesn't represent how people use software.

The main problem with relying on this type of testing is how simple and linear it is. When we use software, we don't take perfectly predefined and clean paths. Instead of performing a series of - submit 5, assert value, select check box, assert value, check for NULL, assert value - testers take a loosely guided path. We meander here and there looking for hints of something interesting and then strike when a clue shows itself. This kind of activity can happen all the time, both on a macro level ("what new features could use all more attention on staging, or even in production, right now?") and at the micro level, exploring a story just a little bit more before the code goes live. That micro-exploring work can even happen with continuous delivery to production, by turning the feature on in staging and "off" in production until the tester has completed an exploration run.

DevOps tends to treat testing as an activity to be completely automated. Over time, as DevOps begins to gain maturity, I see that changing. Human, thinking, in the moment testing might be different each time, and needs to be done by someone, a tester, or developer or someone else, while the things that run every time according to algorithm, the checking, that might be automated. Cutting out the exploring causes us to loose perspective in a way that was probably unanticipated.

The push toward DevOps can be scary for testers; it isn't hard to imagine that the methods and tools in the wrong hands could squeeze our special role out of development groups. The best way to stay relevant is by understanding your unique contribution, being able to explain it, and excelling at it.

So keep calm and Excel On

Read the original blog entry...

More Stories By SmartBear Blog

As the leader in software quality tools for the connected world, SmartBear supports more than two million software professionals and over 25,000 organizations in 90 countries that use its products to build and deliver the world’s greatest applications. With today’s applications deploying on mobile, Web, desktop, Internet of Things (IoT) or even embedded computing platforms, the connected nature of these applications through public and private APIs presents a unique set of challenges for developers, testers and operations teams. SmartBear's software quality tools assist with code review, functional and load testing, API readiness as well as performance monitoring of these modern applications.

IoT & Smart Cities Stories
Andrew Keys is Co-Founder of ConsenSys Enterprise. He comes to ConsenSys Enterprise with capital markets, technology and entrepreneurial experience. Previously, he worked for UBS investment bank in equities analysis. Later, he was responsible for the creation and distribution of life settlement products to hedge funds and investment banks. After, he co-founded a revenue cycle management company where he learned about Bitcoin and eventually Ethereal. Andrew's role at ConsenSys Enterprise is a mul...
In his general session at 19th Cloud Expo, Manish Dixit, VP of Product and Engineering at Dice, discussed how Dice leverages data insights and tools to help both tech professionals and recruiters better understand how skills relate to each other and which skills are in high demand using interactive visualizations and salary indicator tools to maximize earning potential. Manish Dixit is VP of Product and Engineering at Dice. As the leader of the Product, Engineering and Data Sciences team at D...
Dynatrace is an application performance management software company with products for the information technology departments and digital business owners of medium and large businesses. Building the Future of Monitoring with Artificial Intelligence. Today we can collect lots and lots of performance data. We build beautiful dashboards and even have fancy query languages to access and transform the data. Still performance data is a secret language only a couple of people understand. The more busine...
Nicolas Fierro is CEO of MIMIR Blockchain Solutions. He is a programmer, technologist, and operations dev who has worked with Ethereum and blockchain since 2014. His knowledge in blockchain dates to when he performed dev ops services to the Ethereum Foundation as one the privileged few developers to work with the original core team in Switzerland.
René Bostic is the Technical VP of the IBM Cloud Unit in North America. Enjoying her career with IBM during the modern millennial technological era, she is an expert in cloud computing, DevOps and emerging cloud technologies such as Blockchain. Her strengths and core competencies include a proven record of accomplishments in consensus building at all levels to assess, plan, and implement enterprise and cloud computing solutions. René is a member of the Society of Women Engineers (SWE) and a m...
Whenever a new technology hits the high points of hype, everyone starts talking about it like it will solve all their business problems. Blockchain is one of those technologies. According to Gartner's latest report on the hype cycle of emerging technologies, blockchain has just passed the peak of their hype cycle curve. If you read the news articles about it, one would think it has taken over the technology world. No disruptive technology is without its challenges and potential impediments t...
If a machine can invent, does this mean the end of the patent system as we know it? The patent system, both in the US and Europe, allows companies to protect their inventions and helps foster innovation. However, Artificial Intelligence (AI) could be set to disrupt the patent system as we know it. This talk will examine how AI may change the patent landscape in the years to come. Furthermore, ways in which companies can best protect their AI related inventions will be examined from both a US and...
Bill Schmarzo, Tech Chair of "Big Data | Analytics" of upcoming CloudEXPO | DXWorldEXPO New York (November 12-13, 2018, New York City) today announced the outline and schedule of the track. "The track has been designed in experience/degree order," said Schmarzo. "So, that folks who attend the entire track can leave the conference with some of the skills necessary to get their work done when they get back to their offices. It actually ties back to some work that I'm doing at the University of San...
When talking IoT we often focus on the devices, the sensors, the hardware itself. The new smart appliances, the new smart or self-driving cars (which are amalgamations of many ‘things'). When we are looking at the world of IoT, we should take a step back, look at the big picture. What value are these devices providing. IoT is not about the devices, its about the data consumed and generated. The devices are tools, mechanisms, conduits. This paper discusses the considerations when dealing with the...
Bill Schmarzo, author of "Big Data: Understanding How Data Powers Big Business" and "Big Data MBA: Driving Business Strategies with Data Science," is responsible for setting the strategy and defining the Big Data service offerings and capabilities for EMC Global Services Big Data Practice. As the CTO for the Big Data Practice, he is responsible for working with organizations to help them identify where and how to start their big data journeys. He's written several white papers, is an avid blogge...