|By Steve Weisfeldt||
|July 13, 2012 09:05 AM EDT||
Where do I start? What am I doing? (Why am I even here?) Have you ever asked yourself these questions at the beginning of a performance testing project? Have you been one of the many test engineers who have been delivered an application and told to "just test it"? (True story: I was once handed off an application to lead a performance testing project on with these words of guidance, "Just test it to make sure it doesn't suck"). If you are one of the fortunate performance test engineers who have a well-defined process in place in your organization that dictates what are you testing, what you are looking for, etc., then this discussion might not be for you. But for the rest of us, those of us that might need a little guidance in starting our project, read on......
What Are We Testing? What Are We Testing For?
I don't believe there is any one right way to approach a load testing project. And while there may be several different methodologies, I think ultimately they all take us in the direction of identifying a similar starting point. And while my discussion here certainly will not include every detail one should consider, my goal is to provide you with some ideas and practices to get you on your way.
Before we can begin any kind of testing project, we need a game plan. We need to know what we will test and what determines a successful test. We need to determine whether the application we're testing is "good" or "bad". It's common to think that only functional tests have pass or fail conditions but load tests should usually also have criteria defined to dictate whether the application's performance passes or fails.
Essentially what we're talking about here is coming up with the best approach to develop a load/performance testing Test Plan...or at least the "goals" part of a test plan. (A good test plan should include a lot of additional information - schedules, roles of team members, tools to be used, etc. - that I won't discuss here). The key is determining how we're supposed to approach this application-to-test that was just presented to us.
This game plan should be comprised of three pieces. One, what is the activity that the end users of the application do? (What are our use-cases?) Two, how many of these users are there doing whatever they do? (What is our workload, e.g. concurrency and/or transaction rate?) And finally, how fast do we need the application to allow these users to whatever they do? (What are our performance requirements)? Let's take each one of these at a time.
Identify Use Cases
Now that I have this new application in my hands, how can I determine what a user is supposed to do with it? The best approach to answer this question is usually simply to ask someone. Ah, but who to ask? Ideally, a good start would be to find someone involved in developing product requirements. At some point someone had a thought, "We should build an application so that users can do something". Identifying who that person was and what that "something" is, is the key.
I typically like to start with someone in product marketing or product management - someone who has done the work to create the user scenarios that the development team has implemented. If you're lucky, you might be able to find a Marketing (or Product) Requirements Document that has this information documented. Otherwise, a conversation with someone in that department might prove very useful.
If this information is not documented anywhere (gasp!) and the application is one that is currently in production, there will probably be some useful information in the log files captured in the application's infrastructure. Ask your web or application server administrative team if this is information they capture. They may use tools like Omniture, Coremetrics, Web Trends or Google Analytics that can help provide this. If they don't, it's usually possible to turn on some sort of logging for a period of time (a day or two perhaps) so that tracing user activity can be captured. In addition, getting access to actual end-users could be very helpful. As a performance engineer, it's probably not feasible that you have contact with the user community but this is a conversation you might be able to start with another internal resource - customer or technical support teams, sales representative, or (again) marketing folks.
Finally, common sense is an approach that can be very useful. (As obvious as this may seem, I'm constantly surprised by the number of times "common sense" is overlooked). Take a look at the application itself and decide what it is that YOU might do if you were an end-user. If you're testing an online store type of application, it's a pretty good bet that users are going to browse through the product catalog, add items to a shopping cart and make a purchase or two. Online banking customers are probably checking their balance and paying bills... you get the idea.
Generally speaking, you should not waste time trying to identify every single use case through the application since the bulk of user traffic will be captured in only a few transactions. Keep in mind the famous "80/20 rule"; that is 20% of the transactions cover 80% of application's core functionality.
So now you've got a start on identifying the first piece of the load test project puzzle; the use cases to be modeled in your load testing processes and/or load testing tool. Next we need to identify the actual workload that will be used in our load testing - that is, how many simultaneous users that will be sending network traffic against the application we're testing.
The best process to find out this information follows that of identifying the use-cases that we discussed earlier - again, try talking to someone "in the know". The application's specification documents should detail the expected user load or expected transaction rates. If they don't and you are to test an application which is already in production, working with the administrators of the applications architecture can again be helpful. They should be able to give you a sense of what sort of user load the application has experienced. When looking at this data, be sure to use a wide enough data sample to find the peak load the application has seen.
If this is an application that is not in use yet, many of the same principles we discussed in determining use cases apply; talk to marketing people, determine what the expected loads are. Even if you are conducting tests due some expected increased load in the application; a sale or promotion, an expected seasonal rush, rolling out an internal application to a new set of employees, etc., the hope is the expectation of load has been documented somewhere.
You may also be able to determine the number of users to test for by using a formula known as "Little's Law". This formula is only really applicable if you have either determined baseline throughput (transactions per second) and transaction response times or if you have desired requirements for both. But if you do have this information, using this formula is a great way to figure out how many users to test for:
- Number of Users = Response Time * Transactions/second
There are many articles on the web about Little's Law and the different ways it can be applied. If you have access to the "inputs" of this formula, I'd recommend doing some extra research and seeing how it might be of use to you.
Generally speaking, once I determine the peak usage load my application will handle (or hope to handle) in production, I will conduct load tests scaling to 110% of that number.
Establish Performance Requirements
So now we know how to find what use cases to model and how many concurrent users to test for. Now what? Surely we can start to execute some load and performance tests, no? Well yes we can. However, it seems that we still don't know what we are testing for. I.e. why are we even conducting this project?
This is where developing good requirements come in. As I mentioned earlier, we need to figure out just what we're looking for as the results of our testing. To me, test requirements fall into two high-level buckets: Baseline/Benchmark or Performance Requirements.
A Baseline or Benchmark test doesn't usually have any detailed pass or fail criteria. Instead it allows you to provide results that state, "under these circumstances, the application performed like this". Benchmark tests can of course be very useful for examining the impact of infrastructure changes. Conduct an initial baseline test, modify the infrastructure, run the exact test again and compare the "before" and "after" results. The results of your initial baseline test can always serve as a performance benchmark.
Performance requirements on the other hand can help us determined whether the application is ready to be delivered to production.
What just is a performance requirement? As I've indicated, this essentially is the indicator of whether your performance test passed or failed. It may be response time-based (e.g. the Login transaction must always be faster than 3 seconds), it may be load-based (e.g. the website must be able to withstand 1000 concurrent users, how many concurrent users will my site handle before it crashes), it may be transaction rate based (e.g. the application must be able to handle 100 orders per minute) or it may be combinations of all of the above. Requirements may also include details on how the back-end infrastructure behaves (e.g. CPU utilization on the database server may never exceed 50%). If the end-goal of our testing is to make a "go" or "no go" decision on system, the more detailed our performance requirements are, the better.
How can I determine performance requirements? Maybe not surprisingly, the source(s) for this information should be very much the same as what we discussed for determining use cases and user concurrency. The various stakeholders (product management, marketing, business analysts, etc.) hopefully have this information documented somewhere. There may be contractual Service Level Agreements in place between your company and a customer, or between teams within your organization. Hunt them down (both the other teams and the SLAs). Or this information might just be stuffed in the back recesses of some stakeholder's brain. As you might have guessed by now, one of the things I hope to have achieved with this article is facilitated communications between the performance test teams and other groups in the organization - something I notice isn't happening in far too many companies.
And similar to what we've discussed earlier, if performance requirements are not easily available and you're hoping to test beyond a baseline, use common sense. What do YOU think is an acceptable wait time before a web page returns?
There are many studies you can find on the internet which discuss acceptable wait times but I'm not sure these are always valid. Plus they seem to change too quickly. It wasn't too long ago that an 8-second response time was the "cut-off" point between good and bad response time. Waiting 8 seconds now seems excessively long... except perhaps if you are using a mobile device. But my point is that it's all subjective and you should determine requirements specific to YOUR application and YOUR user community.
I may not have provided all the answers to the "where do I start?" question here. But hopefully this discussion has given you some clues as to the information you should gather before you begin your load testing project. Ideally you will now have a better sense of where to get some of this information and what some of the thought processes are to fill in any missing pieces on your own.
Today we can collect lots and lots of performance data. We build beautiful dashboards and even have fancy query languages to access and transform the data. Still performance data is a secret language only a couple of people understand. The more business becomes digital the more stakeholders are interested in this data including how it relates to business. Some of these people have never used a monitoring tool before. They have a question on their mind like “How is my application doing” but no id...
Apr. 30, 2017 07:45 PM EDT Reads: 7,479
Is your application too difficult to manage? Do changes take dozens of developers hundreds of hours to execute, and frequently result in downtime across all your site’s functions? It sounds like you have a monolith! A monolith is one of the three main software architectures that define most applications. Whether you’ve intentionally set out to create a monolith or not, it’s worth at least weighing the pros and cons of the different architectural approaches and deciding which one makes the most s...
Apr. 30, 2017 07:30 PM EDT Reads: 3,079
Developers want to create better apps faster. Static clouds are giving way to scalable systems, with dynamic resource allocation and application monitoring. You won't hear that chant from users on any picket line, but helping developers to create better apps faster is the mission of Lee Atchison, principal cloud architect and advocate at New Relic Inc., based in San Francisco. His singular job is to understand and drive the industry in the areas of cloud architecture, microservices, scalability ...
Apr. 30, 2017 07:00 PM EDT Reads: 3,769
When you decide to launch a startup company, business advisors, counselors, bankers and armchair know-it-alls will tell you that the first thing you need to do is get funding. While there is some validity to that boilerplate piece of wisdom, the availability of and need for startup funding has gone through a dramatic transformation over the past decade, and the next few years will see even more of a shift. A perfect storm of events is causing this seismic shift. On the macroeconomic side this ...
Apr. 30, 2017 06:30 PM EDT Reads: 731
Cloud promises the agility required by today’s digital businesses. As organizations adopt cloud based infrastructures and services, their IT resources become increasingly dynamic and hybrid in nature. Managing these require modern IT operations and tools. In his session at 20th Cloud Expo, Raj Sundaram, Senior Principal Product Manager at CA Technologies, will discuss how to modernize your IT operations in order to proactively manage your hybrid cloud and IT environments. He will be sharing be...
Apr. 30, 2017 01:00 PM EDT Reads: 1,105
Cloud Expo, Inc. has announced today that Aruna Ravichandran, vice president of DevOps Product and Solutions Marketing at CA Technologies, has been named co-conference chair of DevOps at Cloud Expo 2017. The @DevOpsSummit at Cloud Expo New York will take place on June 6-8, 2017, at the Javits Center in New York City, New York, and @DevOpsSummit at Cloud Expo Silicon Valley will take place Oct. 31-Nov. 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA.
Apr. 30, 2017 11:30 AM EDT Reads: 2,945
A Man in the Middle attack, or MITM, is a situation wherein a malicious entity can read/write data that is being transmitted between two or more systems (in most cases, between you and the website that you are surfing). MITMs are common in China, thanks to the “Great Cannon.” The “Great Cannon” is slightly different from the “The Great Firewall.” The firewall monitors web traffic moving in and out of China and blocks prohibited content. The Great Cannon, on the other hand, acts as a man in the...
Apr. 30, 2017 07:30 AM EDT Reads: 583
To more closely examine the variety of ways in which IT departments around the world are integrating cloud services, and the effect hybrid IT has had on their organizations and IT job roles, SolarWinds recently released the SolarWinds IT Trends Report 2017: Portrait of a Hybrid Organization. This annual study consists of survey-based research that explores significant trends, developments, and movements related to and directly affecting IT and IT professionals.
Apr. 30, 2017 05:30 AM EDT Reads: 1,966
NHK, Japan Broadcasting, will feature the upcoming @ThingsExpo Silicon Valley in a special 'Internet of Things' and smart technology documentary that will be filmed on the expo floor between November 3 to 5, 2015, in Santa Clara. NHK is the sole public TV network in Japan equivalent to the BBC in the UK and the largest in Asia with many award-winning science and technology programs. Japanese TV is producing a documentary about IoT and Smart technology and will be covering @ThingsExpo Silicon Val...
Apr. 30, 2017 03:45 AM EDT Reads: 9,497
In his general session at 19th Cloud Expo, Manish Dixit, VP of Product and Engineering at Dice, discussed how Dice leverages data insights and tools to help both tech professionals and recruiters better understand how skills relate to each other and which skills are in high demand using interactive visualizations and salary indicator tools to maximize earning potential. Manish Dixit is VP of Product and Engineering at Dice. As the leader of the Product, Engineering and Data Sciences team at D...
Apr. 30, 2017 03:30 AM EDT Reads: 6,286
Keeping pace with advancements in software delivery processes and tooling is taxing even for the most proficient organizations. Point tools, platforms, open source and the increasing adoption of private and public cloud services requires strong engineering rigor – all in the face of developer demands to use the tools of choice. As Agile has settled in as a mainstream practice, now DevOps has emerged as the next wave to improve software delivery speed and output. To make DevOps work, organization...
Apr. 30, 2017 02:45 AM EDT Reads: 9,297
Enterprise architects are increasingly adopting multi-cloud strategies as they seek to utilize existing data center assets, leverage the advantages of cloud computing and avoid cloud vendor lock-in. This requires a globally aware traffic management strategy that can monitor infrastructure health across data centers and end-user experience globally, while responding to control changes and system specification at the speed of today’s DevOps teams. In his session at 20th Cloud Expo, Josh Gray, Chie...
Apr. 29, 2017 10:45 AM EDT Reads: 3,632
This recent research on cloud computing from the Register delves a little deeper than many of the "We're all adopting cloud!" surveys we've seen. They found that meaningful cloud adoption and the idea of the cloud-first enterprise are still not reality for many businesses. The Register's stats also show a more gradual cloud deployment trend over the past five years, not any sort of explosion. One important takeaway is that coherence across internal and external clouds is essential for IT right n...
Apr. 28, 2017 05:00 PM EDT Reads: 2,008
Back in February of 2017, Andrew Clay Schafer of Pivotal tweeted the following: “seriously tho, the whole software industry is stuck on deployment when we desperately need architecture and telemetry.” Intrigue in a 140 characters. For me, I hear Andrew saying, “we’re jumping to step 5 before we’ve successfully completed steps 1-4.”
Apr. 28, 2017 02:00 PM EDT Reads: 1,994
In his session at 20th Cloud Expo, Scott Davis, CTO of Embotics, will discuss how automation can provide the dynamic management required to cost-effectively deliver microservices and container solutions at scale. He will discuss how flexible automation is the key to effectively bridging and seamlessly coordinating both IT and developer needs for component orchestration across disparate clouds – an increasingly important requirement at today’s multi-cloud enterprise.
Apr. 28, 2017 06:00 AM EDT Reads: 4,492
In large enterprises, environment provisioning and server provisioning account for a significant portion of the operations team's time. This often leaves users frustrated while they wait for these services. For instance, server provisioning can take several days and sometimes even weeks. At the same time, digital transformation means the need for server and environment provisioning is constantly growing. Organizations are adopting agile methodologies and software teams are increasing the speed ...
Apr. 27, 2017 08:30 PM EDT Reads: 3,540
Software as a service (SaaS), one of the earliest and most successful cloud services, has reached mainstream status. According to Cisco, by 2019 more than four-fifths (83 percent) of all data center traffic will be based in the cloud, up from 65 percent today. The majority of this traffic will be applications. Businesses of all sizes are adopting a variety of SaaS-based services – everything from collaboration tools to mission-critical commerce-oriented applications. The rise in SaaS usage has m...
Apr. 22, 2017 06:15 PM EDT Reads: 5,112
The proper isolation of resources is essential for multi-tenant environments. The traditional approach to isolate resources is, however, rather heavyweight. In his session at 18th Cloud Expo, Igor Drobiazko, co-founder of elastic.io, drew upon his own experience with operating a Docker container-based infrastructure on a large scale and present a lightweight solution for resource isolation using microservices. He also discussed the implementation of microservices in data and application integrat...
Apr. 22, 2017 05:45 AM EDT Reads: 6,553
We'd all like to fulfill that "find a job you love and you'll never work a day in your life" cliché. But in reality, every job (even if it's our dream job) comes with its downsides. For you, the constant fight against shadow IT might get on your last nerves. For your developer coworkers, infrastructure management is the roadblock that stands in the way of focusing on coding. As you watch more and more applications and processes move to the cloud, technology is coming to developers' rescue-most r...
Apr. 22, 2017 04:00 AM EDT Reads: 4,285
2016 has been an amazing year for Docker and the container industry. We had 3 major releases of Docker engine this year , and tremendous increase in usage. The community has been following along and contributing amazing Docker resources to help you learn and get hands-on experience. Here’s some of the top read and viewed content for the year. Of course releases are always really popular, particularly when they fit requests we had from the community.
Apr. 22, 2017 03:45 AM EDT Reads: 3,848