Building Your Test Automation Checklist

Your testing team should decide which tests to automate as early as possible.  This information can be used to develop more accurate testing deadlines and better organize your test suite, and even more importantly it can help determine which test automation solution(s) your team adopts.  If your team needs help deciding – “we know a guy”. Certain types of tests are ripe for automated testing, like performance tests, while others are obviously not suitable such as single use “one-off” tests.  There is of course a gray area in-between where pros-and-cons need to be weighed, like testing anything dynamically generated or a semi-frequently changing user interface.

Before meeting with your team to wrangle over these key decisions, skim our Test Automation Checklist below:




  • Older, repetitive tests that have crystal clear pass/fails for verification and/or validation. A good example would be a test case that tries various username and password combinations for a web page login.

  • Data-driven tests and/or any test that requires lengthy, in-depth preparation. Things like configuring and cleaning up large volumes of test data, or the submissions of lengthy forms with various combinations of input data.  Let your test automation solution take the brunt of the repetitive configuration management workload.

  • Sanity and smoke tests, especially any typical paths that your users take through the system under test. For example, an e-commerce website should make doubly sure that users have no issues perusing the sections of its top three best-selling products as this is both a critical and common path for buyers.

  • Time-consuming tests that involve lengthy and repetitive user interactions. Not only is it expensive to tie up a human tester with long, monotonous tests like this, but over time the manual tester becomes glassy-eyed and bored from mindlessly clicking – and in-turn the overall quality of testing drops.

  • Performance tests, as the very nature of determining speed, reliability, scalability, and resource usage is what automated tests are perfectly designed to tackle. The process of finding system benchmarks demands the consistency and precision afforded by automated testing.

READ MORE  Peak Season Roadmap




  • Tests that require instant feedback beyond the depth and scale provided via report from your test automation solution.

  • Test cases that cover brand-new, rarely-used, and/or currently unstable system features.

  • Outlier, corner test cases.

  • Exploratory tests; excluding any repetitive configuration management like data preparation or teardown.

  • Tests that will change often, as the savings reaped from automating is more than negated by overly frequent maintenance.

  • Tests with subjective validation that have more of a “look and feel” check as opposed to a black-and-white pass/fail.

  • Tests that require verification and/or validation of something physical (shipping label, 3D printed model, etc…) that can’t be digitally verified.

The bottom line is automated testing works seamlessly with a variety of tests, such as regression testing. While some tests dpn’t yeild benefits that  others are obviously not suitable such as single use “one-off” tests. We hope our Test Automation Checklist helps break down which ones are and which ones aren’t.

Does your testing team go through a templatized formal process, or perhaps use a checklist like the one above, to determine whether a test case is automated?  Contact us and we will be happy to help!

This post was written by:

James Prior
Technical Pre-Sales Consultant

James has been working in software pre-sales and implementation since 2000, and has more recently settled into focusing on technical pre-sales.