/ #unit-testing 

Unit testing best practices

As part of the onboarding process of every new developer in Tripsta, I am required to deliver a training on unit testing basics. I would like to share the most valuable techniques, methodologies and best practices that I have collected over the years. I believe that they constitute the pillars of a solid foundation that every developer should have.

It may be a bit unconventional (the way I love it), but before we even start, let me clarify the two most important things that tests are not.

Tests are not nice to have

Tests are not negotiable

When I hear phrases like “If we have time, we’ll write some tests too” or “It costs 13 story points, but since there is no time, we can make it 8 if we don’t write tests” I want to quit my job and sell greengrocery for a living. Tests should be an integral part of writing software.

I could use a lot of lines talking about the necessity of writing unit tests, the benefits they provide and how they should be an integral part of the software development process instead of just a verification layer (hint: TDD), but that is another blog post on its own. So, let’s delve into a handful of best practices which I have come to value over the years.

Best practices

Naming

Some people are bad at naming. Some other (such as me) are terrible at it. In any case, taking the time to come up with a proper name for everything in our code is worth it. Regardless of how much time we need. Uncle Bob (Robert Martin) likes to say “with code, we do more reading than writing, by a huge factor. Which of those operations should be made efficient?” Naming matters!! If I were to give you a single piece of advice on naming, it would be the following:

Write the tests you’d want to read

Before stop writing a test and moving to the production code, ask yourself these two questions:

  • Does the test clearly express its behavior?
  • Is it easy to understand?

Usually, a developer that struggles to come up with a proper name will settle for a mediocre one, that does not express behavior properly, or even a bad one, which is hard to understand and misleads its reader.

Tests should be a documentation of the module’s behavior?

Whenever we come across a piece of code that we did not write, we want to be able to read its test suite and figure out what this piece of code does. So, take the time (however much you need) to name your tests (and the variables and functions in them) as best as you can and when you come across a mediocre or a bad name, change it to something better (leaving the code cleaner than you found it, as Uncle Bob would say).

Also, keep in mind that things change and, especially in the world of computer science, they do so rapidly. There is nothing wrong in starting off with a technique and shifting to a new one when you identify benefits to it. For instance, my team and I worked on a greenfield project almost a year ago and initially adopted the following naming convention:

public class AccountTest {

  @Test
  public void account_shouldStoreADepositTransaction_whenGivenOne() throws Exception {
    // Implementation
  }
}

This is a very common naming convention. However, after attending a Sandro Mancuso workshop, I was introduced to the following, far better approach, the benefits of which I presented to my team and we shifted to it.

public class AccountShould {

  @Test public void
  storeADepositTransaction() throws Exception {
    // Implementation
  }
}

So, my advice is to be restless. Scan our (rich) community over the Internet and do not be afraid to shift to new practices and techniques so long as you recognize the value they offer.

Isolation

This whole point I want to make on isolation can be summarized in the following sentence:

Tests should leave the system in the state that they found it

The last thing that we would like to face is 10 failing tests as a result of a single failure that dragged the following 9 tests with it. When we see x tests failing, we should know that there are exactly x problems with the system. Whenever a test is failing, it should clearly declare a single aspect of the software that is not functioning properly and point us to it immediately (by its name and the failure).

In order to achieve this, we should ensure that our test, even when it fails, will restore the system to the exact situation it found it and it won’t leave it dirty. Each test should not know and care if another one run before it (let alone if it failed or not).

Watch the test fail

When should we stop writing the test and start writing the production code? That is a surprisingly hard to answer question for a lot of people when I give the training. Take a minute and think about it before reading further on. There are a couple of questions I like asking myself before I stop writing the test.

  • Does it fail for the right reason? What good is a test if it does not break when the functionality breaks? How can we know that this won’t happen if we do not watch the test fail and make sure that it fails for the right reason?

  • Are the diagnostics clear? There will come a time in the future that this test will break and then someone (most probably us) will have to understand what is wrong and fix it. How hard will it be to fix it if we cannot tell what is wrong? A few months ago, I fell into exactly this pitfall, a test that I had written (quite a few months back) failed against the production code that I had written and all I had in the console was:

Expected: 1
Got: 0

I forced myself in a debugging session to figure out what was wrong. If only I had taken the time to make the diagnostics clear enough…

Triple A

When it comes to designing our unit tests, I strongly recommend that the backbone of the structure should be based on Arrange - Act - Assert (aka AAA). This is a very simple yet solid technique, consisting of the following three steps:

  • Arrange: Initially, we should prepare the system for the upcoming test, making sure that all preconditions are met.

  • Act: Then we should trigger the part of the code that we wish to test.

  • Assert: Finally, we should evaluate whether the code meets our expectations or not. This is the point in which a flag should be raised, declaring the outcome of the test and this flag should either be green or red (no yellow or orange results - more on this on the single assertion section).

Let’s consider the following example (I like using a blank line between the three steps, to make the code more readable):

@Test
public void validatorShouldThrowExceptionWhenBookingIDIsNull() throws Exception {
  Booking booking = BookingBuilder.aBooking().withBookingId(null).build();

  Executable validateBooking = () -> bookingValidator.validate(booking);

  assertThrows(ValidationException.class, validateBooking);
}

We want to test that a Validator class will throw an exception when validating a booking with null booking ID. In order to achieve this, we need a booking, which has a null booking ID. Given that we have it, we should execute the validate(booking) method, which is the part of the code we wish to verify its functionality. Finally, we want to have a way to know if the code behaved in the expected way (threw a ValidationException), so we write an assertion to expect exactly this.

Some of the highlight benefits I find in this approach are the following:

  • It promotes a clear separation among the setup, execute and verification steps.
  • A reader who is familiar with this technique can understand the test fast.
  • Code smells in the test become easier to spot (e.g. a series of act-assert steps - more on this on the single assertion section)
  • Discourages overengineering a test by focusing on the the absolutely needed steps to write it.

Single assertion

How many assertions should a test have? The answer should always be one. What message is a red bar sending to us? That something is wrong in the system. Which is our reflex move? To find out what is this. This is exactly why we want our tests to be shouting out loud the reason for their failure. If a test fails, we should know, just by reading its name, what the problem is. How can this happen if a test fails for multiple reasons though?

Unit tests should fail for one and only one reason

Watch out for a common misconception though. The following test, contains two assertions. Does it violate the single assertion rule?

@Test public void
calculatorShouldGeneratePositiveEvenIntegers() {
  int number = calculator.generate();

  assertThat(number > 0).isTrue();
  assertThat(number % 2).isEqualTo(0);
}

There may be two physical assertions, but they essentially constitute a single logical assertion. This test will fail if and only if the calculator does not generate positive even numbers, as opposed to the following test:

@Test public void
calculatorShouldDivideNumbers() {
  int number = calculator.divide(2, 2);
  assertThat(number).isEqualTo(1);

  Executable divideByZero = () -> calculator.divide(4, 0);
  assertThrows(ArithmeticException.class, divideByZero);
}

This test will both fail when calculator fails to divide correctly and when the error case of dividing by zero does not throw an ArithmeticException. Therefore, in essence, we want to avoid a series of act-assert sequences. A single act followed by more than one physical assertions is fine.

Assert first

Assert first is a technique described in Kent Beck’s excellent book Test Driven Development: By Example. As counterintuitive as it may sound (and feel for someone not used to it), it comes with tons of advantages and it will become second nature given some time. The main idea lies in the fact that writing a test presents multiple problems. The two principal ones are “what is the right answer?” and “how am I going to check it”. Taking this bottom up approach helps us focus on the target and avoid overengineering the test.

Coming back to the arrange - act - assert example, assuming that I have a business requirement that a booking without booking ID is invalid, and I am writing the booking validator logic, I may start as follows:

@Test public void
validatorShouldThrowInvalidBookingExceptionWhenBookingIdIsNull() {
  assertThrows(InvalidBookingException.class, validateBooking);
}

Starting off with the assert, I know both the right answer and how to check it. I am also driven to the proper act for my assertion.

@Test public void
validatorShouldThrowInvalidBookingExceptionWhenBookingIdIsNull() {
  Executable validateBooking = () -> validator.validate(booking);

  assertThrows(InvalidBookingException.class, validateBooking);
}

Now there is only one step missing, the arrange part. I need a booking and it should have a null booking ID.

@Test public void
validatorShouldThrowInvalidBookingExceptionWhenBookingIdIsNull() {
  Booking booking = BookingBuilder.aBooking().withBookingId(null).build();

  Executable validateBooking = () -> validator.validate(booking);

  assertThrows(InvalidBookingException.class, validateBooking);
}

Writing a test in this bottom - up fashion seems strange initially, I know. My suggestion would be to give it a shot, stick with it passed the uncomfortable denial phase (the “I want to write my tests like I’m used to” phase), and it will pay off.

Conclusion

I tried to share some knowledge and some techniques I use when writing unit tests, because I feel that the area is misunderstood. These are the most important topics I wanted to touch, but the list is far from exhaustive.

Whichever the techniques we might use, let’s always keep in mind that tests are an integral part of the system and therefore they should be designed and maintained. We should care for them and we should take the time to write the right tests in the right way.

P.S. I would also love to cover the whole “test state vs test behavior” topic, but this post is already lengthy enough and this is a really long topic