Testing levels

Testing priority triangle

This diagram demonstrates the relative priority of each test type we use. e2e stands for end-to-end.

As of 2019-05-01, we have the following distribution of tests per level:

Test level Community Edition Enterprise Edition Community + Enterprise Edition
Black-box tests at the system level (aka end-to-end or QA tests) 68 (0.14%) 31 (0.2%) 99 (0.17%)
White-box tests at the system level (aka system or feature tests) 5,471 (11.9%) 969 (7.4%) 6440 (10.9%)
Integration tests 8,333 (18.2%) 2,244 (17.2%) 10,577 (17.9%)
Unit tests 32,031 (69.7%) 9,778 (75.1%) 41,809 (71%)

Unit tests

Formal definition: https://en.wikipedia.org/wiki/Unit_testing

These kind of tests ensure that a single unit of code (a method) works as expected (given an input, it has a predictable output). These tests should be isolated as much as possible. For example, model methods that don't do anything with the database shouldn't need a DB record. Classes that don't need database records should use stubs/doubles as much as possible.

Code path Tests path Testing engine Notes
app/assets/javascripts/ spec/javascripts/, spec/frontend/ Karma & Jest More details in the Frontend Testing guide section.
app/finders/ spec/finders/ RSpec
app/graphql/ spec/graphql/ RSpec
app/helpers/ spec/helpers/ RSpec
app/models/ spec/models/ RSpec
app/policies/ spec/policies/ RSpec
app/presenters/ spec/presenters/ RSpec
app/serializers/ spec/serializers/ RSpec
app/services/ spec/services/ RSpec
app/uploaders/ spec/uploaders/ RSpec
app/validators/ spec/validators/ RSpec
app/views/ spec/views/ RSpec
app/workers/ spec/workers/ RSpec
bin/ spec/bin/ RSpec
config/ spec/config/ RSpec
config/initializers/ spec/initializers/ RSpec
config/routes.rb, config/routes/ spec/routing/ RSpec
config/puma.example.development.rb, config/unicorn.rb.example spec/rack_servers/ RSpec
db/ spec/db/ RSpec
db/{post_,}migrate/ spec/migrations/ RSpec More details in the Testing Rails migrations guide.
Gemfile spec/dependencies/, spec/sidekiq/ RSpec
lib/ spec/lib/ RSpec
lib/tasks/ spec/tasks/ RSpec
rubocop/ spec/rubocop/ RSpec
spec/factories spec/factories_spec.rb RSpec

Integration tests

Formal definition: https://en.wikipedia.org/wiki/Integration_testing

These kind of tests ensure that individual parts of the application work well together, without the overhead of the actual app environment (i.e. the browser). These tests should assert at the request/response level: status code, headers, body. They're useful to test permissions, redirections, what view is rendered etc.

Code path Tests path Testing engine Notes
app/controllers/ spec/controllers/ RSpec For N+1 tests, use request specs
app/mailers/ spec/mailers/ RSpec
lib/api/ spec/requests/api/ RSpec
app/assets/javascripts/ spec/javascripts/, spec/frontend/ Karma & Jest More details in the Frontend Testing guide section.

About controller tests

In an ideal world, controllers should be thin. However, when this is not the case, it's acceptable to write a system/feature test without JavaScript instead of a controller test. The reason is that testing a fat controller usually involves a lot of stubbing, things like:

controller.instance_variable_set(:@user, user)

and use methods which are deprecated in Rails 5 (#23768).

About Karma

As you may have noticed, Karma is both in the Unit tests and the Integration tests category. That's because Karma is a tool that provides an environment to run JavaScript tests, so you can either run unit tests (e.g. test a single JavaScript method), or integration tests (e.g. test a component that is composed of multiple components).

White-box tests at the system level (formerly known as System / Feature tests)

Formal definitions:

These kind of tests ensure the GitLab Rails application (i.e. gitlab-foss/gitlab) works as expected from a browser point of view.

Note that:

These tests should only be used when:

For instance, to test the breadcrumbs on a given page, writing a system test makes sense since it's a small component, which cannot be tested at the unit or controller level.

Only test the happy path, but make sure to add a test case for any regression that couldn't have been caught at lower levels with better tests (i.e. if a regression is found, regression tests should be added at the lowest-level possible).

Tests path Testing engine Notes
spec/features/ Capybara + RSpec If your test has the :js metadata, the browser driver will be Poltergeist, otherwise it's using RackTest.

Consider not writing a system test

If we're confident that the low-level components work well (and we should be if we have enough Unit & Integration tests), we shouldn't need to duplicate their thorough testing at the System test level.

It's very easy to add tests, but a lot harder to remove or improve tests, so one should take care of not introducing too many (slow and duplicated) tests.

The reasons why we should follow these best practices are as follows:

Black-box tests at the system level, aka end-to-end tests

Formal definitions:

GitLab consists of multiple pieces such as GitLab Shell, GitLab Workhorse, Gitaly, GitLab Pages, GitLab Runner, and GitLab Rails. All theses pieces are configured and packaged by GitLab Omnibus.

The QA framework and instance-level scenarios are part of GitLab Rails so that they're always in-sync with the codebase (especially the views).

Note that:

Every new feature should come with a test plan.

Tests path Testing engine Notes
qa/qa/specs/features/ Capybara + RSpec + Custom QA framework Tests should be placed under their corresponding Product category

See end-to-end tests for more information.

Note that qa/spec contains unit tests of the QA framework itself, not to be confused with the application's unit tests or end-to-end tests.

Smoke tests

Smoke tests are quick tests that may be run at any time (especially after the pre-deployment migrations).

These tests run against the UI and ensure that basic functionality is working.

See Smoke Tests for more information.

GitLab QA orchestrator

GitLab QA orchestrator is a tool that allows to test that all these pieces integrate well together by building a Docker image for a given version of GitLab Rails and running end-to-end tests (i.e. using Capybara) against it.

Learn more in the GitLab QA orchestrator README.

EE-specific tests

EE-specific tests follows the same organization, but under the ee/spec folder.

How to test at the correct level?

As many things in life, deciding what to test at each level of testing is a trade-off:

Another way to see it is to think about the "cost of tests", this is well explained in this article and the basic idea is that the cost of a test includes:

There are cases where the behaviour you are testing is not worth the time spent running the full application, for example, if you are testing styling, animation, edge cases or small actions that don't involve the backend, you should write an integration test using Jasmine.


Return to Testing documentation