Skip to content

Automated End-to-End Tests

All data flows within Ozone and the actions performed on its HIS apps are rigorously validated using a comprehensive suite of automated end-to-end (E2E) tests.

This section is designed to help Ozone developers familiarize themselves with ozone-e2e, the repository where all end-to-end tests for Ozone are maintained. It will guide you through the architecture of a typical end-to-end test case and provide the necessary knowledge and practical steps to develop your own test cases.

Some facts about ozone-e2e

  • We use Playwright , the automation testing framework.
  • We test our data flows through the actions and effects that they have on Ozone HIS apps.
  • We focus on how these actions and effects are experienced by end users.

Setting up ozone-e2e

Clone the Ozone E2E Repository: Execute the command

git clone https://github.com/ozone-his/ozone-e2e
Navigate to the Project Directory: Change into the directory with
cd ozone-e2e
Install Dependencies: Install the required packages
npm install

Project structure

The actual E2E tests live in the e2e subfolder that is organised as follows:

e2e
 ├── tests
 |   ^ Contains test cases.
 └── utils
     ^ Contains utilities needed to set up and tear down tests,
       as well as methods required by the tests to run.

The data flows resulting from the integration between specific app pairs (e.g., between Odoo and OpenMRS) are typically tested through a designated test file for each pair (e.g., odoo-openmrs-flows.spec.js). All test files are located in the e2e/tests directory. We encourage you to explore and review the current set of tests to understand the testing process better and evaluate the extent of their coverage.

Project configuration

Our E2E test suite is designed to interact with three distinct Ozone environments, allowing seamless switching as needed:

Dev Hosts the latest, bleeding-edge versions of Ozone, updated continuously as the development team commits new code to the Ozone repositories.
QA Used for staging Ozone prior to releases, facilitating final rounds of testing to ensure quality before launch.
Demo Features the latest stable release of Ozone, readily accessible online for demonstration purposes and advertised through the Ozone website.

Two important configuration variables govern the high-level behavior of our E2E test suite:

  1. TEST_ENVIRONMENT: This variable specifies which one of the three environments — dev, qa, or demo — should be targeted for testing.
    E.g.:

    TEST_ENVIRONMENT=dev
    

  2. TEST_PRO: This toggle, set to true for testing Ozone Pro or false for Ozone FOSS, determines the flavour of Ozone HIS under test.
    E.g.:

    TEST_PRO=false
    

All configuration variables are set to the desired values by editing and saving the .env file prior to running the tests.

Additionally, the .env file contains a range of configuration variables that specify the URLs for accessing the various HIS apps in each test environment, as well as the credentials needed to execute the test cases.

Running tests against a local instance

When developing on Ozone, you will likely run a local instance of the HIS and want to execute the E2E tests against it. To do this, update the URL coordinates for the dev environment profile in the .env file.

Typically, you may set the URLs in your local version of the .env file like this:

# OpenMRS
O3_URL_DEV=http://localhost
O3_URL_QA=
O3_URL_DEMO=

# ...

# Odoo
ODOO_URL_DEV=http://localhost:8069
ODOO_URL_QA=
ODOO_URL_DEMO=

# SENAITE
SENAITE_URL_DEV=http://localhost:8081/senaite
SENAITE_URL_QA=
SENAITE_URL_DEMO=
# ...
Continue this pattern for all apps running locally by identifying their local URL in your development environment, and edit the .env file accordingly.

Running Tests

Default test execution

To run all tests in the test suite, use the default command:

npx playwright test

Running specific tests

To focus on testing data flows between specific pairs of HIS apps, you can run tests contained in a specific file. For example, to test the integration data flows between Odoo and OpenMRS:

npx playwright test odoo-openmrs-flows

Selecting your browser

Playwright supports various configurations for running tests across different browsers. Specify the browser using the --project flag. For instance, to run tests in Chromium:

npx playwright test --project=chromium

Interpreting test results

  • Passed: When all tests within the suite pass without encountering any errors, the overall result is marked as "Passed". This means that the application behaved as expected under the test conditions.

  • Failed: If any test within the suite encounters an error or assertion failure during execution, the overall result is marked as "Failed". Determine whether the failure is specific to the test case, a particular app or feature, or the entire application. Once you've identified the root cause of the failure, implement a fix to address the issue. This may involve modifying test assertions, updating application code, or addressing environmental dependencies.

  • Skipped: Sometimes, tests are intentionally skipped based on certain conditions or configurations. These skipped tests are not executed during the test run and are marked as "Skipped" in the test result.

Developing Test Cases

Naming test files

Our test cases cover the data flows between Ozone HIS apps and we have taken the convention to group test cases by pairs of apps. For example all data flows between OpenMRS and SENAITE are grouped together in a file named openmrs-senaite-flows.spec.js that lives in the e2e/tests. This file contains all test cases for data flows going from OpenMRS to SENAITE as well as those going from SENAITE to OpenMRS.

The naming is alphabetical per convention. The data flows going both ways between hypothetical HIS apps Foo and Bar would live in a file named bar-foo-flows.spec.js and so on and so forth for all pairs of HIS apps.

Testing actions and effects

While Ozone relies on its apps default behaviour and features, we do encompass in our E2E test suite the actions performed when using the user experience of the apps and the effects observed throught the user experience of those apps. This ensures that test are truly end-to-end with a focus on the end-user experience.

To write a test case:

  • Identify the user interactions and functionalities to test between a pair of HIS apps.
  • Use Playwright’s API to script the actions within each test scenario, such as navigating pages, clicking buttons, filling forms, and verifying UI elements.

Utilizing Playwright

Leverage Playwright’s comprehensive API to engage with the browser:

  • Navigate to pages.
  • Interact with page elements.
  • Execute actions.
  • Assert expected behaviors using methods like expect, toEqual, toContain, etc.

For detailed usage, refer to Playwright's documentation .

Once you've written your test, use the appropriate test runner to check that the test works as expected. For instance for the OpenMRS-SENAITE flows written in openmrs-senaite-flows.spec.js, run:

npx playwright test openmrs-senaite-flows

Analyzing a sample test case

Again, following our example of the data flows between OpenMRS and SENAITE, let us verify that ordering a lab test for an OpenMRS patient does create the corresponding SENAITE client with an analysis request. Let us look at the sample test case code below:

import { test, expect } from '@playwright/test';
import { OpenMRS, patientName } from '../utils/functions/openmrs';
import { SENAITE } from '../utils/functions/senaite';

let openmrs: OpenMRS;
let senaite: SENAITE;

test.beforeEach(async ({ page }) => {
  openmrs = new OpenMRS(page);
  senaite = new SENAITE(page);

  await openmrs.login();
  await expect(page).toHaveURL(/.*home/);
  await openmrs.createPatient();
  await openmrs.startPatientVisit();
});

test('Ordering a lab test for an OpenMRS patient creates the corresponding SENAITE client with an analysis request.', async ({ page }) => {

  // replay
  await openmrs.goToLabOrderForm();
  await page.click('button:has-text("Add")');
  await page.selectOption('#tab select', '857AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA');
  await openmrs.saveLabOrder();

  // verify
  await senaite.open();
  await expect(page).toHaveURL(/.*senaite/);
  await senaite.searchClient();
  const clientName = `${patientName.firstName} ${patientName.givenName}`;
  const client = await page.$('table tbody tr:nth-child(1) td.contentcell.title div span a:has-text("' + clientName + '")');
  await expect(client).toBeVisible();

});

test.afterEach(async ({ page }) => {
  await openmrs.deletePatient();
  await page.close();
});

We observe that the test structure is broken down between a setup, the actual test case and a cleanup:

Test Setup: Before the actual test, we perform some preliminary actions: logging into Ozone (with SSO), creating a new patient, and starting a visit for the newly created patient.

Test Case: The core of each test case follows the Given-When-Then pattern (organised here as Setup-Replay-Verification). We highly recommend this structured approach as it clearly delineates the setup of the test ("Setup"), the end-user actions performed ("Replay"), and the assertion of outcomes as experienced by the end-user ("Verification"). In our example:

  • Setup: Omitted here. All aspects of the setup have been performed in the beforeEach() method.
  • Replay: Navigation to the lab order form, add a lab test, and save the form.
  • Verification: Navigation to the SENAITE HIS app and search for the client by name. Verify that the client's name is visible in the clients list.

Cleanup: The post-test cleanup consists of deleting the test patient and closing the browser page.