You are viewing an old version of this page. View the current version.

Compare with Current View Page History

« Previous Version 8 Next »

"Test Environment" is a concept introduced in Xray v2.

Here is a short video that explains Test Environments and how you can use them.


What is a Test Environment

Generically speaking, a test environment is an environment that contains all necessary elements, including the SUT, so you can perform testing on it.

Depending on your context, a test environment may represent:

  • a testing stage (e.g. "development", "staging", "preproduction", "production")
  • a device model or device operating system (e.g. "Android", "iOS")
  • an operating system (e.g. "Windows", "macOS", "Linux")
  • browser (e.g. "Edge", "Chrome", "Firefox")

Thus, semantics of what a "Test Environment" represents depends on your specific context.


Please note

In Xray, Test Environments are focused on the execution aspect, providing the means to schedule tests and analyze their results in different environments.

Thus, they're explicitly associated with Test Execution issues.


Benefits of using Test Environments

  • avoid duplication of Tests, whenever you have to run the same test on different environments
  • ability to track the latest status of tests on different environments
  • ability to track coverage on each environment
  • ability to track overall coverage, considering the coverage/results on each environment
  • ability to perform reporting, including traceability, per each environment or globally (i.e. considering all results on all different environments)

Test Environments in Xray at a glance

The current status of a Test is generally calculated by looking at the Test's last execution (more detail here). However, this does not work well if you execute the same Test in two different test environments (e.g., devices) and you want the two results to be consolidated. 

Within a Test Execution, you may specify the Test Environment(s) where the tests will be executed in the respective attribute. A Test Environment is similar to a label, but Xray has special logic to deal with it.

If you use Test Environments, you may reuse the same Test for multiple test environments and create Test Executions for each one. If you don't use Test Environments and you want to track tests for multiple environments (e.g., devices), then the only way to do it is to create multiple tests, one per each test environment.

Let's say that you have executions for two Test Environments: "Android" and "iOS". The test is considered as PASS only if the latest executions for Android and iOS are both PASS; otherwise, it will be FAIL.


test_environments


Note: If you don't use Test Environments, then only the latest result matters for assessing the current status of the Test.

How it works

Let's say we have a Test being executed two times. We will start by creating Test Executions TE1 and TE2.

TE2 will be executed after TE1, so TE1 is considered older.

Below are example scenarios and shows how the overall, consolidated status of the Test is calculated in each case.

If you have a mix of Test Executions (i.e., with and without Test Environment as in scenario D), it will be treated the same way as scenarios A, B, and C.


ScenarioTest Environment(s) of TE 1Test Environment(s) of TE 2Test run status in TE1Test run status in TE2Calculated value for the overall, consolidated status of the Test (i.e. for the "All Environments")Other

A

AndroidiOSPASSPASSPASSThe test will be considered to be PASS in both Android and iOS environments.

B

iOSiOSPASSFAILFAIL

The test will be considered to be FAIL in iOS.

C

iOSiOSFAILPASSPASSThe test will be considered to be FAIL in iOS.

D

iOS-FAILPASSFAILThe test will be considered to be FAIL in iOS and PASS for the empty environment.

E

--PASSFAILFAILThe test will be considered to be FAIL for the empty environment.

F

--FAILPASSPASS The test will be considered to be PASS for the empty environment.

Please note

The empty Test Environment is treated similarly to any other environment having a well-defined name.


How to use

Whenever creating a Test Execution, you must set the Test Environment in which the execution will be executed. You can use this field as a simple label: just add the environment or reuse a previously created one.

Please see some important Tips and Recommendations ahead.


Creating a Test Execution


Test Execution for “android” Test Environment


 

Test Execution for “ios” Test Environment


Tracking the results on different environments

The Test Environments column is shown in your Test Runs table so you can distinguish each execution of the Test between the different environments.

This information can be seen in the Test issue screen (see next screenshot) or in other places that show a list of Test Runs (e.g. Test Plan issue screen).


The same test has been executed in both Test Environments (a Test Execution per Test Environment).


Analyzing the impact of the results on different environments

Results obtained for Test Environments will impact coverage.

Considering the previous screenshot, the "Requirement Status" custom field for the Test issue will show NOK because the Test has failed for one of the environments. This information is independent of the environment picker below within the "Test Coverage" section, which in turn is used to calculate the coverage on request for the selected scope, showing it on the right side along with the corresponding test results.

 


If you want to analyze the coverage for the requirement (e.g. "story") and show the latest results on that environment, just use the picker on the "Test Coverage" section. As seen ahead, this will produce different results because different results were obtained in different environments. 

     

Please check Coverage Analysis to learn more about coverage analysis possibilities.


It is also possible to analyze testing thoroughly considering Test Environments; this analysis can be done using the Traceability Report or the Overall Coverage Report, among others.

The exact behavior upon choosing a specific Test Environment depends on the report itself but, either explicitly or implicitly, Test Runs will be filtered by the selected Test Environment and reports will reflect it.


Traceability Report being used to analyze the results on the "edge" test environment.


Traceability Report being used to analyze the results on the "chrome" test environment.


Analyzing coverage of "requirements" on the "edge" test environment.


Analyzing coverage of "requirements" on the "chrome" test environment.


Using multiple environments at the same time

Sometimes, you may have multiple categorizations for a given environment; in theory, you can think as it being something multidimensional.

Consider a very basic example: whenever performing web/UI based testing you will be using a browser and an operating system and you may want to analyze the results per a browser perspective or per an operating system perspective.


The recommended way to deal with environments having multiple dimensions is to treat each dimension (e.g. browser name, operating system name) individually. In other words, add the values of each dimension to the "Test Environments" field separately.


Whenever you assign "mac" and "edge" to the Test Environments of a given Test Execution, it's equivalent to saying that your Test Run is scheduled for/was run in the "mac" and also in the "edge" environment.

This approach will limit the number of environments to the total number of possible values for each dimension, as opposed to having <number_of_values_dimension_1>*<number_of_values_dimension_2>*... environments.

The drawback of this solution is that you won't be able to analyze the results for an environment tagged as "mac" and "edge" at the same time, for example; you can just analyze results from a specific dimension.


To have in mind

One way to deal with these kinds of environments would be to flatten them and treat them as usual, i.e. you could name the environment such as “windows_edge” or “mac_chrome” but…

  • You could have a ton of composed environments which wouldn’t be manageable at all
  • You couldn’t analyze coverage just from the perspective of one of those variables (e.g. “mac” or “edge”); you would be restricted to analyze it from the perspective of the composed environment

How to use

Assign each environment (e.g. name of operation system, name of browser vendor) as you do for a single environment; in other words, just add the multiple environment names as multiple, distinct labels.


Whenever creating a Test Execution (e.g. from a Test Plan)


Whenever updating an existing Test Execution


Example

Test executed in the context of Test Execution assigned to several environments at the same time

  1. windows, chrome (fail)
  2. windows, edge (pass)
  3. mac, edge (pass)



The calculated status of the test, per Test environment, will be the following.


StatusWhy?
windowsPASSdue to the last result obtained in "windows" environment on CALC-5262
macPASSdue to the last result obtained in the "mac" environment on CALC-5262
edgeFAILdue to the last result obtained in "chrome" environment on CALC-5261
chromePASSdue to the last result obtained in "edge" environment on CALC-5263

"All Environments"

(if analyzing the status of the test without identifying a specific environment)

FAILas the last result for one of the environments ("chrome") was FAIL (i.e. on CALC-5261)


Advanced

Test Environments and the TestRunStatus custom field

The "TestRunStatus" custom field is associated with Test issues and can be used to provide information about the latest status of your test; more info here.

This custom field calculates the status of the test for "all environments" (i.e. the consolidated status), giving you a high-level view; it cannot be configured to show the status for a specific environment.

Internally, this field will store the status of the test for all possible scopes, which besides other things includes the information about the status in all different environments.


If you start using Test Environments in your Test Executions, then it's not only your test status calculation that will change (i.e. the one stored in the TestRunStatus custom field). All custom fields that depend on it (e.g., Requirement status, Test Sets status) will change. Consequently, the requirement coverage calculation and all associated charts/gadgets are also affected.

Tips and Recommendations

Do's

  • Use Test Environments only if you want to run the same Test case in different environments and track results per each environment. 
  • Simplify the names of Test Environments (i.e. lowercase it, shorten it)
    • Example: macOS => mac
  • Evaluate if you really need to assign multiple environments at the same time; using just one is preferable if you can afford that simplicity


Learn more

For advanced Test Environment management capabilities, please check our Integration with Apwide Golive.

Don'ts

  • Don't create hundreds or thousands of Test Environments as it will harden their usage and add some performance overhead
  • Don’t make composed environment names, such as ”<os>_<browser>_<stage>” as it will pollute the environment namespace and harden management
  • Don't try to do data-driven testing using Test Environments; they're not tailored for that


Please note

Besides other usage issues, if you have a large number of environments (>100), it will impact the calculations that need to be done and the size of the Lucene index.

Please try to have a limited, restricted, and well-defined list of Test Environments. 

  • No labels