You are viewing an old version of this page. View the current version.

Compare with Current View Page History

« Previous Version 26 Next »



Xray's roadmap is continuously reviewed and redefined. We update often, depending on the feedback we receive from our clients and internal stakeholders.

Our release plan is available in our Jira issue tracker. Feel free to view and to vote on the issues that you would like to see implemented (account registration required).

Here you can find a list of features that define our main goals for future releases. This doesn't mean that other, potentially smaller features, won't be implemented as well.

Shipped


V1.10


The Test Steps component was the target of a major facelift with the goal of providing a wider, more usable UI that will have the ability to edit all step-related fields at once, in a grid or column-based layout.

Editing, reading and navigating through Test steps are now much easier with this new UI.

Along with the Test steps UI revamp, Xray now also provides the ability to configure and specify new custom step fields for manual test cases that can complement the standard ones (Action/Step, Data, Expected Result). The standard custom fields can also be hidden if desired. All of this can be configured in the project settings.

V1.12


You can now define additional custom fields for Test Runs. These fields can be useful to add extra information to Test Runs usually only available during or after executing Tests.

Test Run custom fields can be configured by project and by Test Type. Therefore, these settings will not affect other projects within your Jira instance. For example, it is possible to have custom fields just for Manual Tests within a project.


Reporting

The Test Runs List report and gadget can already display Test Run Custom Field values for each Test Run.

Within a Test Execution issue, it is also possible to display Test Run Custom Field columns and to filter Test Runs by Test Run Custom Field values.


Learn more about Test Run Custom Fields here.


V2.0


Test parameterization is a powerful practice that allows the same test to be executed multiple times with different parameters. Parameters are similar to input values (variables) that can change with each execution.

Parameterized tests in Xray are defined just like any other test with the addition of some parameter names within the specification using the following notation: ${PARAMETER_NAME}. This notation is used to reference parameters within the test steps.

Precondition issues can also be parameterized by including parameter names in the precondition specification.

The parameters, along with their values, are defined within a dataset. A dataset is a collection of data represented with a tabular view where every column of the table represents a particular variable (or parameter), and each row corresponds to a given record (or iteration) of the dataset. The number of rows in the dataset determines the number of iterations to execute.

A dataset can be defined in the following entities/scopes:

  1. Test (default dataset)
  2. Test Plan - Test
  3. Test Execution - Test (Test Run)

The closest dataset to the test run will be the one used to generate the iterations, effectively overriding any dataset defined in higher levels.



All iterations for a given test are executed within the context of the same test run. Each iteration can be expanded, and the steps executed individually. The step parameters will be replaced by the corresponding iteration values. The steps affect the iteration status that, in turn, affects the overall test run status.


Learn more about parameterized tests here.

In the works


Ability to call an existing Test from the step of another Test and thus use tests as reusable building blocks for the composition of more complex testing scenarios.

The called test can be executed independently or be used as a step of a broader test (e.g. "login as the admin user").


Wish list for future versions


The goal behind the testing board is to provide a centralized hub for accessing Xray entities and activities in order to improve navigation and discoverability. For now, this board features the Test repository, the Test Plan board, and project reports. 

The testing board will be improved to include other Xray entities such as Preconditions, Test Sets, and Test Executions. The Test Run and execution screen will have the testing board in context so that users can navigate to any other test activity.

The Automated Steps Library will be a project-level steps library for BDD Tests/Preconditions, containing all the Gherkin steps referred by these issues.

Thus, it provides a facility to store and manage all steps, encouraging users to easily reuse steps across Tests/Preconditions with step autocompletion and discoverability when editing BDD scenarios and backgrounds with Xray.

This feature will allow test engineers to create or generate environments with variables such as Browser, Operating System, Database, etc. These environments (or configurations) can be managed and assigned to Test Plans and Test Plan folders. Test Executions can then be created automatically for each environment/configuration.

Xray will provide a report to track the progress of Test Runs by environment in the context of a specific Test Plan.

The Xray Connector app for Bamboo will be updated in order to support Xray cloud APIs and connectivity.

This feature will cover a common scenario where manual Test cases evolve to automated Tests. In this case, the only option is to create separate Test issues to cover all the natures for the Test.

With Test natures, multiple Test definitions can coexist within the same Test case. Users will also be able to set the current definition so that Test Runs are always created using the "current" definition.

Currently, the Test Plan is composed of a static list of Test cases. This means you must explicitly add the Tests to the Test Plan. If the Tests are all known and well defined when you start your Test Plan, this is ok. However, if you are working in an agile context where a Test Plan is created for a specific sprint, Tests will only be specified during the sprint and later added to the Test Plan. This process is not ideal because users might forget to add the Tests to the Test Plan.

Dynamic Test Plans can be defined with a JQL query which will be the source for Test cases. Considering the use case described above, we can define a JQL query that will get all Tests covering requirements of a specific sprint.

This report will feature a list of Test Execution issues along with metrics. Possible usage scenarios:

  • analyze both the progress of the Test Execution  and the success rate (i.e., the % of Tests contributing to the requirement's OK status)
  • see the number of manual Tests vs. others in the Test Execution
  • see the overall execution status (i.e., the current status of the Test Runs)
  • see the number of opened/closed linked defects, in the context of the Test Execution

Refer to the Test Executions Report on Xray server.

Right now, a Test Run contains the start and finish date which are set automatically, although users can also change the start date manually. However, users can not stop and start the execution as needed and the total time spent on the Test Run (finish - start) might not reflect the time spent by the user executing the Test.

We plan to provide a feature to allow users to pause and continue with the execution directly on the Test Run screen. 

This report will show a daily historical view of Requirements coverage. Users will be able to analyze the evolution of the Requirement's coverage status over time for a particular analysis scope: Latest, Version or Test Plan (for each Environment). This way they can estimate if they are going to keep the planed released date according to the coverage status trend.








  • No labels