System-level Delta quality

Delta quality is a part of the system maintainability analysis. As such it is default for all systems. It requires more than 1 snapshot to make a comparison. Delta quality tells you

You can reach this view in different ways: Via the top menu, or clicking an a capability on the System or Portfolio Overview pages. See the system-level Overview page or portfolio-level Overview page.

If you have a portfolio, you can assess these trends and changes on portfolio level

Delta quality overview

The Delta quality view shows the impact of code changes on the system for the selected period. By default it shows you “New code”, the first of three tabs.

Within Delta quality, there are three tabs available :

Interpreting new- and changed code quality on the Delta Quality page

New code tends to give you the best indication of whether new development is creating high quality software (assuming that code change heuristics picked up those files correctly as being new, as opposed to moved or renamed).

For Changed code, quality effects may be limited unless deliberate refactoring/renovation has been done.The impact of “Changed code quality” really depends on the size of the change. If 1 line of code has been changed, the code quality will probably stay the same. If that happens to be (older) code with lower quality, it is not necessarily alarming.

A more balanced view will be visible in the “New & changed code” tab, which averages out changes to new and existing code. But here again, the volume of changes and the difference in code quality determines ultimate impact on system rating. A sanity check is to see whether the system metrics have actually changed for the system as a whole.

Also see the Maintainability Overview section on the system maintainability documentation page.

Example: Delta Quality for changed code

As an example below is a partial screenshot for the “Changed code” tab. The layout and type of information is generally the same as for the two other tabs: the left tile shows the number of changed files, their total volume and average maintainability rating in stars. The Changed code tab shows an extra comparison. For each maintainability rating tile, next to the comparison of the system as a whole (whether, as a trend, code quality of existing code is improving), it also shows the difference between the maintainability rating of the file before and after the change. This gives an indication of whether, on average, code quality of modified existing code has been improved. Note that “on average” is significant.

In the tabs for “New code” and “New & changed code”, metrics are only compared to the system quality as it was before.

The filtering menu (here, for Changed files) shows next to the calculated metrics, the file path and Risk change.

You can achieve the same filtering by clicking on the respective column names.

The overview page here is focused on “Changed code”. You can also move to “New code” or “Changed & new code” in the tabs in the top.

The overview shows the different metrics and impact per file. It shows per file whether a certain code quality metric rating (such as “Duplication”) went up, down, or remained equal.

In the top right you can export the metrics for the visible filter to a spreadsheet (.xlsx) by clicking “Export to Excel”. It includes for each file the full file paths and exact rating change (up to 2 decimals). In this case, the export will only contain the changed files. A complete list can be exported if you select the “New & changed code” tab first, since the export adjusts to current viewing filters.

Note the Code Explorer icon next to each file. This will bring you to the Code Explorer tab. This is explained in more detail on the Code Explorer page.