This topic covers how to analyze and address C/C++test’s test execution results.

Sections include:

Reviewing Reported Tasks

After test execution, C/C++test generates a prioritized task list organized by error categories and severities. For tests run in the GUI, tasks are organized into the following categories in the Quality Tasks view:

For tests run from the command line interface, tasks are reported in the Test Generation and Test Execution section of the report. Results imported to your IDE as described in the Importing Results into the UI will also appear in the Quality Tasks view.

Viewing Test Configurations that Trigger Tasks

You can view Test configurations that trigger tasks by right-clicking on a task in the Quality Tasks view and choosing View Test Configuration.

Quickly accessing test configuration from the from the Quality Tasks view is useful for group architects who are customizing tests and want to quickly disable settings that aren’t applicable. Developers importing results from a server-based run may also need to open and review test configurations that trigger tasks.

Reviewing Stack Traces

For each unit testing problem reported, C/C++test reports the stack trace for the test case that caused the problem.

To review one of the lines of code referenced in the stack trace, double-click the node that shows the line number, or right-click the node and choose Go to from the shortcut menu. The editor will then open and highlight the designated line of code.

Reviewing Postconditions

The results of postcondition macros (all the asserted values reported) are displayed in the Quality Tasks view. These postconditions capture the state of test objects or global variables used in the test.

Any test case with reported post conditions can automatically be validated for use in regression testing. Verification changes the *_POST_CONDITION_* macros into assertions that will fail if a subsequent test does not produce the expected (validated) value. This is especially useful for automated generation of a regression base for legacy code. For details on verification, see Verifying Test Cases for Regression Testing.

Understanding and Customizing Task Severity Categorization

Within each category, tasks are organized according to severity to help you identify and focus on the most serious issues.

If you want the Quality Tasks view to display the severity of each task, go to the pull-down menu in the Quality Tasks view, then choose Configure Contents and set it to show Severity

Reviewing Test Case Execution Details

The Test Case Explorer helps you manage a project’s test cases, test suites, and related data sources. It provides detailed test statistics (executed/passed/failed/skipped) and allows you to search/filter the test case tree.

By default, the Test Case Explorer is open in the left side of the UI. If it is not available, you can open it by choosing Parasoft> Show View> Test Case Explorer.

For details on the Test Case Explorer, see the Exploring the C++test UI.

Reviewing Test Case Details

To view test case details, enable the Show> Details option from the Test Case Explorer menu. 

The following information will be displayed in the Test Case Explorer tree:

Opening the Source Code for a Test or Test Suite

To open the source code for a test suite or test case in the Test Case Explorer, right-click its Test Case Explorer node, then choose Open. Or, double-click its Test Case Explorer node.


Correlating Test Case Explorer Nodes to Quality Tasks View Results

To open the test results (if available) for a test case in the Test Case Explorer, right-click its Test Case Explorer node, then choose Show in Quality Tasks. You can also perform the reverse action—to see the Test Case Explorer node that correlates to a result in the Quality Tasks view, right-click the Quality Tasks view node, then choose Show in Tests.

Reviewing Results of Data Source Test Cases

To view detailed information about executed data source test cases, enable Show > Data source tests from the Test Case Explorer menu. This will make C/C++test display information about each executed iteration of the data source test case.

The data source test case element of the tree presents statistics information about executed iterations (e.g., number of passed and failed iterations). Statistics shown for all parent nodes of the data source test case also include information about particular data source test case iterations.

Obtaining Traceability Details

There are several ways to create a report containing the maximum amount of unit test execution and traceability information:

From here...Perform these steps...
Generating test cases automatically
  1. Provide the test case description in the Test Configuration> Generation> General> Test case description field.
  2. Enable Test Configuration> Generation> Test case> Insert code to report test case inputs and outputs.
Creating test cases using Test Case Wizard
  1. Provide a test case description.
  2. Enable Insert code to report test case inputs and outputs.
Executing tests
  1. Enable Test Configuration> Execution> Runtime> Report unit test execution details.
  2. Enable Test Configuration> Execution> Runtime> Include tasks details.
  3. Enable Test Configuration> Execution> Runtime> Include passed assertions details.
Generating reports
  1. Enable Parasoft> Preferences> Reports > Overview of checked files and executed tests.
  2. Choose HTML (C++test's Unit Testing details) as the Report format.
Generating reports from the command line
  1. Set the report.contexts_details=true option in the localsettings file to enable the Executed Test Cases section.
  2. Set the report.format=html_ut_details option in the localsettings file to set up the report format.

Traceability Reporting Tips



Responding to Reported Tasks

Different types of findings require different response strategies. The following table lists the categories used to classify C/C++test’s test execution findings, and links to sections that will help you understand and respond to them.

CategorySubcategoryDescription and Recommended Response
Fix Unit Test ProblemsAssertion FailuresSee Assertion Failures and Timeouts
Runtime ExceptionsSee Runtime Exceptions
Review Unit Test OutcomesUnverified OutcomesSee Unverified Outcomes

Using Quick Fix (R) to Respond to Test Execution Findings

The Quick Fix (R) feature can be used to automate actions commonly performed while reviewing and responding to unit test findings.

To use Quick Fix to respond to a test execution finding:

  1. Locate the detailed task message (the one with the line number) in the Quality Tasks view.
  2. Right-click that task message, then choose the appropriate Quick Fix option from the shortcut menu. C/C++test will then perform the specified action.
  3. Save the modified file.

Understanding Available Quick Fix options

The following section explains the available Quick Fixes. Note that the Quick Fixes available for a specific task depends on the nature of the task.

Assertion Failure Quick Fixes

Unverified Outcome Quick Fixes

Handling a Large Number of Reported Tasks (e.g., After Testing Legacy Code)

If you want to apply the same action to multiple reported problems:

  1. Select the problems you want to apply the action to.
  2. Right-click the selection, then choose the command for the desired response (for example, Verify All Outcomes).

Generating Test Execution Details Reports

The Test Execution Details report is an additional report you can generate from your regular report. It contains details about the following information:

 See Generating the Test Execution Details Report for details.