Overview
Penalty analysis is a great way to identify potential directions for the improvement of products to product developers, for example, by looking for the sensory attributes that may be associated with decreased overall liking (acceptability) of a product. A test compatible with the Penalty analysis includes a
Category Liking question to measure overall liking, and at least one
Category Just About Right (JAR) question to collect data about specific sensory attributes.
Penalty analysis measures the change in product liking due to the product having “too much” or “too little” of an attribute. For example, for a dark chocolate bar, you could measure the bitterness of the chocolate against the overall liking of the chocolate bar.
For consumer testing we typically recommend at least 60 results; however the Penalty analysis will run on a lower N and can still provide valuable information.
Considerations
- It is crucial to include a Liking and JAR question types in the test to be able to run the report.
If it happens that the question types are accidentally set to a different variety of Category Scale questions, even after the data is collected, the question types can be changed to a proper type. Please do not delete results to make this change!
Changing the question type is done in the Build tab, in the Category Question Options, as seen in the image below.

- Penalty Analysis reports are generated as Excel Workbooks.
- Penalty analysis is available only for tests without reps.
- In tests with sections, the report can be generated within individual sections. The report is not available across sections in a single test nor across multiple tests.
- Only complete sample sets will be included in the report. It is not an option to include sample sets that are not complete.
Test Setup
Generate the Report
- Go to the Results area of your test. Filter results if that is required before generating the report.
- Click the Reports tab and select Create report.

- In the 1. Select report type list, select Penalty analysis.

- Click 2. Select options and specify the Mean Drop, Net Penalty, and Penalty Inclusion thresholds you wish to use in the report.
The thresholds that are set in Compusense can be used as a starting point. Whether you increase or decrease the thresholds will be a business decision.
You are able to see the attributes that are trending towards the threshold and can make a business decision from here.

- Mean Drop Thresholds: Represented with red lines in the Mean Drop Chart sheet.
- The Mean Drop threshold indicates how different the mean liking of the consumers who selected JAR for a particular attribute is from the mean liking of consumers who did not select JAR for that attribute.
- The percentage threshold is the total percentage of consumers who selected anything other than JAR.
- Net Penalty Colour Thresholds: If selected, these will clearly indicate which attributes are of interest in the Net Penalty Graphs. A net penalty of <0.25 is considered low impact. Potential and high impact thresholds can be selected by the analyst.
- The Penalty Inclusion Threshold: It allows the analyst to set a penalty analysis percentage threshold. If the sum of the percentages for Too Much and Not Enough for a specific attribute is below the selected penalty analysis percentage, that attribute will be excluded from the Mean Drop Chart.
- Click 3. Select questions.
- Select the Overall Liking attribute. Only one liking attribute can be selected.
If no Liking attributes are listed but you believe they should be, you can review the question setup. It may have happened that a wrong Category question type was selected during the setup. Follow these steps:
In the top left-hand corner, click the test name.

Go to the Build tab, and in the Question options of a Category question that was asking panelists for their overall
liking, check whether it is set as a Category- Liking.
If it is not, change the question type into the Category- Liking type, as
described in the Considerations area on this page.
If unable to make the change, your test may have been set to Complete. If it works for you to do that, go to the Run test tab and undo complete.

Results will not be deleted unless you click the Delete results button. Return to the Build tab as described above and change the question type.

Return back to
the Results area to generate the report. Select the Overall liking attribute.

- Scroll down if necessary to select the JAR attributes that you wish to include in the report.

- In the 4. Select export type, click Create my report.

Report Details
Here we will review each sheet that can be included in the report. The sheet availability in your report can depend on your selections before generating the report, as described in the previous section.
Data
This sheet lists the raw data for all included samples and all included attributes. Information included in this sheet is used for calculations in all other sheets.
Penalty Tables
The penalty table enables you to look at how consumers’ perceptions of the attributes affect liking. The information found in this sheet is used for the graphs on the Mean Drop Chart sheets.
Each sample included in the analysis has its own table in this sheet (you will need to scroll down to see them all).

- Attribute: Three rows for each attribute, one for each level on the collapsed JAR scale. In our example screenshot, the "chocolate JAR" attribute repeats three times because there are three levels (see the next item below) in the attribute.
- Level: While typically a JAR scale would have 5 categories, in this analysis the top two and bottom two options are collapsed; therefore, only three levels are displayed and used in calculations. In our example: Too Weak, Just-about-right, and Too Strong.
- Frequencies: The count of responses for each level within an attribute.
- %: The frequencies converted into percentages.
The formula: (Frequency/N)*100
- Sum Liking: The sum of the liking responses within the attribute and level. In our example, the "chocolate JAR" attribute for the Too Weak level, indicates the sum liking of 25. This is determined by looking at the Data sheet. In the screenshot below, highlighted in yellow, we can see the 4 responses for the sample 1 that were less than a JAR (value lower than 3) for the "chocolate JAR" attribute. For those 4 responses we then look at the Overall Liking scores and add them up: 4 + 6 + 7 + 8 = 25.

- Mean Liking: The mean liking of the sample by consumers who selected the indicated level for the given attribute.
The formula: Sum Liking/Frequency
- Mean Drop: A comparison of the mean liking of the level above or below JAR with the mean liking of the group who selected JAR. Therefore the results are displayed only in the 'too much' and 'not enough' rows; never in the JAR row.
The larger the mean drop, the more that attribute/level affects the liking of the sample.
- Label: This text will appear in the mean drop graphs.
Mean Drop Charts
You will see a mean drop chart sheet for each sample in your study, as well as one sheet for all of your samples together.
If you have selected Penalty Inclusion when generating the report, and if the sum of the percentages for Too much and Not enough for a specific attribute is below the selected penalty inclusion percentage, the attribute will be excluded from the Mean Drop Charts.
If all attributes for a sample are below the penalty inclusion percentage threshold, that whole sample will be excluded from the Mean Drop Charts. That would mean that at the set threshold, that sample's overall liking is not affected by any of the JAR attributes included in the report.
- x axis: This is the frequency percentage (%) of the responses for each level of a specific attribute. See the % column in the Penalty Tables sheet.
- y axis: This is the mean drop value. See the Mean Drop column in the Penalty Tables sheet.
- Colour (these can be edited in Excel using its built in functionality): Each sample is assigned its own colour.
- Shapes (these can be edited in Excel using its built in functionality):
- Levels above JAR are represented by a triangle.
- Levels below JAR are represented by a circle.
- Quadrants: If you selected Mean Drop thresholds, they will be represented as two red lines on the graphs. This creates easy to read quadrants in your graph.
- Attribute levels found in the bottom left quadrant represent attributes levels that are below both predetermined thresholds. These attribute levels are of no concern.
- Attribute levels found in the top right quadrant represent attribute levels that are above both predetermined thresholds. This is the area of primary concern because these attribute levels indicate that they affected panelists overall liking of the sample.
- Attribute levels found in the top left quadrant represent attribute levels that are above the mean drop threshold but below the frequency threshold. This is the area of secondary concern.
- Attribute levels found in the bottom right quadrant represent attribute levels that are below the mean drop threshold but above the frequency threshold. This is another area of secondary concern.
Net Penalty Tables
- Attribute: Selected attributes.
- Category Low: The collapsed lower levels of your scale; lowest value label is shown.
- Net Penalty: Weighted difference of the means between JAR and 'not enough' levels.
The formula: Net Penalty= [Proportion indicated Not Enough] *[Mean Drop]
- Mean Drop: The difference in the means between JAR and 'not enough' levels.
- % Not Enough: Percentage of respondents who selected below JAR.
- % JAR: Percentage of respondents who selected JAR.
- % Too Much: Percentage of respondents who selected above JAR.
- Mean Drop: The difference in the means between JAR and 'too much' levels.
- Net Penalty: Weighted difference of the means between JAR and 'too much' levels.
The formula: Net Penalty= [Proportion indicated Too Much] *[Mean Drop]
- Category High: The collapsed higher levels of your scale; highest value label is shown.
Net Penalty Graphs
This sheet provides a quick and clear way to see which attributes are having the greatest effect on the liking of your product.
The Net Penalty thresholds are selected before running the report.
There is a separate graph for each sample (you will need to scroll down to see them all). Within the graph you will see the following:
- Each attribute that meets your threshold settings.
- The net penalty of the 'too much' and 'not enough' levels.
- The frequency that the JAR level was selected, displayed as a percentage.
The net penalty is calculated by multiplying percentage of panelists
that indicated an attribute other than JAR by the mean drop for that
particular attribute. The calculation combined by impact threshold
produces the graphs.
Troubleshooting Report Issues
For Penalty analysis report to be compatible and available as a
reporting option in your test or section, the test or section must have
at least one Category Liking question and at least one Category JAR
scale in it.
If it happens that you encounter an issue when generating this report, there are some troubleshooting steps that you can perform. Let's explore!
The Penalty Analysis Report is Not an Option!
Explanation: One of the following might be true about your test or section:
- Your test or section is not a Standard test type. Discrimination tests are not compatible with the Penalty analysis report. In conclusion, there is nothing you can do to make the report compatible with such tests or sections if you already collected the data.
If it happened that you accidentally used Choose questions instead of Category questions, you might be able to export the results, create a new test using Category questions and import the results.
--or
- Your test or section is a Standard test type, but it is set up without any Category questions questions. This also makes the report incompatible with the test, therefore it will not be available to be selected from the list of reports.
Reminder, Penalty analysis report cannot be generated if there are no Category liking and Category JAR questions in the test or section.
--or
- Your test or section is a Standard test type and it has Category questions named "Overall Liking" and "JAR", but those Category questions are not set up as Liking type and JAR type. Instead they were accidentally set up as Intensity and Purchase intent question types.
Solution: Don't worry, you can still make it work! If it happens that the question types are accidentally set to a different variety of Category scale questions, even after the data is collected, the question types can be changed to a proper type. If your test is set to Complete, you will have to undo complete to be able to make the change, but please do not delete results to make this change! Merely naming a Category question as "Overall liking" or a "JAR" does not make them liking or JAR type.
Changing the question type is done in the Build tab, in the Category Question Options, as seen in the image below.

The Penalty Analysis Report Generates, but Gives Errors When Trying to Open it!
Problem: You generated and downloaded the report without any issues, but when you tried to open it, you received an error message similar to this: "We found a problem with some content in 'test name Penalty analysis.xlsx'. Do you want us to try to recover as much as we can? If you trust the source of this workbook, click Yes."
Explanation: This usually indicates that the threshold values set before generating the report are excluding all the data out.
Solution: Review the threshold values before generating the report again by following these steps:
- Follow the steps for generating the report and focus on the step 4 where you click 2. Select options.
- The Penalty Inclusion setting is what is likely causing your report not to open. If the Penalty Inclusion is set to a value that excludes all the data, then there is nothing in the report, and it gives errors. For example, if you set the Penalty Inclusion to be 0.2, but your JAR data shows that for all attributes for all samples its frequency is greater than 80% then there is nothing to include in the report, therefore errors generate when trying to open the report.

If the default Penalty Inclusion of 0.2 is your company standard, you can conclude that no attribute affected any of the samples' overall liking.
If your company standard for the Penalty Inclusion is different from 0.2, specify the right value or uncheck the option and generate the report again.
The Penalty Analysis Report Generates, but It's Missing a Sample!
Problem: You generated and downloaded the report without any
issues, but when you opened it up, you noticed that at least one sample was missing from the report!
Explanation: There could be two reasons for this:
- You may have filtered out some samples before generating the report and forgot that you did that.
Solution: Clear of otherwise update the Filters and generate the report again.

Similar to the previous example, this usually indicates that the threshold values set before generating the report are excluding some of the data out.
Solution: Review the threshold values before generating the report again by following these steps:
- Follow the steps for generating the report and focus on the step 4 where you click 2. Select options.
- The Penalty Inclusion setting is what is likely causing your report to exclude some samples. If the Penalty Inclusion
is set to a value that excludes some of the data, then some samples might not get included in the report. For example, if you set the Penalty Inclusion
to be 0.2, but your JAR data for one or more samples is shows that for all attributes for those samples its frequency is greater than 80% then those samples will not be included in the report. In our example image we can see that the Sample 1's JAR frequency is above 80 for all attributes, therefore, the sample is not included in the report; its liking is not affected by anything at the set threshold levels.

If the default Penalty Inclusion of 0.2 is your company standard, you can conclude that the overall liking of sample in question was not affected by any attribute affected.
If your company standard for the Penalty Inclusion is different from 0.2, specify the right value or uncheck the option and generate the report again.