6 ◾ Simple Statistical Methods for Software Engineering
When data quality is low we change the rules of analyses; we do not discard
the data.
Steven’s measurement theory has cast a permanent inuence in statistical
methods.
e lower scales with allegedly inferior data quality found several applications
in market research and customer satisfaction (CSAT) measurement. CSAT data are
collected in almost every software project, and an ordinal scale designed by Likert
[4] is extensively used at present for this purpose. We can improve CSAT data
quality by switching over to the ratio scale, as in the Net Promoter Score approach
invented by Frederick [5] to measure CSAT. CSAT data quality is our own making.
With better quality, CSAT data manifest better resolution that in turn supports a
comprehensive and dependable analysis.
e advent of articial intelligence has increased the scope of lower scale data. In
these days of fuzzy logic, even text can be analyzed, fullling the vision of the German
philosopher Frege, who strived to establish mathematical properties of text. Today, the
lower scales have proved to be equally valuable in their ability to capture truth.
Error
All data contain measurement errors, whether the data are from a scientic laboratory
or from a eld survey. Errors are the least in a laboratory and the most in a eld survey.
We repeat the measurement of a product in an experiment, and we may get results that
vary from trial to trial. is is the “repeatability” error. If many experimenters from dif-
ferent locations repeat the measurement, additional errors may appear because of per-
son to person variation and environmental variation known as “reproducibility” error.
ese errors, collectively called noise, in experiments can be minimized by replication.
e discrepancy between the mean value of measured data and the true value
denotes “bias.” Bias due to measuring devices can be corrected by calibrating the
devices. Bias in estimation can be reduced by adopting the wide band Delphi
method. Bias in regularly collected data is dicult to correct by statistical methods.
Both bias and noise are present in all data; the magnitude varies. Special pur-
pose data such as those collected in experiments and improvement programs have
the least. Data regularly collected from processes and products have the most. If the
collected data could be validated by team leaders or managers, most of the human
errors could be reduced. Statistical cleaning of data is possible, to some extent, by
using data mining approaches, as shown by Han and Kamber [6]. Hundreds of
tools are available to clean data by using standard procedures such as auditing,
parsing, standardization, record matching, and house holding. However, data vali-
dation by team leaders is far more eective than automated data mining technol-
ogy. Even better is to analyze data and spot outliers and odd patterns and let these
data anomalies be corrected by process owners. Simple forms of analysis such as line
graphs, scatter plots, and box plots can help in spotting bad data.