Chapter 8. Staffing Utility: The Concept and Its Measurement

Management ideas and programs often have been adopted and implemented because they were fashionable (for example, Total Quality Management, Quality Circles, reengineering) or commercially appealing, or because of the entertainment value they offered the target audience.[1] In an era of downsizing, deregulation, and fierce global competition, and as the costs of HR programs continue to be examined closely by operating executives, HR executives are under increasing pressure to justify new or continuing programs in more tangible ways. Such justification requires considering the relative payoffs (utilities) to the organization of alternatives for reaching objectives.

This chapter is the first of four chapters that describe techniques for calculating the monetary value of HR programs. This chapter and Chapter 10, “The Payoff from Enhanced Selection,” focus specifically on the payoff from improved staffing. Chapter 11, “Costs and Benefits of HR Development Programs,” focuses on how the logical frameworks for staffing can be adapted to calculate the monetary value of employee training and development. The monetary-value estimation techniques have been particularly well-developed when applied to staffing programs. The combination of analytics based on widely-applicable statistical assumptions, plus a logical approach for combining information to connect to the quality of the work force, and analytical frameworks and tools to understand how work force quality affects pivotal organizational outcomes, has produced sophisticated frameworks.

We begin this chapter by describing the logic underlying the value of staffing decisions, in terms of the conditions that define that value, and that when satisfied lead to high value. After that, we present a broad overview of utility analysis as a way to improve organizational decisions, and decisions about human capital in particular. Note that many of the examples in this chapter will refer to “dollar-valued” outcomes, because the research was conducted in the United States. However, the same concepts apply to any currency.

Recall from Chapter 2, “Analytical Frameworks for HR Measurement,” that utility analysis generally refers to frameworks that help decision makers analyze in a systematic manner the subjective value, or expected utility of alternative outcomes associated with a decision. The expected utility or usefulness of each outcome is obtained by summing a rating of the outcomes importance or value to the decision maker multiplied by the expectation or probability of achieving that outcome. After summing these values across all outcomes, the decision rule is to choose the option with the highest expected utility. The approach to staffing utility measurement is similar; but instead of simple estimates and multiplication, the formulas incorporate more nuanced approaches to probabilities, value estimation, and combinations of the individual elements.

A Decision-Based Framework for Staffing Measurement

Measures exist to enhance decisions. With respect to staffing decisions, measures are important to the decisions of applicants, potential applicants, recruiters, hiring managers, and HR professionals. These decisions include how to invest scarce resources (money, time, materials, and so on) in staffing techniques and activities, such as alternative recruiting sources, different selection and screening technologies, recruiter training or incentives, and alternative mixes of pay and benefits to offer desirable candidates. Staffing decisions also include decisions by candidates about whether to entertain or accept offers, and by hiring managers about whether to devote time and effort to landing the best talent. Increasingly, such decisions are not made exclusively by HR or staffing professionals, but in conjunction with managers outside of HR, and other key constituents.[2]

Effective staffing requires measurements that diagnose the quality of the decisions of managers and applicants. Typical staffing measurement systems fail to reflect these key decisions, so they end up with significant limitations and decision risks. For example, selection tests may be chosen solely based on their cost and predictive relationships with turnover or performance ratings. Recruitment sources may be chosen solely based on their cost and volume of applicants. Recruiters may be chosen based solely on their availability, and evaluated only on the volume of applicants they produce. Staffing is typically treated not as a process, but as a set of isolated activities (recruiting, selecting, offering/closing, and so forth).

Fixing these problems requires a systematic approach to staffing that treats it as a set of decisions and processes that begins with a set of outcomes, identifies key processes, and then integrates outcomes with processes. Consider outcomes, for example. We know that the ultimate value of a staffing system is reflected in the quality of talent that is hired or promoted and retained. In fact, a wide variety of measures exist to examine staffing quality, but generally they fall into seven categories:

  • Cost (for example, cost per hire, cost of assessment activities [tests, interviews, background checks])

  • Time of activities (for example, time to fill vacancies, time elapsed from interview to offer)

  • Volume and yield (for example, total number of applicants, yield of hires from applicants)

  • Diversity and EEO compliance (for example, demographic characteristics of applicants at each stage of the hiring process)

  • Customer/constituent reactions (for example, judgments about the quality of the process and impressions about its attractiveness)

  • Quality attributes of the talent (for example, pre-hire predictive measures of quality [selection tests, interviewer ratings] as well as post-hire measures of potential and competency)

  • Value impact of the talent (for example, measures of actual job performance and overall contribution to the goals of a unit or organization)

This chapter focuses primarily on two of these measures: the quality and value impact of talent. At the same time, it is important not to lose sight of the broader staffing processes within which screening and selection of talent takes place. Figure 8-1 is a graphic illustration of the logic of the staffing process and talent flows.

Logic of staffing processes and talent flows.

Figure 8-1. Logic of staffing processes and talent flows.

Groups of individuals (talent pools) flow through the various stages of the staffing process, with each stage serving as a filter that eliminates a subset of the original talent pool. The top row of Figure 8-1 shows the results of the filtering process, beginning with a potential labor pool that is winnowed through recruitment and selection down to a group that receives offers, and then is winnowed further as some accept offers and remain with the organization.

The “staffing processes” in the lower row show the activities that accomplish the filtering sequence, beginning with building and planning (forecasting trends in external and internal labor markets, inducing potential applicants to develop qualifications to satisfy future talent demands), and ending with on-boarding (orientation, mentoring, removing barriers to performance). Integrating measurement categories with the process steps shown in Figure 8-1 provides a decision-based framework for evaluating where staffing measures are sufficient, and where they may be lacking.

Now that we have presented the “big picture” of the staffing process, let us focus more specifically on one component of that process, employee selection (specifically, on assessing the value of selection by means of utility analysis).

Framing Human Capital Decisions Through the Lens of Utility Analysis

Utility analysis is a framework to guide decisions about investments in human capital.[3] It is the determination of institutional gain or loss (outcomes) anticipated from various courses of action. When faced with a choice among strategies, management should choose the strategy that maximizes the expected utility for the organization.[4] To make the choice, managers must be able to estimate the utilities associated with various outcomes. Estimating utilities traditionally has been the Achilles heel of decision theory,[5] but a less-acute problem in business settings, where gains and losses may be estimated by objective behavioral or cost-accounting procedures, often in monetary terms.

Our objective in this chapter is to describe three different models of staffing utility analysis, focusing on the logic and analytics of each one. Chapter 9, “The Economic Value of Job Performance,” Chapter 10, and Chapter 11, then build on these ideas, emphasizing measures and processes, to communicate results to operating executives and to show how staffing, training, and other HR programs can be evaluated from a return-on-investment perspective.

Overview—The Logic of Utility Analysis

At its core, utility analysis considers three important parameters: quantity, quality, and cost. A careful look at Figure 8-1 shows that the top row refers to the characteristics of candidates for employment as they flow through the various stages of the staffing process. At each stage, the candidate pool can be thought of in terms of the quantity of candidates, the average and dispersion of the quality of the candidates, and the cost of employing the candidates. For example, the “applicant pool” might have a quantity of 100 candidates, with an average quality value of $100,000 per year, and a variability in quality value that ranges from a low of $50,000 to a high of $170,000. This group of candidates might have an anticipated cost (salary, benefits, training, and so on) of 70 percent of their value. After screening and selection, the “offer candidates” might have a quantity of 50 who receive offers, with an average quality value of $150,000 per year, ranging from a low of $100,000 to a high of $160,000. Candidates who receive offers might require employment costs of 80 percent of their value, because we have identified highly qualified and sought-after individuals. Eventually, the organization ends up with a group of “new hires” (or promoted candidates in the case of internal staffing) that can also be characterized by quantity, quality, and cost.

Similarly, the bottom row of Figure 8-1 reflects the staffing processes that create the sequential filtering of candidates. Each of these processes can be thought of in terms of the quantity of programs and practices used, the quality of the programs and practices as reflected in their ability to improve the value of the pool of individuals that survives, and the cost of the programs and practices in each process. For example, the quality of selection procedures is often expressed in terms of their validity, or accuracy in fore casting future job performance. Validity is typically expressed in terms of the correlation (see Chapter 2) between scores on a selection procedure and some measure of job performance, such as the dollar volume of sales. Validity may be increased by including a greater quantity (such as a battery of selection procedures), each of which focuses on an aspect of knowledge, skill, ability, or other characteristic that has been demonstrated to be important to successful performance on a job. Higher levels of validity imply higher levels of future job performance among those selected or promoted, thereby improving the overall payoff to the organization. As a result, those candidates that are predicted to perform poorly never get hired or promoted in the first place.

Decision makers naturally focus on the cost of selection procedures because they are so vividly depicted by the standard accounting systems; but the cost of errors in selecting, hiring, or promoting the wrong person is often much more important. As explained in Chapter 9, the difference in value between an average performer versus a superior performer is often much higher than the difference in the cost of improved staff. In the case of executives, a company often has to pay large fees to headhunters, and poor performance can have serious consequences in terms of projects, products, and customers. That cost can easily run $1 million to $3 million.[6]

In summary, the overall payoff to the organization (utility) from the use of staffing procedures depends on three broad parameters: quantity, quality, and cost. Each of the three staffing utility models that we examine in this chapter addresses two or more of these parameters. The models usually focus on the selection part of the processes of Figure 8-1, but they have implications for the other staffing stages too. Each model defines the quality of candidates in a somewhat different way, so we will start with the models that make relatively basic assumptions, to those that are increasingly more sophisticated.

Utility Models and Staffing Decisions

The utility of a selection device is the degree to which its use improves the quality of the individuals selected beyond what would have occurred had that device not been used.[7] In the context of staffing or employee selection, three of the best-known utility models are those of Taylor and Russell;[8] Naylor and Shine;[9] and Brogden, Cronbach, and Gleser.[10] Each of them defines the quality of selection in terms of one of the following:

  • The proportion of individuals in the selected group who are considered successful

  • The average standard score on a measure of job performance for the selected group

  • The dollar payoff to the organization resulting from the use of a particular selection procedure

The remainder of this chapter considers each of these utility models and its associated measure of quality in greater detail.

The Taylor-Russell Model

Many decision makers might assume that if candidate ratings on a selection device (such as a test or interview) are highly associated with their later job performance, the selection device must be worth investing in. After all, how could better prediction of future performance not be worth the investment? However, if the pool of candidates contains very few unacceptable candidates, better testing may do little good. Or, if the organization generates so few candidates that it must hire almost all of them, again better testing will be of little use. Taylor and Russell translated these observations into a system for measuring the tradeoffs,[11] suggesting that the overall utility or practical effectiveness of a selection device depends on more than just the validity coefficient (the correlation between a predictor of job performance and a criterion measure of actual job performance). Rather, it depends on three parameters: the validity coefficient (r), the selection ratio (SR) (the proportion of applicants selected), and the base rate (BR) (the proportion of applicants who would be successful without the selection procedure).

Taylor and Russell defined the value of the selection system as the “success ratio,” which is the ratio of the number of hired candidates who are judged successful on the job divided by the total number of candidates that were hired. They published a series of tables illustrating the interactive effect of different validity coefficients, selection ratios, and base rates on the success ratio. The success ratio indicates the quality of those selected. The difference between the success ratio and the base rate (which reflects the success ratio without any added selection system) is a measure of the incremental value of the selection system over what would have happened if it had not been used. Let’s develop this logic and its implications in more detail, and show you how to use the tables that Taylor and Russell developed.

Analytics

There are three key, underlying assumptions of this model:

  1. It assumes fixed-treatment selection. (That is, individuals are chosen for one specified job, treatment, or course of action that cannot be modified.) For example, if a person is selected for Treatment A, a training program for slow learners, transfer to Treatment B, fast-track instruction, is not done, regardless of how well the person does in Treatment A.

  2. The Taylor-Russell model does not account for the rejected individuals who would have been successful if hired (erroneous rejections). Because they are not hired, their potential value, or what they might contribute to other employers who now can hire them, is not considered.

  3. The model classifies accepted individuals into successful and unsuccessful groups. All individuals within each group are regarded as making equal contributions. That means that being minimally successful is assumed to be equal in value to being highly successful, and being just below the acceptable standard is assumed to be equal in value to being extremely unsuccessful.

Of course these assumptions may not hold in all situations; but even with these basic assumptions, Taylor and Russell were able to generate useful conclusions about the interplay between testing and recruitment. For example, the Taylor-Russell model demonstrates convincingly that even selection procedures with relatively low validities can increase substantially the percentage of those selected who are successful, when the selection ratio is low (lots of candidates to choose from) and when the base rate is near 50 percent (about half the candidates would succeed without further testing, so there are lots of middle-level candidates who can be sorted by better selection). Let us consider the concepts of selection ratio and base rate in greater detail.

The selection ratio is simply the number of candidates that must be hired divided by the number of available candidates to choose from. A selection ratio (SR) of 1.0 means the organization must hire everyone, so testing is of no value, because there are no selection decisions to be made. The closer the actual SR is to 1.0, the harder it is for better selection to pay off. The opposite also holds true; as the SR gets smaller, the value of better selection gets higher. (For example, a selection ratio of .10 means the organization has ten times more applicants than it needs, and must only hire 10 percent of the available applicants.) The wide-ranging effect that the SR may exert on a predictor with a given validity is illustrated in Figure 8-2. In each case, Xc represents a cutoff score on the predictor. As can be seen in Figure 8-2, even predictors with low validities can be useful if the SR is low, so that the organization needs to choose only the cream of the crop. Conversely, with high selection ratios, a predictor must possess very high validity in order to increase the percentage successful among those selected.

Effect of varying selection ratios on a predictor with a given validity.

Note: The oval is the shape of a scatterplot corresponding to r = 0.50; r = validity coefficient; SR = selection ratio; Xc = cutoff score.

Figure 8-2. Effect of varying selection ratios on a predictor with a given validity.

It might appear that, because a predictor that demonstrates a particular validity is more valuable with a lower selection ratio, one should always opt to reduce the SR (become more selective). However, the optimal strategy is not this simple.[12] When the organization must achieve a certain quota of individuals, lowering the SR means the organization must increase the number of available applicants, which means expanding the recruiting and selection effort. In practice, that strategy may be too costly to implement, as later research demonstrated convincingly.[13]

Utility, according to Taylor and Russell, is affected by the base rate (the proportion of candidates who would be successful without the selection measure). To be of any use in selection, the measure must demonstrate incremental validity by improving on the BR. That is, the selection measure must result in more correct decisions than could be made without using it. As Figure 8-3 demonstrates, when the BR is either very high or very low, it is difficult it for a selection measure to improve upon it.

Effect of varying base rates on a predictor with a given validity.

Note: The oval is the shape of a scatterplot corresponding to r = 0.50; BR = base rate; r = validity coefficient; Yc = minimum level of job performance.

Figure 8-3. Effect of varying base rates on a predictor with a given validity.

In each panel of the figure, Yc represents the minimum level of job performance (criterion cutoff score) necessary for success. That value should not be altered arbitrarily. Instead, it should be based on careful consideration of the true level of minimally acceptable performance for the job.[14] Figure 8-3 illustrates that with a BR of 0.80, it would be difficult for any selection measure to improve on the base rate. In fact, when the BR is 0.80 and half of the applicants are selected, a validity of 0.45 is required to produce an improvement of even 10 percent over base rate prediction. This is also true at very low BRs. (As would be the case, for example, in the psychiatric screening of job applicants.) Given a BR of 0.20, an SR of 0.50, and a validity of 0.45, the percentage successful among those selected is 0.30 (once again representing only a 10 percent improvement in correct decisions). Selection measures are most useful when BRs are about 0.50.[15] As the BR departs radically in either direction from this value, the benefit of an additional predictor becomes questionable. The lesson is obvious: Applications of selection measures to situations with markedly different SRs or BRs can result in quite different predictive outcomes. If it is not possible to demonstrate significant incremental utility by adding a predictor, the predictor should not be used, because it cannot improve on current selection procedures.

Figure 8-4 presents all of the elements of the Taylor-Russell model together. In this figure, the criterion cutoff (Yc) separates the present employee group into satisfactory and unsatisfactory workers. The predictor cutoff (Xc) defines the relative proportion of workers who would be hired at a given level of selectivity. Areas A and C represent correct decisions—that is, if the selection measure were used to select applicants, those in area A would be hired and become satisfactory employees. Those in area C would be rejected correctly because they scored below the predictor cutoff and would have performed unsatisfactorily on the job. Areas B and D represent erroneous decisions; those in area B would be hired because they scored above the predictor cutoff, but they would perform unsatisfactorily on the job, and those in area D would be rejected because they scored below the predictor cutoff, but they would have been successful if hired.

Effect of predictor and criterion cutoffs on selection decisions.

Note: The oval is the shape of the scatterplot that shows the overall relationship between predictor and criterion scores

Figure 8-4. Effect of predictor and criterion cutoffs on selection decisions.

The following ratios were used by Taylor and Russell in developing their tables:

Equation 8-1. 

Equation 8-2. 

Equation 8-3. 

By specifying the validity coefficient, the base rate, and the selection ratio, and making use of Pearson’s “Tables for Finding the Volumes of the Normal Bivariate Surface,”[16] Taylor and Russell developed their tables (see Appendix A). The usefulness of a selection measure thus can be assessed in terms of the success ratio that will be obtained if the selection measure is used. The gain in utility to be expected from using the instrument (the expected increase in the percentage of successful workers) then can be derived by subtracting the base rate from the success ratio (Equation 8-3 minus Equation 8-1). For example, given an SR of 0.10, a validity of 0.30, and a BR of 0.50, the success ratio jumps to 0.71 (a 21 percent gain in utility over the base rate; to verify this figure, see Appendix A).

The validity coefficient referred to by Taylor and Russell is, in theory, based on present employees who have already been screened using methods other than the new selection procedure. It is assumed that the new procedure will simply be added to a group of selection procedures used previously, and it is the incremental gain in validity from the use of the new procedure that is most relevant.

Perhaps the major shortcoming of this utility model is that reflects the quality of the resulting hires only in terms of success or failure. It views the value of hired employees as a dichotomous classification, successful and unsuccessful; and as the tables in Appendix A demonstrate, when validity is fixed, the success ratio increases as the selection ratio decreases. (Turn to Appendix A, choose any particular validity value, and note what happens to the success ratio as the selection ratio changes from 0.95 to 0.05.) Under those circumstances, the success ratio tells us that more people are successful, but not how much more successful. In practice, situations may arise in which one would not expect the average level of job performance to change as a function of higher selection standards, such as food servers at fast-food restaurants. Their activities have become so standardized that there is little opportunity for significant improvements in performance after they have been selected and trained. The relationship between the value of such jobs to the organization and variations in performance demonstrate essentially flat slopes. In such situations, it may make sense to think of the value of hired candidates as either being above the minimum standard or not.

For many jobs, however, one would expect to see improvements in the average level of employee value from increased selectivity. In most jobs, for example, a very high-quality employee is more valuable than one who just meets the minimum standard of acceptability. When it is reasonable to assume that the use of higher cutoff scores on a selection device will lead to higher levels of average job performance by those selected, the Taylor-Russell tables will underestimate the actual amount of value from the selection system. That observation led to the development of the next framework for selection utility, the Naylor-Shine Model.

The Naylor-Shine Model

Unlike Taylor and Russell, the Naylor and Shine utility model does not require that employees be split into satisfactory and unsatisfactory groups by specifying an arbitrary cutoff on the criterion (job performance) dimension that represents minimally acceptable performance.[17] The Naylor-Shine model defines utility as the increase in the average criterion score (for example, the average level of job performance of those selected) expected from the use of a selection process with a given validity and SR. The quality of those selected is now defined as the difference in average level of quality of the group that is hired, versus the average quality in the original group of candidates.

Like Taylor and Russell, Naylor and Shine assume that the relationship between predictor and criterion is bivariate normal (both scores on the selection device and performance scores are normally distributed), linear, and homoscedastic. The validity coefficient is assumed to be based on the concurrent validity model.[18] That model reflects the gain in validity from using the new selection procedure over and above what is presently available using current information.

In contrast to the Taylor-Russell utility model, the Naylor-Shine approach assumes a linear relationship between validity and utility. That is, the higher the validity, the greater the increase in the average criterion score of the selected group compared to the average criterion score that would have been achieved by the candidate group. The basic equation underlying the Naylor-Shine model is shown in Equation 8-4:

Equation 8-4. 

Where yi is the average criterion score (in standard-score units)[19] of those selected, rxy is the validity coefficient, λi is the height of the normal curve at the predictor cutoff, Zxi (expressed in standard-score units), and φi is the selection ratio. Equation 8-4 applies whether rxy represents a correlation between two variables or it is a multiple-regression coefficient.[20]

Using Equation 8-4 as a basic building block, Naylor and Shine present a series of tables (see Appendix B) that specify, for each SR, the standard (predictor) score that produces that SR, the ordinate of the normal curve at that point, and the quotient λii. The tables can be used to answer several important HR questions:

  • Given a specified SR, what will be the average criterion (such as performance) level of those selected?

  • Given a certain minimum cutoff score on the selection device above which everyone will be hired, what will the average criterion level (yi) be?

  • Given a desired improvement in the average criterion score (such as performance) of those selected, and assuming a certain validity, what SR and/or predictor cutoff value (in standard-score units) should be used?

Let us work through some examples, using the tables in Appendix B.

In each of the following examples, assume that rxy, the validity of our predictor, is positive and equal to 0.40. Of course, it is also possible that the validity of a predictor could be negative (for example, higher levels of job satisfaction related systematically to lower intentions to quit). Under these circumstances, the general rule is to reverse the sign of rxy and Zxi everywhere in the calculations.

  1. With a selection ratio of 50 percent (φi = 0.50), what will be the average performance level of those selected?

    Solution: Enter the table at φi = 0.50 and read λii =0.80

    yi = rxyλii = (0.40)(0.80) = 0.32

    Thus, the average criterion score of those selected, using a SR of 0.50, is 0.32 Z-units (roughly one third of a standard deviation) better than the unselected sample.

  2. With a desired cutoff score set at .96 standard deviations below the average of the applicant pool (Zxi = – 0.96), what will the standardized value of the criterion (yi) be?

    Solution: Enter the table at Zxi = –0.96 and read λii = 0.30.

    yi = rxyλii = (0.40) (0.30) = 0.12

    Thus, using this cutoff score on our predictor results in an average criterion score of about one-eighth of a standard deviation (0.12 Z-units) higher than the average of the unselected applicant pool.

  3. If we want to achieve an average standardized level of performance on our criterion (such as job performance) among those selected that is half a standard-deviation higher than the average of the applicant pool (yi = 0.50), and assuming a validity of .40, what SR do we need to achieve, and what is the predictor cutoff value that will achieve that SR?

    Solution: Because yi = rxyλi/φi then

    λi/φi = yi /rxy = 0.50/0.40 = 1.25

    Enter the table at λi/φi = 1.25 and read φi = 0.2578 and Zxi = 0.65. Thus, to achieve an average improvement of 0.50 (one half) standard deviation in job performance, an SR of 0.2578 is necessary (we must select only the top 25.78 percent of applicants), and to achieve that, we must set a cutoff score on the predictor of 0.65 standard deviations above the average among our applicants.

The Naylor-Shine utility approach is more generally applicable than Taylor-Russell because in many, if not most, cases, an organization could expect an increase in average job performance as it becomes more selective, using valid selection procedures. However, “average job performance” is expressed in terms of standard (Z) scores, which are more difficult to interpret than are outcomes more closely related to the specific nature of a business, such as dollar volume of sales, units produced or sold, or costs reduced. With only a standardized criterion scale, one must ask questions such as “Is it worth spending $10,000 to select 50 people per year, to obtain a criterion level 0.50 standard deviations (SDs) greater than what we would obtain without the predictor?”[21] Some HR managers may not even be familiar with the concept of a standard deviation, and would find it difficult to attach a dollar value to a 0.50 SD increase in criterion performance.

Neither the Taylor-Russell nor the Naylor-Shine models formally integrates the concept of selection system cost, nor the monetary gain or loss, into the utility index. Both describe differences in the percentage of successful employees (Taylor-Russell) or increases in average criterion score (Naylor-Shine), but tell us little about the benefits to the employer in monetary terms. The Brogden-Cronbach-Gleser model, discussed next, was designed to address these issues.

The Brogden-Cronbach-Gleser Model

Brogden showed that under certain conditions, the validity coefficient is a direct index of “selective efficiency.” That means that if the criterion and predictor are expressed in standard-score units, rxy represents the ratio of the average criterion score made by persons selected on the basis of their predictor scores (The Brogden-Cronbach-Gleser Modely) to the average score made if one had selected them based on their criterion scores (The Brogden-Cronbach-Gleser Modely). Of course, it is usually not possible to select applicants based on their criterion scores (because one cannot observe their criterion scores before they are hired), but Brogden’s insight means that the validity coefficient represents the ratio of how well an actual selection process does, compared to that best standard. Equation 8-5 shows this algebraically:

Equation 8-5. 

The validity coefficient has these properties when: (1) both the predictor and criterion are continuous (that is, they can assume any value within a certain range, and are not divided into two or more categories); (2) the predictor and criterion distributions are identical (not necessarily normal, but identical); (3) the regression of the criterion on the predictor is linear; and (4) the selection ratio (SR) is held constant.[22]

As an illustration, suppose a firm wants to hire 20 people for a certain job and must choose the best 20 from 85 applicants. Ideally, the firm would hire all 85 for a period of time, collect job performance (criterion) data, and retain the best 20, those obtaining the highest criterion scores. The average criterion score of the 20 selected this way would obviously be the highest obtainable with any possible combination of 20 of the 85 applicants. Such a procedure is usually out of the question, so organizations use a selection process and choose the 20 highest scorers. Equation 8-5 indicates that the validity coefficient may be interpreted as the ratio of the average criterion performance of the 20 people selected on the basis of their predictor scores to the average performance of the 20 who would have been selected had the criterion itself been used as the basis for selection. To put a monetary value on this, if selecting applicants based on their actual behavior on the job, would save an organization $300,000 per year over random selection, a selection device with a validity of 0.50 could be expected to save $150,000 per year. Utility is therefore a direct linear function of validity, when the conditions noted above are met.

Equation 8-5 does not include the cost of selection, but Brogden later used the principles of linear regression to demonstrate the relationships of cost of selection, validity, and selection ratio to utility, expressed in terms of dollars.[23]

Recall that our ultimate goal is to identify the monetary payoff to the organization when it uses a selection system to hire employees. To do this, let’s assume we could construct a criterion measure expressed in monetary terms. We’ll symbolize it as y$. Examples of this might include the sales made during a week/month/quarter by each of the salespersons on a certain job, or the profit obtained from each retail operation managed by each of the store managers across a country, or the outstanding customer debts paid during a week/month/quarter for the customers handled by each of a group of call-center collection agents. If we call that criterion measure “y”, then here is a plain-English and mathematical description of Brogden’s approach.[24]

Step 1: Express the predictor-criterion relationship as a formula for a straight line

Recall the formula for a straight line that most people learn in their first algebra class, shown here as Equation 8-6.

Equation 8-6. 

  • Where:

  • y = dependent variable, or criterion (such as a job performance measure)

  • x = independent variable that we hope predicts our criterion (such as job performance)

  • a = “y intercept” or where the line crosses the y axis of a graph when x = 0

  • b = slope, or “rise over the run” of the line; that is, the change in y (for example, change in sales) for every one-unit change in x (score on a sales-aptitude test).

First, let’s change this equation slightly. Let’s substitute the symbols b0 for a, b1 for b, and y$ for y. In this way, we go from Equation 8-6 above to Equation 8-7.

Equation 8-7. 

Then, let’s add an e after the x, to reflect that there is some random fluctuation or “error” in any straight-line estimate, and we get Equation 8-8.

Equation 8-8. 

Our original formulas (Equations 8-6 and 8-7) described the points that fall exactly on a straight line, but Equation 8-8 describes points that fall around a straight line. Figure 8-5 shows this idea as a straight line passing through an ellipse. The ellipse represents the cloud of score combinations that might occur in an actual group of people, and the line in the middle is the one that gets as close as possible to as many of the points in the cloud. Some people describe this picture as a hot dog on a stick!

Dispersion of actual criterion and predictor scores.

Figure 8-5. Dispersion of actual criterion and predictor scores.

In the context of staffing, x would be each employee’s score on some selection process, and y would be the same employee’s subsequent criterion score (such as performance on the job). If we don’t know yet how someone is going to perform on the job (which we can’t know before the person is hired), a best guess or estimate of how the employee might perform on the job would be the y$ value obtained from plugging the applicant’s x score into Equation 8-7.

The letter e in Equation 8-8 is called “error,” because although our estimate of y$ from Equation 8-7 might be a good guess, it is not likely to be exactly the level of job performance obtained by that applicant later on the job. Note, because y is the actual performance attained by that applicant, then yy$ = e. The “error” by which our original predicted level of job performance, y$, was off from the applicant’s actual job performance y is equal to e.

Ordinary least-squares regression analyses can be used to calculate the “best”-fitting straight line (that is, Equation 8-7), where best means the formula for the straight line Dispersion of actual criterion and predictor scores. (Equation 8-7) that minimizes the sum of all squared errors (e2). This is where the “least-squares” portion of the “ordinary least-squares” label comes from.

Step 2: Standardize x

To get back to the validity coefficient, we need to covert the actual or “raw” scores on our predictor and criterion to standardized form. Starting with Equation 8-7, reprinted here as Equation 8-9, let’s see how this works.

Equation 8-9. 

Let’s first standardize all the applicants’ selection-process scores (that is, take their original scores, subtract the average, and divide by the standard deviation), as shown in Equation 8-10:

Equation 8-10. 

  • Where:

  • xi = selection-process score earned by applicant i.

  • zi = “standard” or Z score corresponding to the xi score for applicant i.

  • = average or mean selection-process score, typically of all applicants, obtained in some sample

  • SDx = standard deviation of xi around , or

When Equation 8-9 is modified to reflect the fact that x is now standardized, it becomes Equation 8-11.

Equation 8-11. 

Step 3: Express the equations in terms of the validity coefficient

Finally, let’s modify Equation 8-11 to show the role of the validity coefficient using this selection process. We want to know the expected value (or the most likely average value) of y$ for the hired applicants. Modifying Equation 8-11 to express all the elements with a capital “E” for expected value, we have:

Equation 8-12. 

Thus E(y$) means the “expected value of the criterion, y, in monetary terms.” Also note that the letter s is now subscripted to the letter z, to show that the criterion scores are from the group of applicants who are actually selected (subscript s stands for “selected”).

Remember that “expected value” typically means “average,” so and . So, substituting these values into Equation 8-11 yields the following:

Equation 8-13. 

We can calculate simply by standardizing the selection test scores of all applicants, averaging just the scores of the individuals who were actually selected (hence the subscript s). When no selection system is used (that is, if applicants had been chosen at random), is expected to be the same as the average of z scores for all applicants. By definition, the average of all z scores in a sample is always 0. So, when = 0 then E(b1) = 0 too, and E(b0) will be the average monetary value of the criterion for individuals selected at random from the applicant pool. The symbol for the expected or average monetary criterion score for all applicants is $, so we can substitute $ for E(b0) in Equation 8-13.

Finally, the value of E(b1) is obtained using multiple-regression software (for example, the regression function in Excel). This is the regression coefficient or beta weight associated with x (as opposed to the “constant,” which is the estimate of E(b0)). By definition, the regression coefficient can also be defined as in Equation 8-14.

Equation 8-14. 

  • Where:

  • rxy = simple correlation between test scores on the employee-selection test x and the criterion measure y.

  • SDy = standard deviation of the monetary value of the criterion (such as job performance).

  • SDx = standard deviation of all applicants’ selection-test scores.

Recall, however, that we standardized applicant test scores in using Equation 8-10 to create the z variable used in Equation 8-11. The standard deviation of z scores is always 1.0. So, substituting 1 for SDx, Equation 8-14 becomes b1 = rxySDy.

Substituting μ$ for E(b0) and rxySDy for b1 in Equation 8-13, we get this.

Equation 8-15. 

Equation 8-15 describes the total expected monetary value of each selected applicant. To calculate the expected average improvement in utility, or the improvement in the monetary value produced by using the staffing system, we can subtract the expected value without using the system, which is μ$, from both sides of the equation. Because μ$ is the monetary value of criterion performance the organization expects when it chooses applicants at random, $μ$ is equal to the expected gain in monetary-valued performance from using the staffing process, as shown in Equation 8-16.

Equation 8-16. 

Step 4: Subtract the costs of the selection process

Selecting applicants requires resources. If we use the letter C to stand for the cost of applying the selection process to one applicant, and the term Na to stand for the total number of applicants to whom the selection process is applied, then the total cost of the selection process is the product of Na and C. If we divide that by the number of applicants actually selected, that gives us the average cost of the selection process per selected applicant. Finally, if we subtract the average selection process cost per selected applicant from the average value expressed in Equation 8-16, we get Equation 8-17.

Equation 8-17. 

Finally, the left-side of Equation 8-17 is often symbolized as ΔU, to stand for the “change in utility” per applicant selected from the selection system, as shown in Equation 8-18.

Equation 8-18. 

Cronbach and Gleser elaborated and refined Brogden’s derivations with respect to utility in fixed-treatment selection, and they arrived at the same conclusions regarding the effects of r, SDy, the cost of selection, and the selection ratio on utility in fixed-treatment selection. Utility properly is regarded as linearly related to validity, and if cost is zero, is proportional to validity.[25] They also adopted Taylor and Russell’s interpretation of the validity coefficient for utility calculations (that is, concurrent validity). The validity coefficient based on present employees presupposes a population that has been screened using information other than the new selection measure. It is to this population that the selection ratio will be applied.

Cronbach and Gleser argued, as did Taylor and Russell and Naylor and Shine, that selection procedures should be judged on the basis of their contribution over and above the best strategy available that makes use of prior information. Thus, any new procedure must demonstrate incremental utility before it is used. Suppose, however, that an organization wants to replace its old selection procedures with new ones. Under such circumstances, the appropriate population for deriving a validity coefficient, SDy, and SR, should be the unscreened population.[26] Figure 8-6 presents a summary of the three utility models that we have discussed.

Table 8.6. Summary of the utility indexes, data requirements, and assumptions of three utility models.

<source>Sources: Cascio, W.F. (1990). Responding to the demand for accountability: A critical analysis of three utility models. Organizational Behavior and Human Performance 25, 42.</source>

Models

Utility Index

Data Requirements

Distinctive Assumptions

Taylor-Russell (1939)

Increase in percentage successful in selected group

Validity, base rate, selection ratio

All selectees classified as either successful or unsuccessful. Equal criterion performance by all members of each group; cost of selection = $0.

Naylor-Shine (1965)

Increase in mean criterion score of selected group

Validity, selection ratio

Validity linearly related to utility; cost of selection = $0.

Brogden-Cronbach-Gleser (1965)

Increase in dollar payoff of selected group

Validity, selection ratio, criterion standard deviation in dollars

Validity linearly related to utility; cost of selection ≠ $0, criterion performance evaluated in dollars.

Note: All three models assume a validity coefficient based on present employees (concurrent validity).

Conclusion

This chapter presents some complex but elegant statistical logic. It’s sometimes hard to follow at first, but as Figure 8-1 shows, it is actually rather intuitive. The idea of each of the three “selection-utility” models is to estimate how much higher the quality of the selected employees will be compared to the candidates for selection. That change in quality depends on how selective the organization can be, how well it predicts future performance, and how much differences in performance quality translate into differences in value to the organization.

The utility models are best used with an understanding of their logic, assumptions, and data requirements. If you make that investment, you have a logical system for making wiser and more strategically relevant decisions about how to select talent both from the outside and within the organization.

Now, these equations would be fine if we actually had a monetary-valued criterion to use in estimating SDy. When a job produces very clear monetary-valued outcomes such as sales, waste, or profit, we might associate these values with each individual on the job and calculate the standard deviation of those values. Still, that would not reflect the standard deviation we might have seen in the pool of applicants, because the people on the job have already been screened in the course of the selection process. Also, even in jobs with obvious monetary outcomes, such as sales, there are other elements of the job that may be quite important, but are not reflected in individual monetary results (such as when sales people actually sell less because they are training their colleagues). In short, the value and the process for estimating SDy go to the heart of a fundamental question in all of human resources and talent management: “How much are differences in performance worth?”

At this point, you might be wondering how organizations can actually estimate the dollar value of differences in performance quality. Indeed, SDy has been the subject of much debate and there are several methods for estimating it. We discuss those in the next chapter. You might also wonder whether this same kind of logic (estimating how much better quality are our employees after a certain HR program, compared to their quality without it) might apply to programs other than selection. We have much more to say about the strategic use of utility analysis in guiding investment decisions about human capital in Chapters 10 and 11.

Exercises

Software that calculates answers to one or more of the following exercises can be found at www.shrm.org/publications/books.

1.

Use the Taylor-Russell tables (Appendix A) to solve these problems by filling in the following table:

Validity

SR

BR

Success Ratio

0.25

0.20

0.30

 

0.55

0.70

0.80

 

0.20

0.70

0.80

 

0.10

0.50

0.50

 

0.55

0.50

0.50

 

2.

Use the Naylor-Shine tables (Appendix B) to solve these problems by filling in the following table:

rxy

φi

zxi

z-baryi

0.35

0.7019

  

0.22

 

–0.30

 

0.47

  

0.65

–0.47

  

0.65

3.

Using the Brogden-Cronbach-Gleser continuous-variable utility model, what is the net gain over random selection (ΔU overall, and per selectee) given the following information?

Quota for selection: 20

SR: 0.20

SDy (standard deviation of job performance expressed in dollars): $25,000

rxy: 0.25

λ(y’): 0.28

cy: $25

Hint: To find N, the number recruited, divide the quota for selection by the SR.

4.

Given the following information on two selection procedures, and using the Brogden-Cronbach-Gleser model, what is the relative difference in payoff (overall and per selectee) between the two procedures? For both procedures, quota = 50, SR = 0.50, λ(y’) = 0.3989, and SDy = $35,000.

ry1: 0.20

c1: $100

ry2: 0.40

c2: $500

5.

You are a management consultant whose task is to do a utility analysis using the following information regarding secretaries at Inko, Inc. The validity of the Secretarial Aptitude Test (SAT) is 0.40, applicants must score 70 or better (Zx of 0.5) to be hired, and only about half of those who apply actually are hired. Of those hired, about half are considered satisfactory by their bosses. How selective should Inko be to upgrade the average criterion score of those selected by Exercises What utility model did you use to solve the problem? Why?

References

1.

Crainer, S., & Dearlove, D. (May/June 2006). Whatever happened to yesterday’s bright ideas? Across the Board, p. 34-40.

2.

Material in this section is largely drawn from Boudreau, J. W., & Ramstad, P. W. (Sept. 2004). Beyond cost-per-hire and time to fill: Supply-chain measurement for staffing. Center for Effective Organizations, University of Southern California, CEO Publication G 04-16 (468).

3.

Boudreau, J. W., & Ramstad, P. M. (2003). Strategic industrial and organizational psychology and the role of utility analysis models. In W. C. Borman, D. R. Ilgen, & R. J. Klimoski (Volume Eds.), Handbook of Psychology, Volume 12, Industrial and Organizational Psychology (p. 193-221). Hoboken, NJ: Wiley.

4.

Brealey, R. A., Myers, S. C., & Allen, F. (2006). Principles of Corporate Finance (8th ed.). Burr Ridge, IL: Irwin/McGraw-Hill.

5.

Cronbach, L. J., & Gleser, G. C. (1965). Psychological Tests and HR Decisions (2nd ed.). Urbana, IL: University of Illinois Press.

6.

Byrnes, N., & Kiley, D. (Feb. 12, 2007). Hello, you must be going. Business Week, p. 30-32. Berner, R. (Feb. 12, 2007). My year at Wal-Mart. Business Week, p. 70-74.

7.

Taylor, H. C., & Russell, J. T. (1939). The relationship of validity coefficients to the practical effectiveness of tests in selection. Journal of Applied Psychology, 23, 565-578.

8.

Naylor, J. C., & Shine, L. C. (1965). A table for determining the increase in mean criterion score obtained by using a selection device. Journal of Industrial Psychology, 3, 33-42.

9.

Brogden, H. E. (1946). On the interpretation of the correlation coefficient as a measure of predictive efficiency. Journal of Educational Psychology, 37, 64-76. Brogden, H. E. (1949). When testing pays off. Personnel Psychology, 2, 171-185. Cronbach, L. J., & Gleser, G. C. (1965). Psychological Tests and Personnel Decisions (2nd ed.). Urbana, IL: University of Illinois Press.

10.

Blum, M. L., & Naylor, J. C. (1968). Industrial Psychology: Its Theoretical and Social Foundations (rev. ed.). New York: Harper & Row.

11.

Taylor & Russell, 1939, op. cit.

12.

Sands, W. A. (1973). A method for evaluating alternative recruiting-selection strategies: The CAPER model. Journal of Applied Psychology, 57, 222-227.

13.

Boudreau, J. W., & Rynes, S. L. (1985). Role of recruitment in staffing utility analyses. Journal of Applied Psychology, 70, 354-366.

14.

Boudreau, J. W. (1991). Utility analysis for decisions in human resource management. In M. D. Dunnette & L. M. Hough (Eds.), Handbook of Industrial and Organizational Psychology (Vol. 2, 2nd ed., p. 621-745).

15.

Taylor & Russell, 1939, op. cit.

16.

Pearson, K. (1931). Tables for statisticians and biometricians: Vol. 2. London: Biometric Laboratory, University College, p. 78-109.

17.

Naylor & Shine, 1965, op. cit.

18.

That means that the new selection procedure is administered to present employees who already have been screened using other methods. The correlation (validity coefficient) between scores on the new procedure and the employees’ job performance scores is then computed.

19.

To transform raw scores into standard scores, the following formula is used: x – / SD, where x is the raw score, x-bar is the mean of the distribution of raw scores, and SD is the standard deviation of the distribution. Assuming the raw scores are distributed normally, about 99 percent of the standard scores will lie within the range –3 to + 3. Standard scores (Z-scores) are expressed in standard-deviation units.

20.

A multiple-regression coefficient represents the correlation between a criterion and two or more predictors.

21.

Boudreau, 1991, op. cit.

22.

Brogden, 1946, op. cit.

23.

Brogden, 1949, op. cit.

24.

The authors would like to thank Professor Craig J. Russell for allowing us to adapt the framework that he developed.

25.

Cronbach & Gleser, 1965, op. cit.

26.

Boudreau, 1991, op. cit.

 

..................Content has been hidden....................

You can't read the all page of ebook, please click here login for view all page.
Reset
18.191.144.194