Harnessing the Transparency of Hiring Algorithms

By: | October 1, 2018 • 3 min read
Peter Cappelli is HRE’s Talent Management columnist and a fellow of the National Academy of Human Resources. He is the George W. Taylor Professor of Management and director of the Center for Human Resources at The Wharton School of the University of Pennsylvania in Philadelphia. He can be emailed at hreletters@lrp.com.

We had an interesting meeting here at Wharton this past week that brought together a group of data-science faculty members and a group of executives who manage HR analytics. The idea was to create a research agenda and try to find an answer to the question: “What’s interesting here?”

The answer: a lot of things.

Let’s take a look at one of them: the practice of hiring, where the interest in data analytics is greatest. After World War II, it was pretty common for a job candidate at a big company to be whisked to its headquarters for a week of assessments (or what we now call pre-employment tests). These would include IQ, personality, in-basket simulations, structured interviews with psychiatrists and so forth.

What happened to all that? It disappeared in many companies, was scaled way back in others and some new companies—many of which are now huge—never bothered doing such assessments at all.


Why was that? We don’t know for sure, but I’m betting at least part of it had to do with the wave of new equal-opportunity legislation and the enforcement of it by government agencies. The watershed event here, in my view, was a series of employment-discrimination lawsuits against AT&T in the 1970s. AT&T may well have had some of the most sophisticated hiring practices at the time, but the courts held that they discriminated against protected groups.

Many companies took note of that, and the perception grew that if you were doing anything sophisticated in testing or prediction, you might well be a target for an investigation. Why? A cynical view is that it would be easier in that case to assess adverse impact: “Let’s just look at the test you were using to hire and promote and see what the impact of that test was.” If, on the other hand, you were doing nothing systematic and hiring managers were all doing something different, it was harder to see if what you were doing was discriminatory, in part, because it was just hard to see what you were doing at all.

That takes us back to data science and the current fascination with machine learning and the algorithms it generates. It is different from the statistical analysis in HR that most of us have at least a passing familiarity with, which typically starts with a hypothesis (e.g., does IQ predict who will be a good hire?) and then looks to at the relationship between the candidates’ IQ scores and their subsequent performance as new employees.

Machine learning doesn’t start with a hypothesis. It starts with a goal, which is to create a model—an algorithm—that will track as closely as possible the performance of employees. The more data we have on those employees, the better. What you get at the end of that process is an algorithm that combines the data on previous candidates in such a way that it does a very good job of figuring out who succeeded, and it can then be used to predict which of the new candidates should be hired. Hiring managers can just plug their information into the algorithm and hire the new candidates with the highest scores.