Hiring in the Age of AI: Hope and Fear

An overview of the potential benefits and possible drawbacks of the increasing use of AI hiring

Novin Ghaffari

Changing Job Markets

The information age has fundamentally altered many processes in society. The availability of vast amounts of information coupled with advanced data processing mechanisms has facilitated an increasing emphasis on “data-driven” strategies among organizations. Large firms and institutions are increasingly turning toward data-derived insights to inform and to support decision making. The hiring process is one domain already feeling the effects of these changes.

The hiring process can be a painful, drawn out, cumbersome, and downright expensive process. And that holds for both job seekers and job recruiters. For job hunters, the search for a decent position is time-consuming, often involving dozens, if not hundreds, of applications each filled with cycles of interviews, document drafting, aptitude and skills assessments, and negotiations. Even competent and accomplished candidates can find themselves drifting for months in search of a fitting role. For the job seekers, there is also an unfair asymmetry at play; their position during the hiring process is usually a big unknown. Recruitees are often left guessing how likely they are to land a position, how stiff the competition is, what work culture characteristics the recruiting firm is looking for, and so on. It is a typical practice for mid to large size firms to take multiple candidates through the entire interview process, knowing full well that most candidates will not be offered any position. These candidates are only being entertained as alternates and to fill hiring process quotas. The whole process can be an immense drain on a job seeker’s time, and as the old adage goes, “time is money.”

The recruitment process can also be stressful and costly on the hiring side. Finding the right candidate for an open position typically involves casting a wide net, sifting through voluminous quantities of resumes, cover letters, and applications. Giving due attention to the right candidates is time-consuming and difficult for firms without large resources to devote to the recruitment process. Hiring for a new type of role presents an added difficulty; if a firm does not have much experience in a technical area, it can be hard to assess the quality and fit of candidates for such a new type of role. And even when firms hone in on a stellar candidate, it is not unusal for top talent to be sitting on multiple job offers. If they decide their services will be more valued elsewhere, that leaves the firm having to restart their hiring process for the position all over again.

AI, Hiring, and Human Biases

With dozens, hundreds, and sometimes even thousands of applicants for a position, a manual review of all candidate applications can be labor intensive, even prohibitively so. Here hiring algorithms can help curtail costs and labor time. Things as simple as filtering based on keywords can reduce the candidate pool a hiring manager must look over. On the other end of the spectrum, approaches like AI deep learning and computer vision can be deployed for in-depth evaluation of a job candidates’ facial expression. However, the process of incorporating data science and AI into hiring practices can be fraught with hiccups and biases.

There is a notion that ‘data science and analytics’ is a neutral, unbiased mechanism for analyzing data and making data-driven decisions. There is a bit of truth here. AI and machine learning, deployed effectively, can help reduce bias. By combing through vast volumes of data, more than a human could process in a feasible amount of time, data analytics and algorithms can facilitate data-driven decision making. By deploying rigorous analytic techniques, rooted in mathematics and statistics, AI and machine learning also remove some of the inherent biases of relying solely on human opinion and intuition. Nonetheless there is still arbitrariness in the choice of algorithm and in the choice of data processing and modeling methodologies. To assume that data analytics can completely remove all biases is a mistake. As Gideon Mann and Cathy O’Neil observe writing for the Harvard Business Review [1],

“Algorithms are, in part, our opinions embedded in code. They reflect human biases and prejudices that lead to machine learning mistakes and misinterpretations. This bias shows up in numerous aspects of our lives, including algorithms used for electronic discovery, teacher evaluations, car insurance, credit score rankings, and university admissions.”

With all of the benefits and detriments, big data and AI is revolutionizing the modern hiring process. A 2018 survey of over 8800 hiring managers conducted by LinkedIn revealed the following trends. A total of 67% of hiring managers claimed to use AI in hiring, claiming it helped reduce biases (43%), deliver the best candidate matches (31%). Respondents claimed AI Is helpful for sourcing candidates (58%), screening candidates (56%), and nurturing candidates (55%). Furthermore, hiring managers pointed out that big data and talent intelligence from past hires has been used to increase retention (56%), evaluate skill gaps (50%), and build better offers (50%) [2].

Many hiring and candidate selection algorithms are trained on past hiring data. The idea: by identifying the data patterns of successful past hires’ applications and searching for similar patterns in current job applicants’ application, a firm can replicate past hiring successes. While this seems like a good strategy, it does suffer from a major pitfall: training on past hire data, an algorithm not only learns past hiring patterns, but also all the biases associated with past hires. For instance, Marianne Bertrand and Sendhil Mullainathan in a study for the National Bureau of Economic Research, found that resumes with white names were 50% more likely to receive callbacks as compared to resumes with African-American names. Furthermore, a high-quality resume with a white-sounding name elicited 30% more callbacks, whereas increases for a high-quality resume with an African American name only showed marginal increases in callbacks. Furthermore, these patterns of biases and discrimination were observed across the board, being equally present with employers who listed ‘Equal Opportunity Employer’ in their job ads [3].

Indeed, the reckless, unscrupulous deployment of AI algorithms on past hiring data means that the learning algorithm will likely train on and learn all the past mistakes, prejudices, and biases in the hiring process. The next case study is one such case.

A Big Data Hiring Failure: The Amazon Case

A naïve hiring algorithm that solely and blindly learns from past hiring decisions would likely replicate these biases moving forward. An example of this was automated job applicant filtering program deployed at Amazon. In 2014, with over half a million employees and growing, Amazon decided to use an algorithmic approach to improve its hiring efficiency. AI algorithms were devised at Amazon’s Edinburgh engineering hub to recognize over 50,000 resume terms, ranking job applicants from one to five stars. Training the algorithms on past hiring data, the AI picked up on and replicated a major facet of Amazon’s past hiring culture: male dominance. The algorithm implicitly upvoted resumes using language cues typically favored by male candidates. In fact, even if Amazon’s past hiring was fair in terms of gender, an AI algorithm can still learn unfair biases for future hiring decisions. That is, even if the best candidates for past positions were indeed mostly male, that does not mean that the best candidate for the next open position will conform to the same profile.

In addition to replicating gender bias, the complexity of the challenge led the hiring algorithms to select random, unqualified candidates on many occasions. After the discovery of these shortcomings, the project was scraped. Recently Amazon has revisited and revamped its automated hiring strategy, using AI for more straightforward tasks (e.g., removing duplicate applications) and for increasing diversity hires [4].

Humantic and HireVue: More Case Studies

Some the most basic and fundamental analytical techniques used to assist in the hiring process are screening and filtering algorithms. In the simplest scenarios, these methods simply screen resumes looking for keywords that indicate good fit and qualification for a role. Naïve implementations of screening systems may cause employers to miss great candidates. However, even naïve implementations can save considerable time and resources. Some job positions receive thousands of applicants; when manually reviewed, each resume is glanced at for an average of a few seconds. Here screening can process more data in those few seconds than a human can manually consider. Adding a little sophistication in the form of some background analytics can provide even more information in little additional time. On the other end of the screening spectrum, Humantic claims to have a system that data mines candidates’ resumes. Using text analysis and psycholinguistics, the service is able to profile the personality of candidates without the need for additional testing. The service is used by more than 100 companies. As another example of the power of analytics, in a few milliseconds an AI system can calculate the approximate commute time based on a candidate’s location [5].

A striking example of AI in hiring is a system that goes far beyond simple filtering of resume terms. HireVue, a recruiting tech firm, has a developed an AI system that uses candidates web or mobile camera and microphone to assess candidates’ eye movements, facial patterns, diction, and tone of voice. These facial movements are used to assess a job candidate’s excitement, and they constitute 29% of a candidate’s score. Based on these measurements, candidates are ranked against each other and given an ‘employability’ score. HireVue claims its AI can hone in on cues beyond just word choice on a resume. Among other things, HireVue claims that its AI systems are less prone to bias than human recruiters. They argue that AI can pick up on tiny signals that might be missed by human recruiter. A standard 30-minute HireVue assessment can generate as many as 500,000 data points. HireVue’s systems are currently used by many major employers, including Hilton and Unilever. Hilton has touted HireVue’s tech’s ability to reduce the average processing time for thousands of job applicants from 6 weeks to 5 days. Unilever claims HireVue’s systems have helped it reduce hiring labor hours by 100,000 and recruitment costs by $1 million [6].

Nonetheless, HireVue’s AI system has drawn a volley of criticism. One neuroscientist, Lisa Feldman Barret, led a meta-study reviewing 1000 published papers on computer vision and human emotion. Barret’s group concluded that these AI systems are highly effective at detecting and differentiating facial expressions, but not at signifying what they mean. Indeed, the significance of specific facial expressions are unique to different individuals, can vary by context, and vary widely in different cultures. Barrett relates the example of a frown, which may be taken as a cue of anger, hence a red flag for sales-type jobs. Nonetheless, a frown may also be indicative of intense focus, a potentially positive indicator of a hard work ethic. HireVue has countered these criticisms by claiming it has on-staff psychologists that work to tailor the scoring algorithm of their AI system to each client’s needs. Simon Mondal, the CEO of a job candidate screening platform, Ideal, which screens 5 million candidates a month, points out that these AI systems must be continually revisited and updated to prevent the encroachment of biases. Patterns always change, and the same is true in the field of hiring prospective candidates. Ultimately, a satisfactory AI screening system should be able to adjust its learning, and this generally requires human oversight. As Mondal points out, “You have to have a way to make sure that you aren’t picking people who are grouped in a specific way and that you’re only hiring those types of people” [5], [6].

Potential concerns with the deployment of AI in hiring practices has even led to protective legislation. In the UK, a poll commissioned by the Royal Arts Society found that 60% of the public opposes the use of automated systems for hiring. Still, a YouGov poll found that only 32% are aware, in general, of the use of AI in decision-making, while only 14% are aware of its increasing adoption in recruitment and only 9% of its use in criminal justice [7].

Public concern has begun converting into legislative action. In Illinois, state representative Jaime Andrade Jr. (D) co-sponsored “Public Act 101-0260,” a bill requiring employers in the state to notify job applicants of how their AI hiring systems work and to get the job applicant’s consent. Governor J.B. Pritzker signed the bill into law in August of 2019, and the bill went into law January of 2020 [6].

In November of 2019, lawyers at the Electronic Privacy Information Center (EPIC) filed a complaint with the FTC against HireVue on the grounds of potential bias, inaccuracy, and lacking transparency. The lawsuit also pointed to possible deceptive trade practices; HireVue claims it does not use facial recognition while EPIC contends its facial analysis constitutes facial recognition. The lawsuit came on the heels of congressional legislation that grants the FTC powers to investigate and regulate “automated decision systems” [5].

Indeed, the rise of AI poses some challenges for antidiscrimination measures; in the worst case, a firm may attempt to hide biased recruiting policies behind algorithmic complexity. Already the Equal Employment Opportunities Commission (EEOC) has investigated several cases of alleged, AI-driven discrimination [8].

The hiring process can be complicated, painstaking, and time-consuming. The use of AI and machine learning offers a tremendous resource to manage and mitigate this complexity. As with all other technology, it brings its own share of promises and threats. The ultimate utility of this tech will largely be driven by how humans choose to adopt, use, and regulate it. Until people figure out how to relate AI and machine learning to hiring in an open, constructive way, the prospect of AI-led hiring will continue creating hope and fear.

References

  1. Hiring Algorithms Are Not Neutral  [link]
    Mann, G. and O’Neil, C., 2016.
  2. LinkedIn 2018 Report Highlights Top Global Trends in Recruiting  [link]
    2018.
  3. Are Emily and Greg More Employable Than Lakisha and Jamal? A Field Experiment on Labor Market Discrimination  [link]
    Bertrand, M. and Mullainathan, S., 2003.
  4. Amazon scraps secret AI recruiting tool that showed bias against women  [link]
    Dastin, J., 2018.
  5. Artificial Intelligence Will Help Determine if You Get Your Next Job  [link]
    Heilweil, R., 2019.
  6. A face-scanning algorithm increasingly decides whether you deserve the job  [link]
    Harwell, D., 2019.
  7. Unilever saves on recruiters by using AI to assess job interviews  [link]
    Booth, R., 2019.
  8. Punching In: Workplace Bias Police Look at Hiring Algorithms  [link]
    Opfer, C., Penn, B. and Diaz, J., 2019.

One response to “Hiring in the Age of AI: Hope and Fear”