• Sectors we work in banner(2)

    Quick Reads

AI and HR - How can employers reduce the risks associated with using artificial intelligence to help manage their workforce?

The term "Artificial Intelligence" may seem intrinsically linked to the world of the future, conjuring images of evil robots, intent to destroy humanity, or humanoid beings, eerily similar to you or me.  In fact, AI is here today, playing a huge part in the world around us.  AI is any electronic device or system which can solve the kind of complex problems we would usually associate with human intelligence, through machine-learning algorithms (an algorithm is essentially a set of instructions for a computer to follow).  This includes anything from Siri and Alexa to a self-driving Tesla, Netflix's recommendations and personalised thumbnails to Google's sense of direction.  

Like most new technologies, AI can make our lives easier and allow us to achieve more.  It's therefore no wonder that organisations are beginning to utilise AI in all aspects of employment and people management.  People Management recently highlighted the growing use of AI in recruitment, where AI can help by automating candidate screening, scheduling interviews and maintaining candidate databases.  It can perform administrative tasks, including managing holiday entitlement, absences and performance data and can even support talent management, by predicting when an employee might leave.  

It's easy to see why employers are keen to utilise AI.  However, the first adopters of any new technologies can face risks and challenges, particularly if they do not understand all of the implications of the systems they are using.  So, what risks should employers be alive to?

I remember reading about an Uber Eats driver who was dismissed by the company's algorithm last year.  Uber Eats was using facial recognition software (a type of AI) in an attempt to stop people other than drivers using an account.  The software failed to recognise one of its drivers and he was dismissed as a result.  Aside from the issues this raises about the gig economy (which we will have to leave for another day), it highlights some of the potential pitfalls in using AI

Facial recognition software is notoriously worse at recognising black people than white people and worse at recognising women than men.  The dismissed Uber Eats driver said the software used by his employer was "racially discriminatory and should be abolished until perfected".   Use of such software, without careful controls, could not only lead to discrimination claims but could also cause enormous reputational damage.  Employers should tread carefully and ensure that they put proper controls in place to double check any decisions made by AI.

AI software runs on algorithms which are trained on large data sets.  The algorithm uses the data to "learn" and improve.  This means that, if you feed AI a biased data set, you get a biased outcome.  Amazon learnt this lesson the hard way in 2014 when it rolled out a new experimental hiring tool, designed to sift through applicants and narrow down the list of potential new recruits.  The AI was trained on existing recruitment data, looking at the patterns contained in CVs submitted by successful and unsuccessful candidates over the previous 10 years .  Therefore, the AI was taught to recruit exactly the sort of people that Amazon had recruited previously, which might not have been a bad thing were it not for the unfortunate fact that Amazon had mainly hired men over those 10 years. The algorithm learned that men are more desirable candidates than women, penalising any applicant who included words such as "women's" in their CV. 

One may think that Amazon could have dealt with the problem by removing gendered language from the CVs before the AI got involved.  However, to date, this has not proved possible. Research has shown that it's essentially impossible to completely hide gender during the recruitment process; AI can identify when an applicant is male or female even where a human being or, indeed, other AI has supposedly de-gendered the CV.  

Amazon scrapped the tech in 2018. Clearly, the use of recruitment AI which exacerbates existing biases is detrimental to diversity and could open employers up to claims for discrimination.

Aside from the discrimination risks, employers are also very often under an obligation to provide explanations to employees for certain decisions, as part of the implied term of trust and confidence between an employer and employee.  How can an employer, who has blindly relied on the results of an algorithm that it does not understand, ever hope to provide such an explanation?

It is therefore vital that employers conduct their due diligence and have some understanding of the tools they are using - What is the purpose of the tool? What characteristics is it assessing?  How is the algorithm being used to make decisions?

Alongside the discrimination risks, there are potential data protection pitfalls.  UK GDPR limits the circumstances in which employers can make solely automated decisions and requires transparency where such decisions are made.  A "solely automated decision" is a decision where there has been no human influence on the outcome.  For example, if an employer's clocking-in system automatically sends a warning to an employee about punctuality, that would be a decision taken solely by automated means.  However, if the system instead sends a flag to an HR manager, who takes the decision to issue a warning, that decision is not solely automated.  

GDPR says that automated decisions are unlawful if they have a "legal effect" or a "similarly significant effect" on the data subject.  Hiring and firing would definitely fall under this umbrella, meaning employers are limited from relying on fully automated decisions in this regard.  There is an exception where automated decision making is "necessary" to enter into a contract, but this is poorly defined.  It is somewhat tricky to see where some level of human intervention would not be possible for employment purposes.  

Even where decisions are not fully automated, use of AI could potentially lead to other data protection issues.  Have a think about the incredibly broad definition of personal data - "any information relating to an identified or identifiable natural person".  Imagine that my employer is training new AI software on its workforce data.   Say that I am the only person named Briony who has ever worked for my employer but one day another Briony applies and the AI software makes certain assumptions about that other Briony, based on what it learnt about me.  Are those assumptions my personal data?  If the AI says "don't hire this new Briony, she is probably a wrongun'", what does that tell you about me and does that amount to my personal data?  

This might be an overly simplistic example - I doubt that anyone would code AI software to make recommendations based on a single data point - but there really are potential issues here and the answers are not always clear cut.  Employers will need to think carefully about how and why they use employee personal data, ensuring they have a legal basis for each processing operation and are fair and transparent about their actions. 

Ultimately, this is an exciting area and not one to fear.  AI can help to make our lives easier and allow us to make better, more informed decisions.  However, it can only do that if we understand how it works and its limitations and, vitally, consider when it is necessary for an actual human being to step in.  

In 2019, Unilever reported that its use of AI had saved its human recruiters approximately 100,000 hours in interviewing time and nearly £1m per year. However, an overreliance on AI when making recruitment decisions case can see employers easily being wrong-footed and inadvertently breaching UK data protection and anti-discrimination laws.

Our thinking

  • Something Changed – Landlord recovers possession of iconic music venue

    Samuel Lear

    Quick Reads

  • Kevin Gibbs and Sadie Pitman write for CoStar on the need for investment in power infrastructure to support new data centres

    Kevin Gibbs

    In the Press

  • New code of practice for the cyber security of AI development

    Rebecca Steer

    Quick Reads

  • The Times and Daily Mail quote Dan Pollard on new changes to the Employment Rights Bill

    Dan Pollard

    In the Press

  • The Good, the Bad and the Ugly - the inheritance tax Consultation on agricultural and business property

    Sarah Wray

    Quick Reads

  • A Ray of Light for Developers - High Court provides some comfort in recent injunction case

    Georgina Muskett

    Insights

  • Sarah Jane Boon and Julia Cox write for Tax Adviser on safeguarding family wealth and the role of pre- and post-nuptial agreements

    Sarah Jane Boon

    In the Press

  • Stepping into the Director's Chair: The Landscape of Risk in Distressed Companies – Misfeasance Trading

    Jessica Boxford

    Insights

  • ESMA Consultation on Guidelines for the criteria to assess knowledge and competence under MiCA

    Charlotte Hill

    Insights

  • EU AI Act: Key provisions now in force

    Racheal Muldoon

    Insights

  • The FCA’s requirements for Payments Firms

    Charlotte Hill

    Insights

  • Mahmood v Standard Chartered Bank – A landmark decision in discrimination and victimisation but what does it mean for discrimination claims in the DIFC?

    Nick Hurley

    Insights

  • Digital Securities Sandbox Update

    Racheal Muldoon

    Insights

  • Property Patter: Challenges for commercial property in 2025

    Emma Humphreys

    Podcasts

  • Naomi Nettleton and Alicia Stanford-Shard write for Property Week on how landed estate owners are increasingly looking to make improvements to their estates to enhance environmental impact and efficiency

    Naomi Nettleton

    In the Press

  • HR Magazine quotes Lucy Lintott on employees who are unable to work due to poor health or disability

    Lucy Lintott

    In the Press

  • Rivals: Filming Locations and Considerations for Landed Estates

    Naomi Nettleton

    Insights

  • New food and drink ads regulation & impact on live sports broadcasts

    Sarah Johnson

    Insights

  • Charles Russell Speechlys advises Puma Growth Partners on its lead investment as part of a $4.3 million funding round for finance-focused legal AI specialist, Semeris

    David Coates

    News

  • A Digital Pound: The Bank of England January 2025 Update

    Racheal Muldoon

    Insights

Back to top