Skip to content

Brought to you by

Dentons logo in black and white

UK People Reward and Mobility Hub

The latest updates in employment, benefits, pensions and immigration

open menu close menu

UK People Reward and Mobility Hub

  • Home
  • Events and training
  • Who We Are
    • Meet the team
  • How we can help

ICO spotlights fairness in AI recruitment: practical points for employers

By Elouisa Crichton
April 9, 2026
  • Artificial intelligence
  • Data protection
  • Equality Act
  • Privacy
Share on Facebook Share on Twitter Share via email Share on LinkedIn

The Information Commissioner’s Office (ICO) has launched a consultation on updates to its guidance on automated decision-making under UK data protection law. While the consultation itself is relatively modest in scope, the ICO has also published a report on the fair and responsible use of automation in recruitment.

Taken together, these developments give a helpful indication of how the ICO is approaching AI in employment contexts, particularly where employers use automated tools to support or inform hiring decisions.

A brief update on the consultation

The consultation focuses on clarifying existing ICO guidance, including how the legislation on automated decision-making applies in practice and what constitutes “meaningful human involvement” in this context.

The ICO is not proposing a new framework. Instead, it underlines a point that often proves difficult in practice: human oversight needs to be genuine and capable of affecting the outcome, rather than a purely procedural step.

The more interesting development: recruitment and AI report

The accompanying report looks more closely at how employers are using automation in recruitment and where risks tend to arise. Drawing on real-world practices, it highlights the gap between what technology can do and how employers are using it day to day.

A consistent theme is that many of the underlying issues are not new, but AI brings them into sharper focus. Bias, for example, is not unique to automated systems, but tools that are not properly tested can reinforce patterns at scale. For instance, training a tool on historical hiring data risks embedding disadvantage if all the previously successful candidates belong to a narrow demographic profile.

Similarly, a lack of transparency has long been a feature of some recruitment processes, but this becomes more acute where decisions are supported by more complex, automated systems. The report includes a case study where the employer has a detailed privacy policy explaining candidates’ rights in relation to the use of automated decision-making, but does not include meaningful information about the logic the tool uses or its potential impact on outcomes.

The ICO also identifies a risk of overreliance on automated outputs. Decision-makers may defer too readily to system-generated recommendations, particularly where the underlying logic is not well understood.

What the ICO expects from employers

The report points to a need for clearer internal ownership of AI systems, supported by appropriate testing and ongoing monitoring. Employers should be able to explain how they use these tools, what data they rely upon, and how they identify and manage risks, particularly around bias.

Transparency is another area of focus. Candidates should not be left uncertain about whether the employer is using AI or how it affects their application. This requires clear and accessible explanations, not just high-level privacy information.

The ICO also emphasises that data quality and relevance remain central. Using more data does not necessarily lead to better outcomes, particularly where it is not closely aligned to the role.

Implications for employers

For employers already using AI in recruitment, this is a useful prompt to review how you are using AI tools within recruitment processes, particularly at key decision points such as screening and shortlisting. Often, the more useful question is not whether you are using a tool, but whether you understand and question its outputs.

Organisations may also wish to consider how they can clearly explain their use of AI to a candidate, a regulator or a tribunal, including how decisions are reached and who is accountable.

There is also a clear overlap with equality law risk. Even where a system meets data protection requirements, its use may still give rise to discrimination issues if outcomes are not properly tested and understood.

Looking ahead

While the consultation may not signal a major shift, the overall message and direction of travel is clear. The emphasis is now on how employers use AI tools in recruitment and the topic is firmly on the regulatory agenda, with increasing focus on governance and fairness. Employers should be prepared for scrutiny from regulators, candidates and, potentially, tribunals.

Share on Facebook Share on Twitter Share via email Share on LinkedIn
Subscribe and stay updated
Receive our latest blog posts by email.
Stay in Touch
Artificial intelligence, Data Protection, Equality Act, Privacy
Elouisa Crichton

About Elouisa Crichton

All posts Full bio

You might also like...

  • Data protection
  • Employment documents

Data protection: consent and the employment contract

The European Parliament has now passed the new General Data Protection Regulation. This will come into effect in 2018. The […]

By Victoria Albon
  • Data protection
  • GDPR

New ICO guidance on data subject access requests for employers

By Claire Maclean
  • Discrimination
  • Equality Act
  • Legislative changes
  • Redundancy and business reorganisation
  • Sex discrimination

Redundancy – Pregnancy and Maternity

By Laura Morrison

About Dentons

Redefining possibilities. Together, everywhere. For more information visit dentons.com

Grow, Protect, Operate, Finance. Dentons, the law firm of the future is here. Copyright 2023 Dentons. Dentons is a global legal practice providing client services worldwide through its member firms and affiliates. Please see dentons.com for Legal notices.

Categories

Dentons logo in black and white

© 2026 Dentons

  • Legal notices
  • Privacy policy
  • Terms of use
  • Cookies on this site