“We must address, individually and collectively, moral and ethical issues raised by cutting-edge research in artificial intelligence and biotechnology, which will enable significant life extension, designer babies, and memory extraction.”
—Klaus Schwab
Merging recruitment with technological innovation always comes with potential legal risks, and this can occur parallel to the time during which employers are looking for faster, more accurate, and cost-effective recruitment platforms that leverage the latest technological innovations to streamlines the hiring process through artificial intelligence (AI) tools.
Employers are turning to AI to transform recruitment and generate a seamless hiring process. This includes automating the candidate sourcing, candidate pool screening, and using AI assessment tools, such as conversational chatbots and video interviewing tools that can measure a candidate’s strengths based on factors such as facial expression, word choice, body language, and vocal tone. However, the use of artificial intelligence (AI) and other automated decision-making (ADM) technology doesn’t come without risk. Employers should tread carefully while implementing such HR Tech solutions.
In this article, we examine the legal considerations and safeguards that are currently being undertaken by the United States of America, Europe, and Singapore.
Legal considerations are generally rooted in two areas of law:
Both these laws are heavily disaggregated civil law considerations.
The two key areas of the Privacy Law to keep in mind:
The EEOC has made clear that employers using AI in their hiring process can be liable for unintended discrimination, and AI vendors regularly include non-liability clauses in their contracts with employers. Therefore, employers need to validate AI tools and take steps to ensure that they do not cause inadvertent discrimination when hiring. Employers should test the capabilities of the AI
Although AI has not yet been federally regulated for adoption, Illinois has just passed the first law of its kind, called the Artificial Intelligence Video Interview Act. Beginning January 1, 2020, the law requires employers to analyze candidate video interviews using AI to:
EU officials stated that AI technology needs proactive regulation now, as it may become difficult to regulate AI later due to the rapid advances in the technology, and insisted on finding a balance between reasonable, commercial, and operational interests of companies, and privacy and anti-discrimination rights of employees.
The Commission proposes to ban completely AI systems that:
Singapore introduced its Model Artificial Intelligence Governance Framework in January 2019 at the World Economic Forum (WEF) in Davos. The two guiding principles of the framework state that decisions made by AI should be “explainable, transparent and fair”; and AI systems should be human-centric. These principles are then developed into four areas of guidance.
The framework translates ethical principles into pragmatic measures that businesses can do.
Why you can trust impress.ai
The European Union (EU) General Data Protection Regulation (GDPR) is a set of industry regulation that became effective on May 25th 2018. The purpose of the legislation was to give EU citizens greater control over the data that they provide online. GDPR covers companies that are operating within the EU and for companies that offer services within the European Union electronically, that track / store personal data in aggregate. With impress.ai operating primarily in non-EU jurisdictions, it provides both GDPR compliant and non-GDPR compliant versions of impress.ai’s’ recruitment automation Software-as-a-Service. Hiring companies, that are clients of impress.ai can require impress ai’s SaaS to be GDPR compliant as part of the service agreement.
Thanks for your interest! We'll get back to you soon
A unified AI platform constructed for recruiters, employers, businesses and people
GET IN TOUCH