03.08.2021 Views

Annuntio#2 - 19.7.2021 - 3.8.2021

The second issue of Annuntio focuses on the pervading use of AI systems at the workplace. AI-driven technologies have proven to be beneficial to both employers and workers. However, with the rapid advance and use of AI systems, it is crucial to assess whether the benefits may germinate into risks and abuses. We raise questions on the impact of such technology on workers’ rights as well as highlighting calls and initiatives for regulation of AI use at the workplace.

The second issue of Annuntio focuses on the pervading use of AI systems at the workplace. AI-driven technologies have proven to be beneficial to both employers and workers. However, with the rapid advance and use of AI systems, it is crucial to assess whether the benefits may germinate into risks and abuses. We raise questions on the impact of such technology on workers’ rights as well as highlighting calls and initiatives for regulation of AI use at the workplace.

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.







harmless uses of AI and ADM, and the

regulation should focus primarily on ‘worker

impact’. At the mention of “high-risk” AI, it is

at this point we have to engage with the

initiative proposed by the European

Commission on the regulation of AI in April

2021 which has been dubbed as the European

Commission’s AI Law. Article 1 of the AI Law

proposes ‘specific requirements for high-risk

AI systems and obligations for operators of

such systems’. The key essence of the

proposal is the regulation of “high-risk” AI

systems. The proposal states clearly that in

classifying an AI system as high-risk, the

‘extent of the adverse impact caused by the AI

system on the fundamental rights protected

by the Charter is of particular relevance’

citing workers’ rights as one of them and AI

systems which are used that impact workers’

rights are deemed as high-risk.

The proposal justifies the position taken on

the basis that ‘AI systems used in

employment, workers management and

access to self-employment, notably for the

recruitment and selection of persons, for

making decisions on promotion and

termination and for task allocation,

monitoring or evaluation of persons in workrelated

contractual relationships, should also

be classified as high-risk, since those systems

may appreciably impact future career

prospects and livelihoods of these persons.’

In Annex III to Article 6(2), it is stipulated that

an AI system is a high-risk AI system if it is

‘intended to be used for recruitment or

selection of natural persons, notably for

advertising vacancies, screening or filtering

applications, evaluating candidates in the

course of interviews or tests’; and where it is

0 7

A N N U N T I O | I S S U E 2




Whilst the conversation on XAI began decades

ago, it saw a resurgence in recent years due to

a response by big tech organisations to a

downward trend in public trust in the sector.

Industry leaders such as Google and IBM

actively promote XAI and have developed

tools to help advance the theory and practice

of responsible and trustworthy AI.

It is important to note that while XAI is

pursued due to general concerns in

transparency and trust-related issues in AI

systems, the Spanish legislative push is also

indicative of revisiting existing legal rights in

traditional fields of law with the intention of

expanding their scope to fit new challenges.

For example, in the field of industrial

relations, especially in common law

jurisdictions, actions by an employer against

an employee are often guided by the principle

of “just cause and excuse” which would mean

that an employer should have a clear basis for

any action against an employee.

This is further expressed with the expectation

of compliance with the rules of natural justice

or the right to a fair hearing, which includes

the right to be informed of the basis of any

form of disciplinary action against them.

It could therefore be argued that new

regulations in AI explainability are merely

extensions of existing legal principles.

Notably, some countries such as Australia

have taken a different approach to regulate AI.

In a recent announcement by Australia's

government, the Minister for Digital

Economy stated that the country's AI ethics

framework will remain voluntary for the

foreseeable future.

The basis for this approach is that Australia

believes that it has sufficient existing

regulations to deal with AI which it regards as

“simply a technology”.

As the degree of explainability that may be

expected of an employer is yet to be seen,

entrenched positions of labour law could act

as the compass in navigating these new rights.

Whether an organisation utilises algorithms as

a determinant factor or merely as assistive

technology, the “right to know” found in the

rider law may be interpreted alongside the

already entrenched laws in industrial relations

thus naturally paving the way for demand in

XAI.

1 0

A N N U N T I O | I S S U E 2

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!