Rotterdam school of Management, Erasmus University compact logo

Article: Thursday 19 February

When artificial intelligence is introduced at work, the biggest decisions are often already made. By the time workers feel the effects on schedules or job content, the technology can be hard to reverse. The Dockers’ AI Toolkit, developed for the International Transport Workers’ Federation (ITF) by José Luis Gallegos, PhD candidate at Rotterdam School of Management and visiting fellow at Cornell University, is designed to intervene much earlier.

The toolkit offers practical guidance for employees, labour advocates, and particularly trade union members, on how to negotiate the use of AI before it reshapes work. Although rooted in the port sector, its relevance extends beyond dock work: AI is increasingly used to allocate tasks, monitor workers and support organisational decisions. “AI at work should not be something workers are forced to adapt to after the fact,” Gallegos says. “It should be something they help govern from the outset.”

From automation to authority

Ports have long been testing grounds for new technologies, from containerisation to automated cranes and digital planning systems. In his PhD research, Gallegos examined earlier waves of automation at the Port of Rotterdam and observed how unions played a decisive role in ensuring that the introduction of automation did not result in increased inequality or job displacement. AI, he argues, raises the stakes further because it also reshapes decision-making. It influences planning, performance evaluation, and workflow coordination, affecting office-based roles as well as work on the terminal floor.

A key strength of the toolkit is that it looks at AI as a whole, rather than as a series of separate problems. Instead of dealing with job loss, surveillance, bias, and data use one by one, it brings them together in a single framework. That matters because the same AI system can be used in very different ways. It can be set up to improve safety, cut costs, increase efficiency, or tighten control over workers.

These decisions are typically made early—when systems are procured or designed—yet worker representatives are often not involved at that stage. The toolkit therefore concentrates on intervening while objectives, boundaries, and accountability structures are still open for discussion. The toolkit breaks negotiations into clear steps and includes model clauses that unions can adapt for collective agreements. “Negotiating AI with worker representatives is not about blocking or slowing technological change,” Gallegos emphasises. “It is about collectively setting the rules under which it operates.”

AI at work should not be something workers are forced to adapt to after the fact. It should be something they help govern from the outset.

Beyond consultation

The guide is clear that simply informing workers after decisions have been made is not enough. When AI changes how work is organised, late consultation offers little real protection. Instead, the toolkit argues for firm rights that give worker representatives a real say both before and after AI is introduced. This includes the right to approve or reject high-risk systems and the ability to bring in independent experts who can assess technical, legal or ethical risks. The toolkit also calls for clear limits. Certain uses of AI should not be allowed at all, such as biometric surveillance, emotion-recognition tools or fully automated decisions that affect people’s jobs.

Governing AI over time

AI systems do not stay the same once they are introduced. They learn from new data, are updated by suppliers and are often used in more ways than first intended. To deal with this, the toolkit proposes setting up joint technology review committees, with equal representation from management and workers. These committees would regularly review the effects of AI systems, keep an eye on how they are used in practice and arrange checks when problems arise.

When data becomes labour

Another key contribution of the toolkit concerns worker data. Every time workers adjust a system or correct an error, they generate data that can be used to train AI. Over time, this data reflects their experience and practical know-how. As the toolkit notes, workers are “hired to move boxes, not to train AI”. Yet their knowledge can end up feeding systems that later deskill or replace parts of their work. For this reason, the guide treats data as a form of labour and explores ways workers could be compensated, for example through wage premiums, agreements on how data may be used, or jointly managed transition funds.

What this means in practice

Although developed for dockworkers’ unions, the toolkit offers broader lessons for organisations introducing AI at work. This is especially relevant as new regulatory frameworks in Europe, including the EU AI Act, and similar initiatives elsewhere begin to clarify responsibilities for organisations deploying AI in employment settings. In this context, AI implementation is increasingly understood not only as a technical decision, but also as one that carries organisational and governance implications.

The toolkit highlights several practical recommendations:

Involve worker representatives early, before procurement or system design decisions are fixed
Define in writing what AI systems may and may not be used for
Institutionalise human oversight and clear accountability for AI-driven decisions
Link AI adoption to job protection and transition measures
J.L. (José Luis) Gallegos
PhD Candidate
Rotterdam School of Management (RSM)
Erasmus University Rotterdam
Photo
José Luis Gallegos
Pile of books with vibrant bookmarks protruding from various pages, symbolizing in-depth research.

Related articles

RSM Discovery

Want to elevate your business to the next level using the latest research? RSM Discovery is your online research platform. Read the latest insights from the best researchers in the field of business. You can also subscribe to the newsletter to receive a bimonthly highlight with the most popular articles.

Do you want to learn more about this subject?

Check out these RSM education programmes

Your contact for more information:
Danielle Baan

Science Communication and Media Officer

Erasmus University campus in autumn, showcasing its iconic red trees, viewed from across the campus pool.