Professional Documents
Culture Documents
The Exploited Labor Behind Artificial Intelligence
The Exploited Labor Behind Artificial Intelligence
Behind Artificial
Intelligence
Supporting transnational worker
organizing should be at the center of the
fight for “ethical AI.”
Adrienne Williams and Milagros Miceli are researchers at the Distributed AI Research
(DAIR) Institute. Timnit Gebru is the institute’s founder and executive director. She
was previously co-lead of the Ethical AI research team at Google.
Tech companies that have branded themselves “AI first” depend on heavily
surveilled gig workers like data labelers, delivery drivers and content
moderators. Startups are even hiring people to impersonate AI systems like
chatbots, due to the pressure by venture capitalists to incorporate so-called AI
into their products. In fact, London-based venture capital firm MMC Ventures
surveyed 2,830 AI startups in the EU and found that 40% of them didn’t use
AI in a meaningful way.
Far from the sophisticated, sentient machines portrayed in media and pop
culture, so-called AI systems are fueled by millions of underpaid workers
around the world, performing repetitive tasks under precarious labor
conditions. And unlike the “AI researchers” paid six-figure salaries in Silicon
Valley corporations, these exploited workers are often recruited out of
impoverished populations and paid as little as $1.46/hour after tax. Yet despite
this, labor exploitation is not central to the discourse surrounding the ethical
development and deployment of AI systems. In this article, we give examples
of the labor exploitation driving so-called AI systems and argue that
supporting transnational worker organizing efforts should be a priority in
discussions pertaining to AI ethics.
In 2009, however, Jia Deng and his collaborators released the ImageNet
dataset, the largest labeled image dataset at the time, consisting of images
scraped from the internet and labeled through Amazon’s newly introduced
Mechanical Turk platform. Amazon Mechanical Turk, with the motto “artificial
artificial intelligence,” popularized the phenomenon of “crowd work”: large
volumes of time-consuming work broken down into smaller tasks that can
quickly be completed by millions of people around the world. With the
introduction of Mechanical Turk, intractable tasks were suddenly made
feasible; for example, hand-labeling one million images could be automatically
executed by a thousand anonymous people working in parallel, each labeling
only a thousand images. What’s more, it was at a price even a university could
afford: crowdworkers were paid per task completed, which could amount to
merely a few cents.
The ImageNet dataset was followed by the ImageNet Large Scale Visual
Recognition Challenge, where researchers used the dataset to train and test
models performing a variety of tasks like image recognition: annotating an
image with the type of object in the image, such as a tree or a cat. While non-
deep-learning-based models performed these tasks with the highest accuracy
at the time, in 2012, a deep-learning-based architecture informally dubbed
AlexNet scored higher than all other models by a wide margin. This catapulted
deep-learning-based models into the mainstream, and brought us to today,
where models requiring lots of data, labeled by low-wage gig workers around
the world, are proliferated by multinational corporations. In addition to
labeling data scraped from the internet, some jobs have gig workers supply the
data itself, requiring them to upload selfies, pictures of friends and family or
images of the objects around them.
Content moderators, for example, are responsible for finding and flagging
content deemed inappropriate for a given platform. Not only are they essential
workers, without whom social media platforms would be completely unusable,
their work flagging different types of content is also used to train automated
systems aiming to flag texts and imagery containing hate speech, fake news,
violence or other types of content that violates platforms’ policies. In spite of
the crucial role that content moderators play in both keeping online
communities safe and training AI systems, they are often paid miserable wages
while working for tech giants and forced to perform traumatic tasks while
being closely surveilled.
Every murder, suicide, sexual assault or child abuse video that does not make
it onto a platform has been viewed and flagged by a content moderator or an
automated system trained by data most likely supplied by a content
moderator. Employees performing these tasks suffer from anxiety, depression
and post-traumatic stress disorder due to constant exposure to this horrific
content.
Similar to social media platforms which would not function without content
moderators, e-commerce conglomerates like Amazon are run by armies of
warehouse workers and delivery drivers, among others. Like content
moderators, these workers both keep the platforms functional and supply data
for AI systems that Amazon may one day use to replace them: robots that stock
packages in warehouses and self-driving cars that deliver these packages to
customers. In the meantime, these workers must perform repetitive tasks
under the pressure of constant surveillance — tasks that, at times, put their
lives at risk and often result in serious musculoskeletal injuries.
Amazon warehouse employees are tracked via cameras and their inventory
scanners, and their performance is measured against the times managers
determine every task should take, based on aggregate data from everyone
working at the same facility. Time away from their assigned tasks is tracked
and used to discipline workers.
Like warehouse workers, Amazon delivery drivers are also monitored through
automated surveillance systems: an app called Mentor tallies scores based on
so-called violations. Amazon’s unrealistic delivery time expectations push
many drivers to take risky measures to ensure that they deliver the number of
packages assigned to them for the day. For instance, the time it takes someone
to fasten and unfasten their seatbelt some 90-300 times a day is enough to put
them behind schedule on their route. Adrienne and many of her colleagues
buckled their seat belts behind their backs, so that the surveillance systems
registered that they were driving with a belt on, without getting slowed down
by actually driving with a belt on.
In 2020, Amazon drivers in the U.S. were injured at a nearly 50% higher rate
than their United Parcel Service counterparts. In 2021, Amazon drivers were
injured at a rate of 18.3 per 100 drivers, up nearly 40% from the previous year.
These conditions aren’t only dangerous for delivery drivers — pedestrians and
car passengers have been killed and injured in accidents involving Amazon
delivery drivers. Some drivers in Japan recently quit in protest because they
say Amazon’s software sent them on “impossible routes,” leading to
“unreasonable demands and long hours.” In spite of these clear harms,
however, Amazon continues to treat its workers like machines.
In addition to tracking its workers through scanners and cameras, last year,
the company required delivery drivers in the U.S. to sign a “biometric consent”
form, granting Amazon permission to use AI-powered cameras to monitor
drivers’ movements — supposedly to cut down on distracted driving or
speeding and ensure seatbelt usage. It’s only reasonable for workers to fear
that facial recognition and other biometric data could be used to perfect
worker-surveillance tools or further train AI — which could one day replace
them. The vague wording in the consent forms leaves the precise purpose open
for interpretation, and workers have suspected unwanted uses of their data
before (though Amazon denied it).
The “AI” industry runs on the backs of these low-wage workers, who are kept
in precarious positions, making it hard, in the absence of unionization, to push
back on unethical practices or demand better working conditions for fear of
losing jobs they can’t afford to lose. Companies make sure to hire people from
poor and underserved communities, such as refugees, incarcerated people and
others with few job options, often hiring them through third party firms as
contractors rather than as full time employees. While more employers should
hire from vulnerable groups like these, it is unacceptable to do it in a predatory
manner, with no protections.
Data labeling jobs are often performed far from the Silicon Valley headquarters
of “AI first” multinational corporations — from Venezuela, where workers label
data for the image recognition systems in self-driving vehicles, to Bulgaria,
where Syrian refugees fuel facial recognition systems with selfies labeled
according to race, gender, and age categories. These tasks are often outsourced
to precarious workers in countries like India, Kenya, the Philippines or Mexico.
Workers often do not speak English but are provided instructions in English,
and face termination or banning from crowdwork platforms if they do not fully
understand the rules.
These corporations know that increased worker power would slow down their
march toward proliferating “AI” systems requiring vast amounts of data,
deployed without adequately studying and mitigating their harms. Talk of
sentient machines only distracts us from holding them accountable for the
exploitative labor practices that power the “AI” industry.
Thus, we advocate for funding of research and public initiatives that aim to
uncover issues at the intersection of labor and AI systems. AI ethics
researchers should analyze harmful AI systems as both causes and
consequences of unjust labor conditions in the industry. Researchers and
practitioners in AI should reflect on their use of crowdworkers to advance their
own careers, while the crowdworkers remain in precarious conditions. Instead,
the AI ethics community should work on initiatives that shift power into the
hands of workers. Examples include co-creating research agendas with
workers based on their needs, supporting cross-geographical labor organizing
efforts and ensuring that research findings are easily accessed by workers
rather than confined to academic publications. The Turkopticon platform
created by Lilly Irani and M. Six Silberman, “an activist system that allows
workers to publicize and evaluate their relationships with employers,” is a
great example of this.
Journalists, artists, and scientists can help by drawing clear the connection
between labor exploitation and harmful AI products in our everyday lives,
fostering solidarity with and support for gig workers and other vulnerable
worker populations. Journalists and commentators can show the general
public why they should care about the data annotator in Syria or the
hypersurveilled Amazon delivery driver in the U.S. Shame does work in certain
circumstances and, for corporations, the public’s sentiment of “shame on you”
can sometimes equal a loss in revenue and help move the needle toward
accountability.
This type of solidarity between highly-paid tech workers and their lower-paid
counterparts — who vastly outnumber them — is a tech CEO’s nightmare.
While corporations often treat their low-income workers as disposable, they’re
more hesitant to lose their high-income employees who can quickly snap up
jobs with competitors. Thus, the high-paid employees are allowed a far longer
leash when organizing, unionizing, and voicing their disappointment with
company culture and policies. They can use this increased security to advocate
with their lower-paid counterparts working at warehouses, delivering packages
or labeling data. As a result, corporations seem to use every tool at their
disposal to isolate these groups from each other.