The Exploited Labor Behind Artificial Intelligence

Suppor­ting trans­na­ti­o­nal worker orga­ni­zing should be at the center of the fight for “ethi­cal AI.”

The public’s unders­tan­ding of arti­fi­cial inte­lli­gence (AI) is largely shaped by pop culture — by block­bus­ter movies like “The Termi­na­tor” and their dooms­day scena­rios of machi­nes going rogue and destroying huma­nity. This kind of AI narra­tive is also what grabs the atten­tion of news outlets: a Google engi­neer clai­ming that its chat­bot was senti­ent was among the most discus­sed AI-rela­ted news in recent months, even reaching Step­hen Colbert’s milli­ons of viewers. But the idea of supe­rin­te­lli­gent machi­nes with their own agency and deci­sion-making power is not only far from reality — it distracts us from the real risks to human lives surroun­ding the deve­lop­ment and deploy­ment of AI systems. While the public is distrac­ted by the spec­ter of none­xis­tent senti­ent machi­nes, an army of preca­ri­zed workers stands behind the suppo­sed accom­plish­ments of arti­fi­cial inte­lli­gence systems today.

Many of these systems are deve­lo­ped by multi­na­ti­o­nal corpo­ra­ti­ons loca­ted in Sili­con Valley, which have been conso­li­da­ting power at a scale that, jour­na­list Gideon Lewis-Kraus notes, is likely unpre­ce­den­ted in human history. They are stri­ving to create auto­no­mous systems that can one day perform all of the tasks that people can do and more, without the requi­red sala­ries, bene­fits or other costs asso­ci­a­ted with employing humans. While this corpo­rate execu­ti­ves’ utopia is far from reality, the march to attempt its reali­za­tion has crea­ted a global under­class, perfor­ming what anth­ro­po­lo­gist Mary L. Gray and compu­ta­ti­o­nal social scien­tist Sidd­harth Suri call ghost work: the down­played human labor driving “AI”.

Tech compa­nies that have bran­ded them­sel­ves “AI first” depend on heavily survei­lled gig workers like data labe­lers, deli­very drivers and content mode­ra­tors. Star­tups are even hiring people to imper­so­nate AI systems like chat­bots, due to the pres­sure by venture capi­ta­lists to incor­po­rate so-called AI into their products. In fact, London-based venture capi­tal firm MMC Ventu­res surveyed 2,830 AI star­tups in the EU and found that 40% of them didn’t use AI in a meaning­ful way.

Far from the sophis­ti­ca­ted, senti­ent machi­nes portrayed in media and pop culture, so-called AI systems are fueled by milli­ons of under­paid workers around the world, perfor­ming repe­ti­tive tasks under preca­ri­ous labor condi­ti­ons. And unlike the “AI rese­ar­chers” paid six-figure sala­ries in Sili­con Valley corpo­ra­ti­ons, these exploi­ted workers are often recrui­ted out of impo­ve­ris­hed popu­la­ti­ons and paid as little as $1.46/hour after tax. Yet despite this, labor exploi­ta­tion is not central to the discourse surroun­ding the ethi­cal deve­lop­ment and deploy­ment of AI systems. In this arti­cle, we give exam­ples of the labor exploi­ta­tion driving so-called AI systems and argue that suppor­ting trans­na­ti­o­nal worker orga­ni­zing efforts should be a prio­rity in discus­si­ons pertai­ning to AI ethics.

We write this as people inti­ma­tely connec­ted to AI-rela­ted work. Adri­enne is a former Amazon deli­very driver and orga­ni­zer who has expe­ri­en­ced the harms of survei­llance and unre­a­lis­tic quotas esta­blis­hed by auto­ma­ted systems. Mila­gros is a rese­ar­cher who has worked closely with data workers, espe­ci­ally data anno­ta­tors in Syria, Bulga­ria and Argen­tina. And Timnit is a rese­ar­cher who has faced reta­li­a­tion for unco­ve­ring and commu­ni­ca­ting the harms of AI systems.

Trea­ting Workers Like Machi­nes

Much of what is currently descri­bed as AI is a system based on statis­ti­cal machine lear­ning, and more speci­fi­cally, deep lear­ning via arti­fi­cial neural networks, a metho­do­logy that requi­res enor­mous amounts of data to “learn” from. But around 15 years ago, before the proli­fe­ra­tion of gig work, deep lear­ning systems were consi­de­red merely an acade­mic curi­o­sity, confi­ned to a few inter­es­ted rese­ar­chers.

In 2009, howe­ver, Jia Deng and his colla­bo­ra­tors rele­a­sed the Image­Net data­set, the largest labe­led image data­set at the time, consis­ting of images scra­ped from the inter­net and labe­led through Amazon’s newly intro­du­ced Mecha­ni­cal Turk plat­form. Amazon Mecha­ni­cal Turk, with the motto “arti­fi­cial arti­fi­cial inte­lli­gence, ” popu­la­ri­zed the pheno­me­non of “crowd work”: large volu­mes of time-consu­ming work broken down into smaller tasks that can quickly be comple­ted by milli­ons of people around the world. With the intro­duc­tion of Mecha­ni­cal Turk, intrac­ta­ble tasks were suddenly made feasi­ble; for exam­ple, hand-labe­ling one million images could be auto­ma­ti­cally execu­ted by a thou­sand anony­mous people working in para­llel, each labe­ling only a thou­sand images. What’s more, it was at a price even a univer­sity could afford: crowd­wor­kers were paid per task comple­ted, which could amount to merely a few cents.

“So-called AI systems are fueled by milli­ons of under­paid workers around the world, perfor­ming repe­ti­tive tasks under preca­ri­ous labor condi­ti­ons.”

The Image­Net data­set was follo­wed by the Image­Net Large Scale Visual Recog­ni­tion Challenge, where rese­ar­chers used the data­set to train and test models perfor­ming a vari­ety of tasks like image recog­ni­tion: anno­ta­ting an image with the type of object in the image, such as a tree or a cat. While non-deep-lear­ning-based models perfor­med these tasks with the highest accu­racy at the time, in 2012, a deep-lear­ning-based archi­tec­ture infor­mally dubbed Alex­Net scored higher than all other models by a wide margin. This cata­pul­ted deep-lear­ning-based models into the mains­tream, and brought us to today, where models requi­ring lots of data, labe­led by low-wage gig workers around the world, are proli­fe­ra­ted by multi­na­ti­o­nal corpo­ra­ti­ons. In addi­tion to labe­ling data scra­ped from the inter­net, some jobs have gig workers supply the data itself, requi­ring them to upload selfies, pictu­res of friends and family or images of the objects around them.

Unlike in 2009, when the main crowd­wor­king plat­form was Amazon’s Mecha­ni­cal Turk, there is currently an explo­sion of data labe­ling compa­nies. These compa­nies are raising tens to hundreds of milli­ons in venture capi­tal funding while the data labe­lers have been esti­ma­ted to make an average of $1.77 per task. Data labe­ling inter­fa­ces have evol­ved to treat crowd­wor­kers like machi­nes, often pres­cri­bing them highly repe­ti­tive tasks, survei­lling their move­ments and punis­hing devi­a­tion through auto­ma­ted tools. Today, far from an acade­mic challenge, large corpo­ra­ti­ons clai­ming to be “AI first” are fueled by this army of under­paid gig workers, such as data labo­rers, content mode­ra­tors, ware­house workers and deli­very drivers.

Content mode­ra­tors, for exam­ple, are respon­si­ble for finding and flag­ging content deemed inap­pro­pri­ate for a given plat­form. Not only are they essen­tial workers, without whom social media plat­forms would be comple­tely unusa­ble, their work flag­ging diffe­rent types of content is also used to train auto­ma­ted systems aiming to flag texts and imagery contai­ning hate speech, fake news, violence or other types of content that viola­tes plat­forms’ poli­cies. In spite of the crucial role that content mode­ra­tors play in both keeping online commu­ni­ties safe and trai­ning AI systems, they are often paid mise­ra­ble wages while working for tech giants and forced to perform trau­ma­tic tasks while being closely survei­lled.

Every murder, suicide, sexual assault or child abuse video that does not make it onto a plat­form has been viewed and flag­ged by a content mode­ra­tor or an auto­ma­ted system trai­ned by data most likely supplied by a content mode­ra­tor. Employees perfor­ming these tasks suffer from anxi­ety, depres­sion and post-trau­ma­tic stress disor­der due to cons­tant expo­sure to this horri­fic content.

Besi­des expe­ri­en­cing a trau­ma­tic work envi­ron­ment with none­xis­tent or insuf­fi­ci­ent mental health support, these workers are moni­to­red and punis­hed if they devi­ate from their pres­cri­bed repe­ti­tive tasks. For instance, Sama content mode­ra­tors contrac­ted by Meta in Kenya are moni­to­red through survei­llance soft­ware to ensure that they make deci­si­ons about violence in videos within 50 seconds, regard­less of the length of the video or how distur­bing it is. Some content mode­ra­tors fear that failure to do so could result in termi­na­tion after a few viola­ti­ons. “Through its prio­ri­ti­za­tion of speed and effi­ci­ency above all else, ” Time Maga­zine repor­ted, “this policy might explain why videos contai­ning hate speech and inci­te­ment to violence have remai­ned on Face­book’s plat­form in Ethi­o­pia.”

Simi­lar to social media plat­forms which would not func­tion without content mode­ra­tors, e-commerce conglo­me­ra­tes like Amazon are run by armies of ware­house workers and deli­very drivers, among others. Like content mode­ra­tors, these workers both keep the plat­forms func­ti­o­nal and supply data for AI systems that Amazon may one day use to replace them: robots that stock packa­ges in ware­hou­ses and self-driving cars that deli­ver these packa­ges to custo­mers. In the mean­time, these workers must perform repe­ti­tive tasks under the pres­sure of cons­tant survei­llance — tasks that, at times, put their lives at risk and often result in seri­ous muscu­los­ke­le­tal inju­ries.

“Data labe­ling inter­fa­ces have evol­ved to treat crowd­wor­kers like machi­nes, often pres­cri­bing them highly repe­ti­tive tasks, survei­lling their move­ments and punis­hing devi­a­tion through auto­ma­ted tools.”

Amazon ware­house employees are trac­ked via came­ras and their inven­tory scan­ners, and their perfor­mance is measu­red against the times mana­gers deter­mine every task should take, based on aggre­gate data from everyone working at the same faci­lity. Time away from their assig­ned tasks is trac­ked and used to disci­pline workers.

Like ware­house workers, Amazon deli­very drivers are also moni­to­red through auto­ma­ted survei­llance systems: an app called Mentor tallies scores based on so-called viola­ti­ons. Amazon’s unre­a­lis­tic deli­very time expec­ta­ti­ons push many drivers to take risky measu­res to ensure that they deli­ver the number of packa­ges assig­ned to them for the day. For instance, the time it takes some­one to fasten and unfas­ten their seat­belt some 90–300 times a day is enough to put them behind sche­dule on their route. Adri­enne and many of her colle­a­gues buck­led their seat belts behind their backs, so that the survei­llance systems regis­te­red that they were driving with a belt on, without getting slowed down by actu­ally driving with a belt on.

In 2020, Amazon drivers in the U.S. were inju­red at a nearly 50% higher rate than their United Parcel Service coun­ter­parts. In 2021, Amazon drivers were inju­red at a rate of 18.3 per 100 drivers, up nearly 40% from the previ­ous year. These condi­ti­ons aren’t only dange­rous for deli­very drivers — pedes­tri­ans and car passen­gers have been killed and inju­red in acci­dents invol­ving Amazon deli­very drivers. Some drivers in Japan recently quit in protest because they say Amazon’s soft­ware sent them on “impos­si­ble routes, ” leading to “unre­a­so­na­ble demands and long hours.” In spite of these clear harms, howe­ver, Amazon conti­nues to treat its workers like machi­nes.

In addi­tion to trac­king its workers through scan­ners and came­ras, last year, the company requi­red deli­very drivers in the U.S. to sign a “biome­tric consent” form, gran­ting Amazon permis­sion to use AI-powe­red came­ras to moni­tor drivers’ move­ments — suppo­sedly to cut down on distrac­ted driving or spee­ding and ensure seat­belt usage. It’s only reaso­na­ble for workers to fear that facial recog­ni­tion and other biome­tric data could be used to perfect worker-survei­llance tools or further train AI — which could one day replace them. The vague wording in the consent forms leaves the precise purpose open for inter­pre­ta­tion, and workers have suspec­ted unwan­ted uses of their data before (though Amazon denied it).

The “AI” industry runs on the backs of these low-wage workers, who are kept in preca­ri­ous posi­ti­ons, making it hard, in the absence of unio­ni­za­tion, to push back on unet­hi­cal prac­ti­ces or demand better working condi­ti­ons for fear of losing jobs they can’t afford to lose. Compa­nies make sure to hire people from poor and under­ser­ved commu­ni­ties, such as refu­gees, incar­ce­ra­ted people and others with few job opti­ons, often hiring them through third party firms as contrac­tors rather than as full time employees. While more employers should hire from vulne­ra­ble groups like these, it is unac­cep­ta­ble to do it in a preda­tory manner, with no protec­ti­ons.

“AI ethics rese­ar­chers should analyze harm­ful AI systems as both causes and conse­quen­ces of unjust labor condi­ti­ons in the industry.”

Data labe­ling jobs are often perfor­med far from the Sili­con Valley head­quar­ters of “AI first” multi­na­ti­o­nal corpo­ra­ti­ons — from Vene­zu­ela, where workers label data for the image recog­ni­tion systems in self-driving vehi­cles, to Bulga­ria, where Syrian refu­gees fuel facial recog­ni­tion systems with selfies labe­led accor­ding to race, gender, and age cate­go­ries. These tasks are often outsour­ced to preca­ri­ous workers in coun­tries like India, Kenya, the Philip­pi­nes or Mexico. Workers often do not speak English but are provi­ded instruc­ti­ons in English, and face termi­na­tion or banning from crowd­work plat­forms if they do not fully unders­tand the rules.

These corpo­ra­ti­ons know that incre­a­sed worker power would slow down their march toward proli­fe­ra­ting “AI” systems requi­ring vast amounts of data, deployed without adequa­tely studying and miti­ga­ting their harms. Talk of senti­ent machi­nes only distracts us from holding them accoun­ta­ble for the exploi­ta­tive labor prac­ti­ces that power the “AI” industry.

An Urgent Prio­rity For AI Ethics

While rese­ar­chers in ethi­cal AI, AI for social good, or human-cente­red AI have mostly focu­sed on “debi­a­sing” data and foste­ring trans­pa­rency and model fair­ness, here we argue that stop­ping the exploi­ta­tion of labor in the AI industry should be at the heart of such initi­a­ti­ves. If corpo­ra­ti­ons are not allo­wed to exploit labor from Kenya to the U.S., for exam­ple, they will not be able to proli­fe­rate harm­ful tech­no­lo­gies as quickly — their market calcu­la­ti­ons would simply dissu­ade them from doing so.

Thus, we advo­cate for funding of rese­arch and public initi­a­ti­ves that aim to unco­ver issues at the inter­sec­tion of labor and AI systems. AI ethics rese­ar­chers should analyze harm­ful AI systems as both causes and conse­quen­ces of unjust labor condi­ti­ons in the industry. Rese­ar­chers and prac­ti­ti­o­ners in AI should reflect on their use of crowd­wor­kers to advance their own care­ers, while the crowd­wor­kers remain in preca­ri­ous condi­ti­ons. Instead, the AI ethics commu­nity should work on initi­a­ti­ves that shift power into the hands of workers. Exam­ples include co-crea­ting rese­arch agen­das with workers based on their needs, suppor­ting cross-geograp­hi­cal labor orga­ni­zing efforts and ensu­ring that rese­arch findings are easily acces­sed by workers rather than confi­ned to acade­mic publi­ca­ti­ons. The Turkop­ti­con plat­form crea­ted by Lilly Irani and M. Six Silber­man, “an acti­vist system that allows workers to publi­cize and evalu­ate their rela­ti­ons­hips with employers, ” is a great exam­ple of this.

Jour­na­lists, artists, and scien­tists can help by drawing clear the connec­tion between labor exploi­ta­tion and harm­ful AI products in our every­day lives, foste­ring soli­da­rity with and support for gig workers and other vulne­ra­ble worker popu­la­ti­ons. Jour­na­lists and commen­ta­tors can show the gene­ral public why they should care about the data anno­ta­tor in Syria or the hyper­sur­vei­lled Amazon deli­very driver in the U.S. Shame does work in certain circums­tan­ces and, for corpo­ra­ti­ons, the public’s senti­ment of “shame on you” can some­ti­mes equal a loss in reve­nue and help move the needle toward accoun­ta­bi­lity.

Suppor­ting trans­na­ti­o­nal worker orga­ni­zing should be at the center of the fight for “ethi­cal AI.” While each work­place and geograp­hi­cal context has its own idiosyn­cra­sies, knowing how workers in other loca­ti­ons circum­ven­ted simi­lar issues can serve as inspi­ra­tion for local orga­ni­zing and unio­ni­zing efforts. For exam­ple, data labe­lers in Argen­tina could learn from the recent unio­ni­zing efforts of content mode­ra­tors in Kenya, or Amazon Mecha­ni­cal Turk workers orga­ni­zing in the U.S., and vice versa. Further­more, unio­ni­zed workers in one geograp­hic loca­tion can advo­cate for their more preca­ri­ous coun­ter­parts in anot­her, as in the case of the Alpha­bet Workers Union, which inclu­des both high paid employees in Sili­con Valley and outsour­ced low wage contrac­tors in more rural areas.

“This type of soli­da­rity between highly-paid tech workers and their lower-paid coun­ter­parts — who vastly outnum­ber them — is a tech CEO’s night­mare.”

This type of soli­da­rity between highly-paid tech workers and their lower-paid coun­ter­parts — who vastly outnum­ber them — is a tech CEO’s night­mare. While corpo­ra­ti­ons often treat their low-income workers as dispo­sa­ble, they’re more hesi­tant to lose their high-income employees who can quickly snap up jobs with compe­ti­tors. Thus, the high-paid employees are allo­wed a far longer leash when orga­ni­zing, unio­ni­zing, and voicing their disap­point­ment with company culture and poli­cies. They can use this incre­a­sed secu­rity to advo­cate with their lower-paid coun­ter­parts working at ware­hou­ses, deli­ve­ring packa­ges or labe­ling data. As a result, corpo­ra­ti­ons seem to use every tool at their dispo­sal to isolate these groups from each other.

Emily Cunning­ham and Maren Costa crea­ted the type of cross-worker soli­da­rity that scares tech CEOs. Both women worked as user expe­ri­ence desig­ners at Amazon’s Seattle head­quar­ters cumu­la­ti­vely for 21 years. Along with other Amazon corpo­rate workers, they co-foun­ded the Amazon Employees for Climate Justice (AECJ). In 2019, over 8,700 Amazon workers publicly signed their names to an open letter addres­sed to Jeff Bezos and the company’s board of direc­tors deman­ding climate leaders­hip and concrete steps the company needed to imple­ment to be alig­ned with climate science and protect workers. Later that year, AECJ orga­ni­zed the first walkout of corpo­rate workers in Amazon’s history. The group says over 3,000 Amazon workers walked out across the world in soli­da­rity with a youth-led Global Climate Strike.

Amazon respon­ded by announ­cing its Climate Pledge, a commit­ment to achi­eve net-zero carbon by 2040 — 10 years ahead of the Paris Climate Agre­e­ment. Cunning­ham and Costa say they were both disci­pli­ned and thre­a­te­ned with termi­na­tion after the climate strike — but it wasn’t until AECJ orga­ni­zed acti­ons to foster soli­da­rity with low-wage workers that they were actu­ally fired. Hours after anot­her AECJ member sent out a calen­dar invite invi­ting corpo­rate workers to listen to a panel of ware­house workers discus­sing the dire working condi­ti­ons they were facing at the begin­ning of the pande­mic, Amazon fired Costa and Cunning­ham. The Nati­o­nal Labor Rela­ti­ons Board found their firings were ille­gal, and the company later sett­led with both women for undis­clo­sed amounts. This case illus­tra­tes where execu­ti­ves’ fears lie: the unflin­ching soli­da­rity of high-income employees who see low-income employees as their comra­des.

In this light, we urge rese­ar­chers and jour­na­lists to also center low-income workers’ contri­bu­ti­ons in running the engine of “AI” and to stop misle­a­ding the public with narra­ti­ves of fully auto­no­mous machi­nes with human-like agency. These machi­nes are built by armies of under­paid labo­rers around the world. With a clear unders­tan­ding of the labor exploi­ta­tion behind the current proli­fe­ra­tion of harm­ful AI systems, the public can advo­cate for stron­ger labor protec­ti­ons and real conse­quen­ces for enti­ties who break them.

By Adri­enne Williams, Mila­gros Miceli and Timnit Gebru Octo­ber 13, 2022

Adri­enne Williams and Mila­gros Miceli are rese­ar­chers at the Distri­bu­ted AI Rese­arch (DAIR) Insti­tute. Timnit Gebru is the insti­tu­te’s foun­der and execu­tive direc­tor. She was previ­ously co-lead of the Ethi­cal AI rese­arch team at Google.

Image: Nash Weera­se­kera for Noema Maga­zine