Ethicists fire back at 'AI Pause' letter they say 'ignores the actual harms'
A group of well-known AI ethicists have written a counterpoint to this week's controversial letter asking for a six-month "pause" on AI development, criticizing it for a focus on hypothetical future threats when real harms are attributable to misuse of the tech today.
Thousands of people, including such familiar names as Steve Wozniak and Elon Musk, signed the open letter from the Future of Life institute earlier this week, proposing that development of AI models like GPT-4 should be put on hold in order to avoid "loss of control of our civilization," among other threats.
Timnit Gebru, Emily M. Bender, Angelina McMillan-Major and Margaret Mitchell are all major figures in the domains of AI and ethics, known (in addition to their work) for being pushed out of Google over a paper criticizing the capabilities of AI. They are currently working together at the DAIR Institute, a new research outfit aimed at studying and exposing and preventing AI-associated harms.
But they were not to be found on the list of signatories, and now have published a rebuke calling out the letter's failure to engage with existing problems caused by the tech.
"Those hypothetical risks are the focus of a dangerous ideology called longtermism that ignores the actual harms resulting from the deployment of AI systems today," they wrote, citing worker exploitation, data theft, synthetic media that props up existing power structures and the further concentration of those power structures in fewer hands.
The choice to worry about a Terminator- or Matrix-esque robot apocalypse is a red herring when we have, in the same moment, reports of companies like Clearview AI being used by the police to essentially frame an innocent man. No need for a T-1000 when you've got Ring cams on every front door accessible via online rubber-stamp warrant factories.
While the DAIR crew agree with some of the letter's aims, like identifying synthetic media, they emphasize that action must be taken now, on today's problems, with remedies we have available to us:
What we need is regulation that enforces transparency. Not only should it always be clear when we are encountering synthetic media, but organizations building these systems should also be required to document and disclose the training data and model architectures. The onus of creating tools that are safe to use should be on the companies that build and deploy generative systems, which means that builders of these systems should be made accountable for the outputs produced by their products.
The current race towards ever larger "AI experiments" is not a preordained path where our only choice is how fast to run, but rather a set of decisions driven by the profit motive. The actions and choices of corporations must be shaped by regulation which protects the rights and interests of people.
It is indeed time to act: but the focus of our concern should not be imaginary "powerful digital minds." Instead, we should focus on the very real and very present exploitative practices of the companies claiming to build them, who are rapidly centralizing power and increasing social inequities.
Incidentally, this letter echoes a sentiment I heard from Uncharted Power founder Jessica Matthews at yesterday's AfroTech event in Seattle: "You should not be afraid of AI. You should be afraid of the people building it." (Her solution: become the people building it.)
While it is vanishingly unlikely that any major company would ever agree to pause its research efforts in accordance with the open letter, it's clear judging from the engagement it received that the risks — real and hypothetical — of AI are of great concern across many segments of society. But if they won't do it, perhaps someone will have to do it for them.