We urge policymakers to instead draw on solid scholarship that investigates the harms and risks of AI—and the harms caused by delegating authority to automated systems, which include the unregulated accumulation of data and computing power, climate costs of model training and inference, damage to the welfare state and the disempowerment of the poor, as well as the intensification of policing against Black and Indigenous families. Solid research in this domain—including social science and theory building—and solid policy based on that research will keep the focus on the people hurt by this technology.
YES
I don’t think Ai Is going to go skynet. I know it’s going to be used to disenfranchise Black people, destroy creative fields, and generate mis- and disinformation because it’s already doing it
This is such an important distinction. Current AI is incapable of wanting to cause any of that harm, yet it’s already happening. The danger won’t be skynet, it will be and always has been human greed and ignorance.
Advocacy groups like the EFF and ACLU have been raising the alarm about the surveillance state for decades. It has very little to do with AI, and regulating AI will not likely have any effect on state-level abuse, because guess what? They already operate above the law and beyond reason. That’s the real problem.
We urge policymakers to instead draw on solid scholarship that investigates the harms and risks of AI—and the harms caused by delegating authority to automated systems, which include […] damage to the welfare state and the disempowerment of the poor […]
I got an example of that yersterday, when calling FedEx’s over the phone to schedule a pickup.
Contrary to calls I’ve made in the past, it was answered by a chatbot instead of humans. This raise various problems, in particular for already-disenfranchised people who rely on customer service being available over the phone.
- A disclaimer says “This call may be recorded for training purpose”, there was no asking for consent, no way to opt-out. Does that mean my voice and conversation will be fed into a database to develop DL/LLM AI systems? There’s no simple way to know, or to refuse without hanging up (ie opt-out = being denied service).
- The chatbot had a hard time understanding my inquiry, forcing me to repeat/rephrase it multiples times. In the end it replied it’s not possible to schedule pickups over the phone. This was possible back when humans were on the phone. Internet access is now required. Meaning disenfranchised people with no (easy/dependable) Internet are being left out.
- The chatbot used a generic greeting, didn’t introduce itself properly, and was evasive when I asked “are you a human?”. Not being forthcoming will confuse some old or non-tech-savy people who won’t easily realize it’s a robot, not a human.
I don’t know, I’m much more concerned about the possibility that we develop huge automation capabilities that end up being controlled by very few people.
As for the specific issues in the article - yes, they’re real problems. But every advance in communication and information technology makes it easier to surveil or defame, and can be used for bad policing.
Right now there’s a push to regulate the internet to “prevent CSAM” by blocking encryption, and I’m afraid a push to regulate AI will not get better results.
Sure, we can ban predictive policing and demands some amounts of transparency (and the EU already wants to do that). But if we try to go further and impose restrictions on the AI models themselves, this will most likely solidify that AI is controlled by few powerful corporations. After all, highly regulated models by definition can’t be free and open.
But every advance in communication and information technology makes it easier to surveil or defame, and can be used for bad policing.
So we should just give up and accept surveillance and defamation without trying to come up with any legal counter measures?
Right now there’s a push to regulate the internet to “prevent CSAM” by blocking encryption, and I’m afraid a push to regulate AI will not get better results.
Totally agree that KOSA and the like are awful, but the existence of shitty regulations doesn’t negate the need for positive regulations.
But if we try to go further and impose restrictions on the AI models themselves, this will most likely solidify that AI is controlled by few powerful corporations. After all, highly regulated models by definition can’t be free and open.
I just don’t see this? So regulating powerful companies use of AI will… solidify their power? I’m not connecting the dots here but that might be on me, as I think there are plenty of highly regulated spaces that still see innovation.
Your assessment seems spot on to me. I’m connecting some projected dots to late stage capitalism. Perhaps the AIs will trickle down and such if we hold off on regulations.
Of course it’s possible for the government to impose regulations without sticking their face in and motorboating the AI’s contents. Google, Microsoft et al. would love to prevent this from happening because they actually do have their faces in there.
Some regulation proposals seem fine to me, like the proposed EU AI act.
But for some of the problems the article lists, like defamation or porn generation, you just can’t prevent if you have free and open models out there. You can make these things harder - and people already work on that - but if I have a free and open model, I can also change it (and remove restrictions).
The only way to stop those uses would be to keep AI tightly controlled in a walled garden. In capitalism, those walled gardens will belong to companies.
I think their examples are the kind of actual outcomes that should be addressed. Yours absolutely also are, but their core concept is “focus on the actual shit we see being fucked up instead of imaginary future shit that’s might theoretically eventually be a concern but is just not a real world issue with anything resembling the tech we have today”.
Didn’t read the article but headlines like this are extremely naive. If a person takes seriously the existential risk factor of AGI/ASI then that’s by definition the most important thing one can pay attention to. That’s like asking a deeply religious person to stop worrying about hell.
But to build on your analogy: we don’t make regulations based on a religious doctrine anymore in most countries. If your religion says no one is allowed to wear mixed fabrics or eat pork, that’s fine if you’re not doing that, but we’re not banning those things for all of society.
There’s about as much proof for an existential AGI threat as there is for a deity, so let’s not make policies based on either, and focus instead on real potential and already proven harms of AI