Ex-Google researcher: AI workers need whistleblower protection




Artificial intelligence expert Timnit Gebru on the challenges researchers can face at Big Tech companies, and how to protect workers and their research.

Artificial intelligence research leads to new cutting-edge technologies, 
but it’s expensive.
Big Tech companies, which are powered by AI and have deep pockets, often take on this work  — but that gives them the power to censor or impede research that casts them in an unfavorable light, according to Timnit Gebru, a computer scientist, co-founder of the nonprofit organization Black in AI and the former co-leader of Google’s Ethical AI team.  

The situation imperils both the rights of AI workers at those companies and the quality of research that is shared with the public, said Gebru, speaking at the recent 
EmTech MIT conference hosted by MIT Technology Review.

“It’s all the incentive structures that are not in place for you to challenge the status quo,” she said.
Gebru was forced out at Google last December (Gebru said she was fired, while Google said she resigned) after co-writing a paper about the risks of large AI language models, such as environmental impacts and the difficulty in finding embedded biases. Google’s search engine runs on such a large language model.

Citing concerns, Google told Gebru to retract the paper from a conference or remove her name and the name of other Google researchers,
 according to The New York Times. Gebru refused to so without a fuller explanation from Google, which led to Google announcing her departure.

During her recent talk, Gebru highlighted what she views as the labor rights concerns of AI workers, how to protect them, and why academia isn’t always a better route for researchers. Ultimately, she said, the goal is better and more equitable artificial intelligence.
 
“The moment you push a little hard, you’re out”

Gebru’s research centers on unintended negative impacts of artificial intelligence. A paper she co-authored with MIT Media Lab researcher
 Joy Buolamwini explored bias in facial recognition algorithms.

After 
joining Google in 2018, “I had issues from the very beginning,” Gebru said.  She said some people had doubts that she would be able to change a company as large as Google. “I was thinking, ‘Okay, maybe I can carve out a small piece … that is safe for people in marginalized groups,’” she said. “What I learned is that it's impossible, because the moment you push a little hard, you're out. So if you survive, it's because maybe you're not poking … at a thing that they find super important.”

It’s important to hold tech companies accountable from the outside, Gebru said. 
“We can't have the current dynamic that we have and expect any sort of nonpropaganda tech to come out of tech companies," she said. "Because when you start censoring research, then that's what happens, right? The papers that come out end up being more like propaganda.”

Problems extend outside of Big Tech  

Since leaving Google, Gebru has been working to
 develop an independent research institute. While many AI researchers work in academia, Gebru said that in her experience, that avenue poses its own concerns related to gatekeeping, harassment, and an incentive structure that doesn’t reward long-term research.  

There are also concerns about tech companies funding AI research at academic institutions. Gebru cited
 The Grey Hoodie Project, a research paper by Mohamed Abdalla of the University of Toronto and Moustafa Abdalla of Harvard Medical School. The researchers compared the way Big Tech (large technology companies like Google, Amazon, and Facebook) is funding and leading AI research with how big tobacco companies funded research in an effort to dispel concerns about the health effects of smoking.

“At an independent research institute, you can do research that the company does not think is going to make it money right now. You can do research that really shows fundamental flaws in whatever technology that a company might be using,” Gebru said.




 

No comments

intech company. Powered by Blogger.