RightsCon report: Machine learning systems that discriminate violate human rights, says declaration

Machine learning software has been touted as the next wave of innovation, promising to help governments and businesses make faster and more accurate decisions.

But human rights activists and technology groups warned Wednesday that creating systems that discriminate should be treated as a violation of human rights.

It came with the release at the RightsCon conference of the so-called Toronto Declaration on preventing machine learning from being used to support discrimination.

Machine learning systems – sometimes called artificial intelligence – are more than pattern recognition software, say adherents of the declaration. Used wrongly – deliberately or inadvertently — by data scientists and software developers, they can violate privacy, data protection, freedom of expression, participation in cultural life and equality before the law.

“Systems that make decisions and process data can also implicate economic, social, and cultural rights,” says the document, ranging from deciding who gets jobs, healthcare or jobs.

As a result, countries and the private sector have obligations to promote, protect and respect human rights by ensuring machine learning systems aren’t used to support discrimination on grounds that include race, colour, sex, language, religion, political or other opinion, national or social origin, property or birth.

The declaration also says victims of human rights violations or abuses must have access to prompt and effective remedies, although these aren’t spelled out.

“There are many, many examples” today of discriminatory machine learning systems,” Anna Bacciarelli, Amnesty International’s advisor on technology and human rights and a member of the drafting team, said in an interview. “They are proliferating at present.”

Anna Bacciarelli, Amnesty International

She cited a 2016 report by ProPublica which raised questions about the fairness of risk assessment software used in some U.S. courts to help judges decide which accused are likely to commit another crime.

“We’re basically just showing states, private actors, tech companies how they need to apply their products to international human rights law,” she said of the declaration.

The document was drafted by a group of 11 people before the conference, including members of Amnesty International; Access Now, a digital rights group which organizes the annual conference; and Privacy International. It was honed by a larger group before being released that included people from technology companies.

It will be discussed at the conference, and a final version will be released at the end of the month. It is also expected that guidelines for developers and data scientists will also be created.

The non-binding declaration could be adopted by the United Nations, say supporters.

Immediate reaction was mixed. Human rights advocates were enthusiastic – it has already been endorsed by Human Rights Watch and Amnesty International. Others were cautious.

Steve Crown, Microsoft’s vice president and deputy general counsel, called the declaration a “hugely important step.”

“Artificial intelligence has so much opportunity to improve the human condition.” he said in an interview,” and because its so powerful, [also] opportunities for misuse … so we have to be thoughtful about the misuse or unintended consequences as well as all the great things.”

But he also said that “we don’t want to see pre-emptive regulation before we understand how to use the [legal] tools we already have.”

Key to the consensus on the document, said Dinah PoKempner, Human Rights Watch’s general counsel, is that much of the language on discrimination and equality was taken from the International Covenant on Civil and Political Rights adopted by many countries.

“It’s not as though we just invented something,” she said at a session. “This has been there, it’s there in your law. It’s there to be used and elaborated by national, local and municipal governments, and it’s there to be incorporated by corporations, ethicists and [software] designers as well. It’s not a recipe for every situation and every application. We will face many challenges in the future, but we don’t go into the future blind.”

Would you recommend this article?

Share

Thanks for taking the time to let us know what you think of this article!
We'd love to hear your opinion about this or any other story you read in our publication.


Jim Love, Chief Content Officer, IT World Canada

Featured Download

Howard Solomon
Howard Solomon
Currently a freelance writer, I'm the former editor of ITWorldCanada.com and Computing Canada. An IT journalist since 1997, I've written for several of ITWC's sister publications including ITBusiness.ca and Computer Dealer News. Before that I was a staff reporter at the Calgary Herald and the Brampton (Ont.) Daily Times. I can be reached at hsolomon [@] soloreporter.com

Featured Articles

Cybersecurity in 2024: Priorities and challenges for Canadian organizations 

By Derek Manky As predictions for 2024 point to the continued expansion...

Survey shows generative AI is a top priority for Canadian corporate leaders.

Leaders are devoting significant budget to generative AI for 2024 Canadian corporate...

Related Tech News

Tech Jobs

Our experienced team of journalists and bloggers bring you engaging in-depth interviews, videos and content targeted to IT professionals and line-of-business executives.

Tech Companies Hiring Right Now