Artificial Intelligence and Human Rights: We Need to Talk about the Use Phase




As business, government, and civil society make progress in addressing the human rights impacts and ethical questions arising from the use of artificial intelligence (AI), we believe that one supremely important constituency needs to participate much more actively: the “non-technology” companies integrating AI into their business operations, strategies, and plans. 

Without these participants, dialogue about AI and human rights risks being too focused on the development of AI, with insufficient attention given to the companies deploying AI. Our aim with this blog is to explain why.

In August last year, BSR published three reports setting out the importance of taking a human rights-based approach to the development, deployment, and use of AI. Since then, we’ve put this advice into practice in our work with BSR member companies in the U.S., Asia, and Europe to develop policies on AI and human rights, engage with civil society organizations, and undertake human rights due diligence of AI solutions. We’ve had the opportunity to consider a wide range of scenarios—including algorithmic decision making, facial recognition, and sentiment analysis—as well as a wide variety of application areas, including retail, national security, human resources, and transportation systems. As can be imagined for technologies that are evolving so rapidly, it’s been a time of extraordinary learning.

In many ways, this work has confirmed predominant assumptions that exist around how technology companies can fulfill the responsibility to respect the human rights impacts arising from the use of their products and services. Across many different settings, our recommendations have coalesced around some common themes: scrutinize the quality of training data, examine to whom you sell products and services and refuse sales to those most likely to misuse them, and establish acceptable use policies that place restrictions on how products and services may be used. We’ve also recommended system-wide approaches, such as advocating for rights-protecting laws and regulations, increasing disclosure and transparency, and providing best practice guidance for users.

These are all important responsibilities held by technology companies, and nothing that follows should suggest otherwise. However, we’ve found that the common thread running throughout these recommendations is the notion that technology companies should use their leverage to prevent the misuse of products, services, and technologies by influencing the actions of others—and that no matter how much effort is deployed, there is no guarantee of success.

Decisions made today about the deployment of AI will bring significant consequences for the realization of human rights long into the future.

This observation has led us to one simple question: In addition to trying to influence the actions of others, shouldn’t we also be working more directly with the companies, governments, and organizations that are directly deploying AI themselves? In the terms of the UN Guiding Principles on Business and Human Rights, why would we spend most of our time working with the companies that are contributing to or directly linked to human rights impacts, and much less of our time working with the companies that might be causing them?

Companies in all these industries should be taking a human rights-based approach to their use of AI.

The second of the three reports we published last year on the importance of a human rights-based approach to AI anticipated these issues. In it, we listed the human rights risks and opportunities arising from the use of AI in the financial services, health care, retail, transportation, agriculture, and extractives industries, and proposed sector-wide impact assessments for each industry. One year on, we are doubling down on this point of view, such as during our recent participation in the Skoll World Forum, Sustainable Brands Paris, and our own BSR Connect events

Decisions made today about the deployment of AI will bring significant consequences for the realization of human rights long into the future—and this means that undertaking due diligence of AI across all industries now is a matter of urgency and not something that can be put off into a distant future. Today, we are joining the Partnership on AI, and we look forward to making good on this perspective by working more closely with the Partnership on AI and BSR member companies across all industries to assess the human rights impacts arising from their use of AI.

Denne artikel er del af et tema:

I fokus: ansvarlig brug af nye teknologier

Automatisering, kunstig intelligens, robotteknologi, intelligente enheder, osv kan give mange fordele. Men alle har de implikationer for diverse menneskerettigheder og for jobmarkedet. Læs om, hvordan man kan arbejde ansvarligt med de nye muligheder.



What the SBTi Battle Portends: The Decisive Decade Becomes the Dilemma Decade



How Financial Institutions Can Manage Human Rights-Related Risks



How Companies Can Navigate China’s ESG Reporting Guidelines



Climate Scenario Analysis: It’s about More than CSRD Compliance



Scaling Investment in Carbon Credits with Integrity



Advancing a Just Transition: Lessons from Company Practices