|Author (Person)||Forster, Madeleine|
|Publisher||Royal Institute of International Affairs [Chatham House]|
|Series Title||Chatham House Papers|
|Publication Date||September 2022|
|Content Type||Research Paper|
Artificial intelligence (AI) is being introduced to help decision-making in high-risk fields. This includes decision-making about asylum and refugee protection, where automated ways of processing people and predicting risks in contested circumstances hold great appeal.
This field, even more than most, will act as a test case for how AI protects or fails to protect human rights. Wrong or biased decisions about refugee status can have life and death consequences, including the return of refugees to places where they face persecution, contrary to international law. Existing refugee decision-making systems are already complex and are often affected by flaws, including lack of legal remedies – issues that can be exacerbated when overlayed with AI.
This paper examines the primary protections being proposed to make AI more responsive to human rights, including the upcoming EU AI Law. Can innovation and protection of human rights really be combined in asylum systems and other domains that make decisions about the future of vulnerable communities and minorities? This is a question not just for governments but also for private sector providers, which have independent human rights responsibilities when providing AI products in a politically charged and changeable policy field that decides the future of vulnerable communities and minorities.
|Subject Categories||Justice and Home Affairs, Values and Beliefs|
|Subject Tags||Asylum | Refugees, Fundamental | Human Rights|