The White House unveiled a bill of rights for artificial intelligence (AI) that lays out voluntary guidelines for how companies can avoid the misuse of AI. The new AI bill of rights results from a year’s initiative to help businesses, researchers, and policymakers work together to ensure that AI ethics are used are implemented to benefit Americans.
The “Blueprint for an AI Bill of Rights” is a 73-page document developed by researchers, technologists, advocates, journalists, and policymakers. It contains a set of voluntary guidelines that public and private organizations can use to form an AI ethics code to protect people from abuse and discrimination in algorithms. “Too often, these tools limit our opportunities and prevent our access to critical resources or services. These problems are well-documented. In America and around the world, systems supposed to help with patient care have proven unsafe, ineffective, or biased,” according to the White House.
Healthcare AI ethics have been specifically mentioned concerning access to critical resources or services in the handout and data-sensitive domains and calls for enhanced protection and restrictions for data across such sensitive domains.
What’s in the AI Bill of Rights?
This Blueprint for an AI Bill of Rights offers five broad principles regarding technical protections and practices to help guide the development and implementation of AI systems, including AI relevant to behavioral health. Each principle on the new blueprint further contains three supplemental sections, which detail problems to be addressed, expected outcomes, automated systems, and practical steps that can be implemented to realize the vision of AI ethics.
Safe & Effective Systems
The first principle of the Blueprint for an AI Bill of Rights states, “You (American people) should be protected from unsafe or ineffective systems.” Meeting this goal requires the identification of risks, pre-deployment testing, and monitoring of AI automated systems.
Algorithmic Discrimination Protections
The second principle states, “You (American people) should not face discrimination by algorithms, and systems should be used and designed in an equitable way.” This principle is important in making healthcare algorithms socially responsible and equitable. This principle recognizes bias in algorithms that may lead to discrimination and violate legal protections.
The third principle states, “You should be protected from abusive data practices via built-in protections, and you should have agency over how data about you is used.” This asks for design choices, user permissions, transfer, and deletion of data in line with reasonable expectations, to ensure and safeguard the collection of data by developers of AI systems. This principle emphasizes the need for organizations to use a language for asking permission that is understandable to users.
Notice and Explanation
The fourth principle states, “You should know that an automated system is being used and understand how and why it contributes to outcomes that impact you.” This calls for clear descriptions of the overall system functioning and the role of automation in that system.
Human Alternatives, Consideration, and Fallback
This principle states, “You should be able to opt-out, where appropriate, and have access to a person who can quickly consider and remedy problems you encounter.” This principle asks for human alternatives so users can opt-out whenever they choose.
HHS Announces Proposed Rule to Strengthen Nondiscrimination in Health Care
Complementing the voluntary guidelines released by the White House, the United States Department of Health and Human Services (HHS) released its “Trustworthy AI (TAI) Playbook” (PDF) to Prevent Algorithmic Health Discrimination and Protect Consumers. HHS Chief AI Officer Oki Mek had this to say about the Trustworthy AI Playbook and its relevance to AI ethics:
HHS has a significant role to play in strengthening American leadership in Artificial Intelligence (AI). As we use AI to advance the health and wellbeing of the American people, we must maintain public trust by ensuring that our solutions are ethical, effective, and secure. The HHS Trustworthy AI (TAI) Playbook is an initial step by the Office of the Chief AI Officer (OCAIO) to support trustworthy AI development across the Department.
- Behavioral professionals interested in other resources are directed to consider these options:
- AI in Mental Health: Benefits of Technologies for AI Mental Health
- What is Artificial Intelligence in Healthcare? Will It Replace Your Job?
- Artificial Intelligence in Behavioral and Mental Health Care