Senate Bill 5116

Source

Section 1

The legislature finds that:

  1. Washington is a technology leader on a national and global level and holds a distinctive position in creating frameworks around technology that enhance innovation while protecting consumers and promoting fairness, accountability, and transparency for all Washingtonians.

  2. Automated decision systems are rapidly being adopted to make or assist in core decisions in a variety of government and business functions, including criminal justice, health care, education, employment, public benefits, insurance, and commerce.

  3. These automated decision systems are currently unregulated, may be deployed without public notice, and vendors selling the systems may require restrictive contractual provisions that undermine government transparency and accountability.

  4. The average Washington resident is unlikely to understand processes used by these automated decision systems, yet these systems are increasingly used to make core government and business decisions impacting the civil rights and liberties of Washingtonians, raising significant concerns around due process, fairness, accountability, and transparency.

  5. A growing body of research shows that reliance on automated decision systems without adequate transparency, oversight, or safeguards can undermine market predictability, harm consumers, and deny historically disadvantaged or vulnerable groups the full measure of their civil rights and liberties.

  6. In order to enhance innovation and ensure the use of these systems in ways that benefit Washington residents, the legislature intends to ensure the fair, transparent, and accountable use of automated decision systems.

Section 2

The definitions in this section apply throughout this chapter unless the context clearly requires otherwise.

  1. "Agency" or "public agency" means any entity meeting the definition of "public agency" as defined in RCW 42.30.020.

  2. "Algorithmic accountability report" means the report with content enumerated in section 5(6) of this act.

  3. "Artificial intelligence-enabled profiling" means the automated or semiautomated process by which the external or internal characteristics of an individual are analyzed to determine, infer, or characterize an individual's state of mind, character, propensities, protected class status, political affiliation, religious beliefs or religious affiliation, immigration status, or employability.

  4. "Automated decision system" means any electronic software, system, or process designed to automate, aid, or replace a decision-making process that impacts the welfare or rights of any Washington resident, and that would otherwise be performed by humans. Automated decision systems include, without limitation, tools that analyze data sets to generate scores, predictions, classifications, or recommended actions that are used by agencies to automate, aid, or replace decision-making processes that impact the welfare or rights of any Washington resident. Automated decision systems do not include tools that do not automate, aid, or replace such decision-making processes, including without limitation junk email filters, firewalls, antivirus software, calculators, spreadsheets, databases, data sets, or other compilations of data.

  5. "Automated final decision system" means an automated decision system that makes final decisions, judgments, or conclusions without human intervention.

  6. "Automated support decision system" means an automated decision system that provides information to inform the final decision, judgment, or conclusion of a human decisionmaker.

  7. "Use" means to operate an automated decision system or to contract with a third party to operate an automated decision system to automate, aid, or replace any decision-making process that would otherwise be made by an agency.

Section 3

By January 1, 2022, the director of the office of the chief information officer shall adopt rules pursuant to chapter 34.05 RCW regarding the development, procurement, and use of automated decision systems by a public agency. These rules must incorporate the minimum standards and procedures set forth in sections 4 and 5 of this act with respect to automated decision systems including, but not limited to, a definition of systematic discrimination or less favorable treatment as set forth in section 4(1) of this act. In adopting the rules, the director of the office of the chief information officer must consult with representatives of communities whose rights are disproportionately impacted by automated decision systems as demonstrated by current studies.

Section 4

The following provisions apply to a public agency's development, procurement, or use of an automated decision system:

  1. A public agency may not develop, procure, or use an automated decision system that discriminates against an individual, or treats an individual less favorably than another, in whole or in part, on the basis of one or more factors enumerated in RCW 49.60.010. A public agency may not develop, procure, or use an automated final decision system to make a decision impacting the constitutional or legal rights, duties, or privileges of any Washington resident, or to deploy or trigger any weapon.

  2. A public agency may not operate, install, or commission the operation or installation of equipment incorporating artificial intelligence-enabled profiling in any place of public resort, accommodation, assemblage, or amusement, as defined in RCW 49.60.040, or use artificial intelligence-enabled profiling to make decisions that produce legal effects or similarly significant effects concerning individuals. Decisions that include legal effects or similarly significant effects concerning consumers include, without limitation, denial or degradation of consequential services or support, such as financial or lending services, housing, insurance, educational enrollment, criminal justice, employment opportunities, health care services, and access to basic necessities, such as food and water.

  3. A public agency shall develop, procure, or use an automated decision system only after the public agency first completes an algorithmic accountability report and after that the director of the office of the chief information officer approves the report, as set forth in section 5 of this act.

  4. A public agency that develops, procures, or uses an automated decision system must follow any conditions set forth in the relevant approved algorithmic accountability report. In addition, the public agency must, at a minimum:

    1. Give clear notice in plain language to an individual impacted by the automated decision system of the following:

      1. The fact that the system is in use;

      2. The system's name, vendor, and version;

      3. What decision or decisions it will be used to make or support;

      4. Whether it is an automated final decision system or automated support decision system and whether and through what process a human verifies or confirms decisions made by the automated decision system;

    2. What policies and guidelines apply to its deployment; and

    1. How an individual may contest any decision made involving the automated decision system as required pursuant to this section;
    1. Ensure the automated decision system and the data used to develop the system are made freely available by the vendor before, during, and after deployment for agency or independent third-party testing, auditing, or research to understand its impacts, including potential bias, inaccuracy, or disparate impacts;

    2. Ensure that any decision made or informed by the automated decision system is subject to appeal, immediate suspension if a legal right, duty, or privilege is impacted by the decision, and potential reversal by a human decisionmaker through a timely process not to exceed 20 days, and clearly described and accessible to an individual impacted by the decision; and

    3. Ensure the agency can explain the basis for its decision to any impacted individual in terms understandable to a layperson including, without limitation, by requiring the vendor to create such an explanation.

  5. A procurement contract for an automated decision system entered into by a public agency must ensure that the minimum standards set forth in this section are able to be effectuated without impairment, including requiring the vendor to waive any legal claims that may impair these minimum standards. Such a contract may not contain nondisclosure or other provisions that prohibit or impair these minimum standards.

Section 5

  1. A public agency intending to develop, procure, or use an automated decision system must produce an algorithmic accountability report for that system, and that system must be approved by the director of the office of the chief information officer prior to deployment, according to the procedures set forth in this section.

  2. Agencies already using an automated decision system as of the effective date of this section must comply with all provisions and procedures set forth in this chapter by January 1, 2022. If such an agency is not in compliance by that date, the agency must immediately cease use of the automated decision system until such time as compliance is achieved.

  3. The agency intending to develop, procure, or use an automated decision system must submit the algorithmic accountability report to the director of the office of the chief information officer prior to any use of the automated decision system. The director of the office of the chief information officer must post the algorithmic accountability report on the office of the chief information officer's public website and invite public comment on the algorithmic accountability report for a period of no less than 30 days.

  4. After receiving public comment, the director of the office of the chief information officer must determine whether the intended use of the automated decision system meets the minimum standards set forth in section 4 of this act. On the basis of that determination, the director of the office of the chief information officer may approve the use of the automated decision system in accordance with rules and procedures set forth in the algorithmic accountability report, deny it, or make changes to rules and procedures set forth in the algorithmic accountability report prior to approval.

  5. Except as provided in subsection (2) of this section, no agency may develop, procure, or use an automated decision system prior to obtaining the approval required in this section, and after approval, such an agency may use the automated decision system only in accordance with the policies and procedures set forth in the approved algorithmic accountability report.

  6. Each algorithmic accountability report must include clear and understandable statements of the following:

    1. The automated decision system's name, vendor, and version;

    2. A description of the automated decision system's general capabilities, including reasonably foreseeable capabilities outside the scope of the agency's proposed use and whether the automated decision system is used or may be used to deploy or trigger any weapon;

    3. [Empty]

      1. The type or types of data inputs that the technology uses; (ii) how that data is generated, collected, and processed; and (iii) the type or types of data the system is reasonably likely to generate;
    4. Whether the automated decision system has been tested by an independent third party, has a known bias, or is untested for bias;

    5. A description of the purpose and proposed use of the automated decision system, including: What decision or decisions it will be used to make or support; whether it is an automated final decision system or automated support decision system; and its intended benefits, including any data or research demonstrating those benefits;

    6. A description of how the agency plans to comply with each requirement set forth in section 4 of this act;

    7. Whether the automated decision system makes decisions affecting the constitutional or legal rights, duties, or privileges of any Washington resident;

    8. A description of any potential impacts of the automated decision system on civil rights and liberties and potential disparate impacts on marginalized communities, and a mitigation plan;

    9. A clear use and data management policy, including specific protocols for the following:

      1. How and when the automated decision system will be deployed or used and by whom including, but not limited to: The factors that will be used to determine where, when, and how the technology is deployed; and other relevant information, such as whether the technology will be operated continuously or used only under specific circumstances. If the automated decision system will be operated or used by another entity on the agency's behalf, the algorithmic accountability report must explicitly include a description of the other entity's access and any applicable protocols;

      2. Any additional rules that will govern use of the automated decision system and what processes will be required prior to each use of the automated decision system;

      3. How automated decision system data will be securely stored and accessed, and whether an agency intends to share access to the automated decision system or the data from that automated decision system with any other entity, and why;

      4. How the agency will ensure that all personnel who operate the automated decision system or access its data are properly trained and able to ensure compliance with the use and data management policy prior to use of the automated decision system;

    10. A description of any public or community engagement held and any future public or community engagement plans in connection with the automated decision system; and

    1. A description of the fiscal impact of the automated decision system, including: Initial acquisition costs; ongoing operating costs such as maintenance, licensing, personnel, legal compliance, use auditing, data retention, and security costs; any cost savings that would be achieved through the use of the technology; and any current or potential sources of funding, including any subsidies, incentives, or free products being offered by vendors or governmental entities.

Section 6

  1. By December 1, 2021, the office of the chief information officer shall make publicly available on its website a comprehensive inventory of all automated decision systems that have been proposed for or are being used, developed, or procured by state agencies, including all algorithmic accountability reports relating to such systems. Beginning January 1, 2022, the office of the chief information officer shall make publicly available on its website metrics on all approvals, conditional approvals, or denials of agency proposals to develop, procure, or use automated decision systems, including written explanations of each decision.

  2. [Empty]

    1. Beginning January 1, 2022, the office of the chief information officer shall conduct an annual audit that includes the following:

      1. Whether each agency that uses, develops, or procures an automated decision system has complied with the terms of its approved algorithmic accountability report;

      2. Descriptions of any known or reasonably suspected violations of any algorithmic accountability report policies;

      3. Any systematic issues, such as bias, disproportionate impacts on marginalized or vulnerable communities, raised by use of automated decision systems;

      4. Publishing recommendations, if any, relating to revision to this chapter or to specific automated decision system algorithmic accountability reports.

    2. The first annual audit must be made publicly available on the chief information officer's website by March 1, 2022, and annually thereafter on or before March 1st.

  3. Beginning January 1, 2022, each agency using an automated decision system must publish on its website annual metrics regarding the number of requests for human review of a decision rendered by the automated decision system it received and the outcome of the human review.

Section 7

Any person who is injured by a material violation of this chapter may institute proceedings against the public agency deploying the automated decision system in a court of competent jurisdiction for injunctive relief, including restoration of the government benefit in question, declaratory relief, or a writ of mandate to enforce this chapter. Actionable injuries under this section include, but are not limited to, denial or interference with: Any government benefit, direct or indirect financial harm, physical harm or threats to persons or property, discrimination in goods, services, or economic opportunity, interference with constitutional or statutory rights or privileges, interference with the right to vote or participate in free and fair elections, or other impacts on human welfare; if any of the foregoing occur due to the use of an automated decision system that does not meet the standards set forth in this chapter.

Section 8

This section adds a new section to an existing chapter 49.60. Here is the modified chapter for context.

It is an unfair practice for any automated decision system to discriminate against an individual, or to treat an individual less favorably than another, in whole or in part, on the basis of one or more factors enumerated in RCW 49.60.010. For the purposes of this section, "automated decision system" has the same meaning as defined in section 2 of this act.

Section 10

This section modifies existing section 43.386.901. Here is the modified chapter for context.

Sections 1 through 9 and 11 through 13 of this act take effect July 1, 2026.


Created by @tannewt. Contribute on GitHub.