Automated legal guidance at federal agencies: report for the Administrative Conference of the United States
When individuals have questions about Federal benefits, services, and legal rules, they are increasingly seeking help from government chatbots, virtual assistants, and other automated tools. Current forms of automated legal guidance platforms include the U.S. Citizenship and Immigration Services’ Emma, the U.S. Department of Education’s Aidan, and the Internal Revenue Service’s Interactive Tax Assistant. Most scholars who have studied artificial intelligence and federal government agencies have not focused on the government’s use of technology to offer guidance to the public. The absence of scholarly attention to automation as a means of communicating government guidance is an important gap in the literature, given the strong influence that these communications can have on individuals’ decisions about the law.
This report describes the results of a qualitative study of automated legal guidance across the federal government, which included semi-structured interviews with both agency technology experts and lawyers. This study was conducted under the auspices of the Administrative Conference of the United States (ACUS). During the study, the authors reviewed the automated legal guidance activities of all federal agencies and conducted in-depth research on agencies that are already using well-developed chatbots, virtual assistants, or other related tools to assist the public in understanding or following relevant law. After identifying the agencies that are primary adopters of automated legal guidance, they conducted interviews with multiple individuals from each agency, as well as representatives from the U.S. General Services Administration.
The paper finds that automated legal guidance offers agencies an inexpensive way to help the public navigate through complex legal regimes. However, the authors also find that automated legal guidance may mislead members of the public about how the law will apply in their individual circumstances. In some cases, agencies exacerbate this problem by, among other things, making guidance seem more personalised than it is, not recognising how users may rely on the guidance, and not adequately disclosing that the guidance cannot be relied upon as a legal matter. In many respects, this is not a problem of agencies’ own making. Rather, agencies are faced with the difficult task of translating complex statutory and regulatory regimes for a public that has limited capacity to understand them. Agencies also often lack sufficient resources to engage in more personalised outreach. Fundamentally, the authors identify a tension between agencies’ reasonable desires to promote automated legal guidance and its under-appreciated limitations.
This report charts a path forward, offering policy recommendations, organised into five categories: transparency; reliance; disclaimers; process; and accessibility, inclusion, and equity. The authors believe this report, and the detailed policy recommendations that flow from it, will be critical for evaluating existing, as well as future, development of automated legal guidance by governments.