Last updated September 1st, 2023
Abstract: The Federal AI Inventory Analysis project for 2023 is a comprehensive examination of the public AI projects reported in the federal inventory during that collection cycle. The complexity of this task is due to the variation in reporting formats and the inclusion of non-machine-readable reports from different Departments and independent agencies.
The analysis includes data collection from published AI use case inventories, data preparation, which required a blend of manual parsing and automated processing, and the use of a high-powered language model (LLM) for summarizing the collected data. Themes were then extracted from these summaries to categorize projects into specific domains like security, infrastructure, and healthcare. Visualization was employed to understand the general clustering of projects, using OpenAIβs text-embedding model. The project also included a mechanism for daily data integrity checks and provided transparent cost calculations for API usage. The outcomes include detailed and summarized reports of AI projects, interactive visualizations, and insights into prominent AI themes across various federal Departments and agencies, thereby contributing to a transparent understanding of AI's role within the federal government.
Disclaimer: This project extensively utilized multiple Language Models (LLMs). While the results are valuable, they offer a broad overview of each project, theme, and the Federal AI portfolio. This analysis exclusively covers publicly released projects.
Reports: The methodology documented below has been utilized to create the following reports:
- Machine-readable consolidated dataset
- AI Themes throughout the Federal Government
- Complete Text of AI Projects by Agencies
- Summarized GPT Text for Project Descriptions
- Prominent AI Highlights across Departments
Projects were downloaded from AGENCY INVENTORIES OF AI USE CASES, provided by the National AI Initiative Office (NAIIO).
Consistent with this principle of transparency, EO 13960 directed agencies to conduct an annual inventory of their AI use cases, and to publish them to the extent possible. As specified by EO 13960, agencies' inventories are not required to include AI use cases that are classified, sensitive, used in defense or national security systems, used by the Department of Defense or Intelligence Community, embedded within common commercial products, or used for AI research and development activities.
Each inventory was downloaded and saved and cleaned to a common format. All data were collected on September 1st, 2023 and may not reflect any updates on ai.gov or linked reports. Basic stats, information, and links can be found here.
Most Departments provided some form of machine readable table or easily parsed web version. Some only provided PDFs (Commerce, Energy, Treasury, Justice, NASA) and these required considerable human parsing along with a custom script using Camelot. Entries that were exact duplicates in the Title and Summary were removed HHS (6), DOS (6), DOE (5).
After manual parsing and cleanup the record level data was saved to data/record_level_information_FedAI_2023.csv
Use_Case_ID
(automatically assigned by this project)Department_Code
(Acronym for the Department)Agency
(or Bureau / Institute as appropriate)Office
(or Center as appropriate)Title
Summary
Development_Stage
(if provided)Techniques
(methodologies, models, or AI techniques if provided)Source_Code
(link to source code if provided)
Counts per Department / Independent agency are
Department/Agency | Count |
---|---|
Department of Energy | 178 |
Department of Health and Human Services | 157 |
Department of Commerce | 49 |
Department of Homeland Security | 41 |
Department of Veterans Affairs | 40 |
Department of Agriculture | 39 |
Department of Interior | 38 |
National Aeronautics and Space Administration | 33 |
Department of State | 31 |
Department of Labor | 18 |
Department of Transportation | 14 |
Department of Treasury | 14 |
Social Security Administration | 14 |
U.S. Agency for International Development | 14 |
U.S. General Services Administration | 12 |
National Archives and Records Administration | 5 |
Department of Justice | 4 |
U.S. Office of Personnel Management | 4 |
U.S. Environmental Protection Agency | 3 |
Department of Housing and Urban Development | 1 |
Department of Education | 1 |
A high powered LLM (in this case Chat GPT: gpt-3.5-turbo) was used to summarize each response as the quality of the summaries provided by the agencies is variable. For example, consider the response from the IRS and the associated summary:
Projected Contract Award Date Web App Projected contract award dates are generated with a machine learning model that statistically predicts when procurement requests will become signed contracts. Input data includes funding information, date / time of year, and individual Contract Specialist workload. The model outputs projected contract award timeframes for specific procurement requests. 'When will a contract be signed?' is a key question for the IRS and generally for the federal government. This tool gives insight about when each request is likely to turn into a contract. The tool provides a technique other federal agencies can implement, potentially affecting $600 billion in government contracts. Weblink: https://www.irs.gov/newsroom/irs- announces-use-of-projected-contract-award-date-web-app-that-predicts- when-contracts-will-be-signed.
The IRS has developed a web app that uses a machine learning model to predict when procurement requests will become signed contracts. The tool provides valuable insight for the IRS and other federal agencies on when contracts are likely to be signed, potentially impacting $600 billion in government contracts.
Using the summarized response, we iteratively asked for a set of high level topics. Each project was scored across the themes holistically by asking the LLM to return a JSON object. Observationally, this is a high recall medium precision task, so we further refined each positive response with an explanation of why the theme matched the project. By taking the final refinement step, about 25% of the projects were removed from their original theme. Project could belong to any number of themes, including none.
Theme | Projects | |
---|---|---|
π¬ | Scientific Research | 417 |
π§ | Infrastructure | 149 |
π³ | Environmental | 117 |
π | Geospatial | 111 |
π₯ | Healthcare | 110 |
π | Cyber Intelligence | 34 |
π€ | Customer Service Or Engagement | 31 |
π‘ | Threat Intelligence | 30 |
π | Language Services | 23 |
π΅οΈββοΈ | Fraud | 13 |
π± | Wearables | 2 |
The projects were also analyzed with respect to a set of GSA categories.
Icon | Category | Count |
---|---|---|
π¬ | Science and Technology | 522 |
π | Health and Medical | 158 |
β‘ | Energy | 75 |
π | Environment and Natural Resources | 68 |
π§ | Mission-enabling | 55 |
π | Transportation | 23 |
π₯ | Veteran Care and Services | 22 |
π | Finance, Economy | 15 |
π | Law and Justice | 14 |
π | Space | 12 |
π | Grants | 12 |
π | Emergency Management | 12 |
π | Zoological | 8 |
π | General Admin | 7 |
π | Diplomacy and Trade | 5 |
π± | Education and Workforce | 4 |
π | Benefit Programs | 2 |
To interactively explore the dataset visit:
To get a sense of the general clustering, each of the summarized projects were embedded through the latest OpenAI model (text-embedding-ada-002) and projected onto two dimensions using UMAP. A numpy array containing the embeddings is saved here. Topics were loosely clusted in the reduced space and labels were assigned to the clusters via KeyBERT.
Costs were calculated from a final run of the program, intermediate API calls during the exploration phase were not recorded.
Cost : $1.51
Tokens : 755,928
Calls : 2860
Each day, a github action is called to check the hash of the ai.gov source website. If the latest hash has not changed from ee6b92b6c6514b4a4f855b7c83b9c52f
then the data is up-to-date. If the hash has changed, some aspect of the website has been updated though it may not reflect a change in the data.