AI Values Statement

A perspective on the uses and potential pitfalls of applying AI to Smithsonian data products

Technology is not neutral.

The use of Artificial Intelligence (AI) tools1 to describe, analyze, visualize, or aid discovery of information from Smithsonian collections, libraries, archives, and research data reflects the biases and positionality of the people and systems who built each tool, as well as those that collected, cataloged, and described any data used for their training. These tools might hold extensive value in their use at the Smithsonian, but there are issues that will limit the applicability and reliability of their use due to the way they were planned and created.

We seek to only begin AI projects2 that implement tools and algorithms that are respectful to the individuals and communities that are represented by the information in our museum, library, and archival collections. We aim to be proactive in identifying and documenting biases and methodologies when building and implementing such tools and making the documentation available to audiences that will interact with the resulting products. We recognize that technology evolves over time and that our efforts must also evolve to ensure our ethical framework stays relevant and robust. We encourage any person, community, or stakeholder involved with or affected by said tools and algorithms to provide feedback and point out any concerns.

We acknowledge the opportunities that AI tools present for cultural heritage organizations:

We urge anyone contemplating an AI project to consider:

We strive to promote the following actions when implementing AI tools:

We strive to recognize the following when implementing AI tools:

We strive to promote the following when partnering with outside organizations on AI tools or projects:

Version drafted in Spring 2022 by members of the Smithsonian AI & ML community of practice. Comments and suggestions welcomed at SI-DataScience@si.edu.

References

Bender, E. M., Friedman, B. 2018. Data Statements for Natural Language Processing: Toward Mitigating System Bias and Enabling Better Science. https://aclanthology.org/Q18-1041.

Denton, E., Hanna, A., Amironesei, R., Smart, A., Nicole, H., Scheuerman, M. K. 2020. Bringing the People Back In: Contesting Benchmark Machine Learning Datasets. Proceedings of ICML Workshop on Participatory Approaches to Machine Learning (https://arxiv.org/pdf/2007.07399.pdf).

Murphy, O., Villaespesa, E. 2020. AI: A Museum Planning Toolkit (https://themuseumsainetwork.files.wordpress.com/2020/02/20190317_museums-and-ai-toolkit_rl_web.pdf).

Schwartz, R., Dodge, J., Smith, N. A., Etzioni, O. 2019. Green AI. https://doi.org/10.48550/arxiv.1907.10597.

Stanford Special Collections and University Archives Statement on Potentially Harmful Language in Cataloging and Archival Description (https://library.stanford.edu/spc/using-our-collections/stanford-special-collections-and-university-archives-statement-potentially).

Footnotes

  1. The term “AI tools” includes a variety of technologies that seek to create decision-making software. Some examples include facial and speech recognition, machine learning based optical character recognition, language translation, natural language processing, image recognition, object detection and segmentation, and data clustering. Common commercial examples include virtual assistants such as Siri or Alexa, website search and recommendation algorithms, and tagging and identification of people in images on social media platforms.↩︎

  2. The term “AI project” refers to an intentional effort to utilize or create an AI tool in research or in an existing workflow.↩︎