Introduction
My name is Andrew Weymouth, the Digital Initiatives Librarian for University of Idaho and today I’ll be discussing two practical applications implementing computer vision that can facilitate processing large numbers of digital visual resources.
✺
The team I’m a part of in the Center for Digital Initiatives and Learning consists of three other people and a small group of student workers, collaborating with the five members of our Special Collections department to create and maintain the 150 digital collections on our main institutional repository as well as the many other, more customized fellowship projects. I also develop digital tools and workflows to enhance transcription, tagging, and image processing, in order to make the university’s audio, text, and visual resources more discoverable for researchers.

Both of the tools I’ll be speaking about today have their origins in finding practical solutions to repetitive, digital work that comes after the hands-on arranging, research, cataloguing and description work in the archive. Considering I work with technology quite a bit, I would broadly self identify as critical towards AI applications to cultural heritage work, which I think you’ll find reflective in the coding approach, where sets of algorithms and neural networks are implemented on guardrails within a programmatic workflow, to make results more scalable and reproducible.
✺
Definitions
It might be helpful to begin with some definitions of what resources these tools implement. Both tools are written in Python
, a general purpose, readable programming language, and they are hosted in GitHub
, a version controlled platform for hosting software development projects.

The first tool I’ll discuss implements an artificial intelligence resource. At the moment, most of us would associate AI with large language models
that are trained on text and reproduce human language styled output. In contrast, this tool is a more general neural network
, computer architecture in which processors are interconnected in a manner suggestive of the human brain. The second tool utilizes machine learning
which is a computational method that enables a computer to learn to perform tasks by analyzing a large dataset without being explicitly programmed, in that case to focus on pattern recognition. All of the other resources these tools utilize are Python libraries
, wrappers that distribute any of the above models as well as path utilities
, which are untrained, deterministic functions.
✺