Skip to main content

Digital Collection Assessment & Use

by Tam Troup
This lesson introduces students to metadata development and reuse, using the DPLA digital collections and the DPLA API to harvest metadata. It includes handouts.
Digital Collection Assessment & Use
·
Contributors (10)
T
J
J
S
A
& 5 more
Published
Sep 19, 2019
DOI
10.21428/65a6243c.42628a94

Session Specifics

Three hour workshop with two ten minute breaks

Instructional Partners

Discipline-specific teaching faculty

Audience

Mid-track undergraduates majoring in fields requiring a high-level of visual literacy about the cultural or physical world (anthropology, art, history, biology, geology, pre-medicine, etc.); graduate students; MLIS students learning about metadata development.

Curricular Context

This course was designed using a crosswalking method shared by Marcia Lei Zeng (Post-It note paper to tag, label, and simulate the crosswalking experience) and expanded upon in Metadata workshops held by the author.

The course supports the identification of metadata by individuals with local knowledge or subject-specific knowledge.

Learning Outcomes

The primary goal of this lesson is to familiarize students with the methods of metadata development and reuse, and to instill the confidence in their ability to contribute to curated knowledge. This is accomplished by following the following four steps:

  1. Introduction to the effort required to create simple and robust metadata.

  2. Use of the DPLA API to harvest metadata.

  3. Practice of empathy in the assessment and use of digital collections, identify bias and how bias may be addressed, and identify gaps in access.

  4. Enhancement of metadata for use by a specific user group, specifically through the identification of keywords, enhanced description, coverage, or additional fields.

Preparation

Audience

Instructor

Prepare printed images from the DPLA:

  • Images for students to discuss -- instructor can select any image shared with the DPLA, note that these image will be discussed in section 1.0 and cognitive biases will be assessed. Instructor should select images which could challenge student’s identification of the content.

API practice

Note: it is very important for the instructor to gain some experience working with the API prior to teaching the course; the DPLA documentation is excellent.

Create Sample datasets (backup).

Prepared bookmarks/open tabs/software:

DPLA

dp.la API Codex

Excel

OpenRefine

RegEx

REST API Tutorial

Tab Save

Materials

  • Computer with World Wide Web access

  • Document software (Microsoft, Google, etc.)

  • Paper, pencils, Post-Its of various colors

  • DPLA API key

  • Supplementary handouts (see Additional Instructional Materials below)

Session Outline

1.0 Developing Metadata

Introduction to the effort required to create simple and robust metadata

  1. Students —

    • Organize into small groups, each group selects a single image to discuss.

  2. Instructor —

    • Introduce the Panofsky-Shatford matrix (generics, specifics, abstracts)—see handout in Additional Instructional Materials below.

    • Discuss cognitive biases and the effect on the transfer of information—see handout .

  3. Students —

    • Discuss two or three cognitive biases and the potential effect of these biases on the process of creating metadata for the image in hand.

    • Discuss in small groups how the image could be “tagged” what “labels” could apply to the “tags,” what research may be required to identify more information, etc.

    • Record the labels and the tags on Post-Its.

  4. Students —

    • Reconvene as a large group, and crosswalk/map all the groups labels/concepts.

    • Discuss difficulties mapping labels.

  5. Instructor —

    • Introduce the idea of standardized information (examples: hand drawn map, 19th century map, Rand-McNally map, Google map).

    • Discuss the ideas communicated through standardized information.

    • Share and discuss the standards funnel — see handout in Additional Instructional Materials below.

    • How is information created by humans?

    • How is information discovered and used by humans?

    • What information can be contributed?

    • How is information crawled and used by machines?

  6. Students —

    • Use a pro & con grid (on paper or whiteboard) to list the benefits and drawbacks of standardized or centrally managed metadata — see handout in Additional Instructional Materials below.

2.0 Harvesting Metadata

Use DPLA API to harvest metadata

  1. Instructor —

    • Briefly introduce the idea of exposing metadata for use and reuse.

  2. Students —

    • Discuss possible ethical and legal issues to consider when using metadata created by others.

  3. Instructor —

    • Demonstrate harvest and creating calls.

  4. Students —

    • Collect affiliated images from JSON results with an image scraper (e.g., Tab Save extension).

  5. Students —

    • Use a concept mapping approach to describe the systems (machines, protocols, and people) used to harvest metadata.

3.0 Assessing Metadata

Practice empathy in the assessment and use of digital collections, identify bias and how bias may be addressed, and identify gaps in access.

  1. Students —

    • Discuss communities and attributes of communities (cultural, professional, knowledge domains, sub-Reddits, Twitter cultures, etc.)

  2. Instructor —

    • Demonstrate various tools and methods of metadata assessment (word cloud guessing, sorting, visualization, etc.)

  3. Students —

    • Use various tools and methods to investigate the metadata.

  4. Instructor —

    • Demonstrate the assessment rubric — see handout.

  5. Students —

    • Investigate a few original collections that metadata is drawn from, and use the rubric (available as part of Handouts in Additional Instructional Materials below) to assess the standards and policies related to metadata development.

  6. Students —

    • Discuss and Identify possible cultural bias, knowledge gaps, inclusive/exclusive approaches to the metadata.

  7. Students —

    • Identify gaps in access — Who may not discover this collection because some information is missing? Who is the primary contributor? Who is the primary audience of this collection? Who is excluded from sharing their knowledge? Are the labels useful for finding information?

4.0 Enhancing Metadata

Enhance metadata for use by a specific user group. Keywords, enhanced description, coverage, or additional fields.

  1. Instructor —

    • Discuss sharing metadata with the collection creators and working collaboratively to enhance metadata.

  2. Students —

    • Use Criterion 3 of the assessment rubric to identify methods of making knowledge about the dataset more robust.

  3. Instructor —

    • Demonstrate various tools and methods of refining and enhancing data (Excel, OpenRefine, or Regex depending on instructor’s experience and comfort level).

  4. Students —

    • Identify methods of making access to the collection more robust, identify schemas, local elements, controlled vocabularies, local terms, which might be used.

Assessment

  • Students identify issues which surround the curation of information by discussing biases and knowledge gaps.

  • Students successfully use tools to collect and refine data.

  • Students are empowered to research and to contribute to record-keeping.

Reflection

Metadata records are living records, which should be revisited and revised when new information is acquired or new perspectives require re-cataloging. As the myriad communities responsible for metadata creation, curation, and use continues to evolve, they need to consider best practices which allow for knowledge experts to collaborate and contribute. Perhaps we need interfaces which allow experts to fork and create enhanced collections. Whatever solutions are devised, subject experts should be involved as thinkers, contributors, and collaborators.

Additional Instructional Materials

Handouts

Handouts

Comments
0
comment

No comments here