Purpose: Annotating is considered a ‘scholarly primitive’ among different fields in the humanities. Nevertheless, the debate on digital annotations has mostly focused on the annotation of textual data, whereas existing models for representing annotations of images still lack sufficient semantic richness to anchor the annotation itself to multiple conceptual levels. We address the challenge of defining a data model to overcome the problem of ‘semantic deficit’ in this application domain. Finally, we implement an annotation client for testing multi-level semantic annotations. Methodology: To define a data model for representing digital annotations, we analysed applications which support annotation images through IIIF protocol, focusing on digital representations of palimpsests. We then extended the Web Annotation Data Model by introducing domain standards such as LRMer, CIDOC-CRM, and HiCO. We also validated the model through SPARQL queries corresponding to five competency questions to report on satisfiability. Finally, we developed a prototype annotation client as a plugin for Mirador to evaluate its performances in real-world scenarios. Findings: The results indicate that our model can effectively disambiguate between a target image and multiple conceptual levels of the entity itself, proving to be decisive in the representation of entities that coexist in the same material item (e.g., palimpsests). Additionally, the model allows users to describe annotations as interpretative acts, incorporating scholarly criteria and multiple viewpoints. An interface plugin enables scholars without technical expertise to create structured annotations that comply with the model. Value: The proposed approach facilitates the detailed management of the relationships between digital resources and their annotations, improving interoperability and information accessibility in the Semantic Web domain. Future developments will concern further extensions of the model, considering information about versioning, provenance, and authoritativeness of the digital annotations on images, as well as support for meta-annotations and iconological levels of interpretation.

Pedretti, C.T., Bocchi, M.F., Tomasi, F., Vitali, F. (2024). What Do We Annotate When We Annotate? Towards a Multi-Level Approach to Semantic Annotations. Amsterdam : IOS Press BV [10.3233/ssw240030].

What Do We Annotate When We Annotate? Towards a Multi-Level Approach to Semantic Annotations

Pedretti, Carlo Teo;Bocchi, Maria Francesca;Tomasi, Francesca;Vitali, Fabio
2024

Abstract

Purpose: Annotating is considered a ‘scholarly primitive’ among different fields in the humanities. Nevertheless, the debate on digital annotations has mostly focused on the annotation of textual data, whereas existing models for representing annotations of images still lack sufficient semantic richness to anchor the annotation itself to multiple conceptual levels. We address the challenge of defining a data model to overcome the problem of ‘semantic deficit’ in this application domain. Finally, we implement an annotation client for testing multi-level semantic annotations. Methodology: To define a data model for representing digital annotations, we analysed applications which support annotation images through IIIF protocol, focusing on digital representations of palimpsests. We then extended the Web Annotation Data Model by introducing domain standards such as LRMer, CIDOC-CRM, and HiCO. We also validated the model through SPARQL queries corresponding to five competency questions to report on satisfiability. Finally, we developed a prototype annotation client as a plugin for Mirador to evaluate its performances in real-world scenarios. Findings: The results indicate that our model can effectively disambiguate between a target image and multiple conceptual levels of the entity itself, proving to be decisive in the representation of entities that coexist in the same material item (e.g., palimpsests). Additionally, the model allows users to describe annotations as interpretative acts, incorporating scholarly criteria and multiple viewpoints. An interface plugin enables scholars without technical expertise to create structured annotations that comply with the model. Value: The proposed approach facilitates the detailed management of the relationships between digital resources and their annotations, improving interoperability and information accessibility in the Semantic Web domain. Future developments will concern further extensions of the model, considering information about versioning, provenance, and authoritativeness of the digital annotations on images, as well as support for meta-annotations and iconological levels of interpretation.
2024
Knowledge Graphs in the Age of Language Models and Neuro-Symbolic AI.
370
385
Pedretti, C.T., Bocchi, M.F., Tomasi, F., Vitali, F. (2024). What Do We Annotate When We Annotate? Towards a Multi-Level Approach to Semantic Annotations. Amsterdam : IOS Press BV [10.3233/ssw240030].
Pedretti, Carlo Teo; Bocchi, Maria Francesca; Tomasi, Francesca; Vitali, Fabio
File in questo prodotto:
File Dimensione Formato  
SSW-60-SSW240030.pdf

accesso aperto

Descrizione: Contributo in Atti di Convegno
Tipo: Versione (PDF) editoriale / Version Of Record
Licenza: Licenza per Accesso Aperto. Creative Commons Attribuzione (CCBY)
Dimensione 867.03 kB
Formato Adobe PDF
867.03 kB Adobe PDF Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11585/1009700
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus ND
  • ???jsp.display-item.citation.isi??? ND
social impact