Webinars
Chinese Students in Archives: Radical Empathy & Metadata


Ruohua Han, Yingying Han
This webinar explores the representation of early Chinese international students at the University of Illinois, focusing on archival materials and metadata. Using the principle of radical empathy, the study identifies key issues— inaccuracies, inconsistencies, and inappropriateness—affecting discoverability and engagement. Introducing the “Three-I Framework,” the session highlights systemic factors shaping archival records and advocates for holistic, empathetic strategies beyond metadata enhancement to address subtle marginalization and improve access to these historical narratives.
Centering LGBTQ+ Voices in Catalog Research


Brian Dobreski, Karen Snow
Libraries are addressing barriers faced by marginalized communities through improved catalog metadata. The Inclusive Catalog Use Lab (ICUL) focuses on user-based research to understand LGBTQ+ needs. Dr. Karen Snow and Dr. Brian Dobreski will share ICUL’s findings, highlighting user studies, focus groups, and solutions to support inclusive access for diverse communities.
On the Duality of Provenance Metadata

Amelia Acker
Explore how metadata is exploited to craft deceptive narratives and influence public discourse on platforms. Explore tactics used by bad actors to manipulate social media and evade content moderation by obfuscating provenance information. Understand metadata's duality in preventing data misuse and its exploitation for disinformation campaigns. Discover connections between responsible data management in open access and potential misuse. Information professionals and researchers play a crucial role by promoting data literacy and critical engagement with digital content.
Metadata in Scholarly Research: Collection, Management, and Standardization

Matt Mayernik
In this webinar, Dr. Mayernik will discuss findings and insights from multiple projects focused on metadata collection, management, and standardization within the context of scientific data curation. The webinar will discuss important socio-technical dynamics that influence what metadata get generated in different situations, and how data archives and repositories work to ensure quality and standardization.
AI & NLP for Open-Source Archival Linked Data Workflows

Jennifer Proctor
Join our webinar to harness AI and NLP for efficient archival linked data creation. Explore computer vision, NLP, and Python techniques to transform metadata into linked data, saving time and uncovering hidden collection details. Gain insights on tools for automatic data generation, appreciate the value of linked data in rediscovery, and learn its role as a new gateway to your collections.
Critical Data Modeling

Karen M. Wickett
In this DCMI webinar, Dr Karen Wickett will present critical data modeling: the use of data modeling and systems analysis techniques to build critical interrogations of information systems. Dr Wickett will describe the results of critical data modeling of a police arrest record dataset and discuss how conceptual modeling can help us synthesize critical information studies and identify new opportunities for modeling and critique.
Getting Started With SQLite

Nishad Thalhath
SQLite is a powerful relational database engine widely used in many applications. SQLite is the most used database engine in the world and implements most of the SQL standards. This webinar will cover the features of SQLite, the basics of SQL, the SQLite file format, the SQLite command line interface, and popular tools for managing and interacting with SQLite databases.
Linking Challenges for Pseudonyms

Charlene Chou
This presentation examines and explores how to improve the linking relationships of persons with more than one identity in VIAF, LCNAF (LC National Authority File) and Wikidata through definition comparisons and case studies in terms of the semantic web.
Querying Wikidata: All the Knowledge in the World

Daniel Garijo Verdejo
Knowledge Graphs have become a common asset for representing world knowledge in data driven models and applications. Wikidata is one of the largest crowdsourced Knowledge Graphs, with over 90 million entities and 1.5 billion edges.
Centering Relationality in Indigenous Knowledge Organization

Sandy Littletree
In this talk, Dr. Littletree will discuss a conceptual framework that centers on the concept of relationality, a framework that can be used to understand Indigenous systems of knowledge in the LIS context.
Applying FAIR Principles to Ontologies

María Poveda-Villalón
About the webinar This webinar addresses ontologies for the semantic web and how FAIR principles could be applied to ontologies. This includes metadata best practices and registries to publish the ontologies as Linked Open Vocabularies. Objectives are: To become familiar with ontologies for the semantic web and linked data principles; To become familiar with FAIR principles; To understand how FAIR principles could be applied to ontologies.
Natural history data curation: maintaining infrastructures and turning documents into datasets.

Andrea Thomer
About the webinar Natural history collections constitute an important and complex knowledge infrastructure. The heterogeneous data in these collections include specimens, genomes, field notes, ecological monitoring data, and more -- all of which have huge potential for reuse in biodiversity, medicine, climate science, agriculture, and many other domains. However, facilitating this integrative reuse requires first maintaining the fragile infrastructures used to store and share these data, and second, novel approaches to migrating data from legacy (and often analog) formats into datasets that are more fit-for-use.
Information and Knowledge Organisation in Digital Humanities: Global Perspectives


Koraljka Golub, Ying-Hsang Liu
This webinar, organized by the Dublin Core Metadata Initiative Education Committee, will provide an overview of and reflections on the edited book, Information and Knowledge Organisation in Digital Humanities: Global Perspectives. This volume explores the potential uses of, and challenges involved in, applying the organisation of information and knowledge in the various areas of Digital Humanities. By focussing on how information is described, represented, and organised in both research and practice, this work furthers the transdisciplinary nature of digital humanities.
Introduction to OpenRefine

Elizabeth Wickes
About the webinar OpenRefine is a freely available desktop application for cleaning and processing data. This webinar will introduce the basic features of OpenRefine and discuss its use in some common scenarios for metadata cleaning and improvement.
Modeling Intangible Entities in the Cultural Domains for Digital Archiving

Shigeo Sugimoto
About the webinar There is a variety of new cultural domains for which we need to design metadata schemas. It is widely recognized that data models are crucial to design metadata schemas – we need to identify objects for which we create metadata. This webinar aims to present a set of generalized data models designed for metadata about resources in cultural domains such as intangible cultural heritage and popular culture. The presenter will first introduce the generalized data models and then discuss data models designed in the Media Arts Database project and related projects in which they are involved.
Introduction to the DDI Metadata Standard




Jane Fry, Arofan Gregory, Jared Lyle, Barry Radler
About the webinar In this presentation, we discuss the Data Documentation Initiative (DDI) standard for describing the data produced by surveys and other observational methods in the social, behavioral, economic, and health sciences. We spend the bulk of our time highlighting DDI-Codebook and DDI-Lifecycle. DDI-Codebook is a more light-weight version of the standard, intended primarily to document simple survey data. DDI-Lifecycle is designed to document and manage data across the entire life cycle, from conceptualization to data publication, analysis and beyond. It encompasses all of the DDI-Codebook specification and extends it to more complex, linked, and longitudinal data. Based on XML Schemas, DDI-Lifecycle is modular and extensible.
Introduction to KNIME


Magnus Pfeffer, Kai Eckert
KNIME Analytics Platform is an open source software for working with all kinds of data. It uses visual workflows that are created with an intuitive, drag and drop style graphical interface, without the need for coding.
中国元数据发展:多领域应用及实践途径





Lei Sa / 萨蕾, Cuijuan Xia / 夏翠娟, Zhiqiang Wang / 王志强, Ruhua Huang / 黄如花, Lu An / 安璐
本次网络研讨会题为“中国元数据发展:多领域应用及实践途径”, 于2020年8月29日举办。会议由武汉大学信息管理学院、武汉大学宣传部、都柏林核心元数据创始计划(DCMI)教育委员会和中国图书馆学会学术委员会共同主办,武汉大学信息管理学院承办。会议使用中文进行主旨演讲、中英双语PPT呈现,萨蕾(国家图书馆)、夏翠娟(上海图书馆)、王志强(中国标准化研究院)、黄如花(中国图书馆学会、武汉大学)以及安璐(武汉大学)五位专家在会上做主旨报告。本次研讨会的中文视频以及英文同声传译视频录像可于DCMI教育委员会YouTube频道观看。
Metadata Development in China





Lei Sa / 萨蕾, Cuijuan Xia / 夏翠娟, Zhiqiang Wang / 王志强, Ruhua Huang / 黄如花, Lu An / 安璐
"Metadata Development in China", a webinar held on 29 August 2020, was co-organized by Wuhan University, the DCMI Education Committee, and the Library Society of China and hosted by the School of Information Management at Wuhan University. The webinar featured presentations in Chinese language, using slides in both Chinese and English, by Lei Sa (National Library of China), Cuijuan Xia (Shanghai Library), Zhiqiang Wang (China National Institute of Standardization), Ruhua Huang (Library Society of China and Wuhan University, and Lu An (Wuhan University). The webinar is available on YouTube both in the original Chinese and with English interpretation.
Introduction to the W3C Data Catalog Ontology (DCAT)

Peter Winstanley
The W3C Data Catalogue ontology (DCAT) is a recommendation for how to publish data catalogues on the web. It can be used by itself, of form the basis of a richer application profile. Although originally focusing on open government data, the ontology and application profiles are now seeing use in a wider set of uses.
Getty Vocabulary Program’s OpenRefine reconciliation service

Jonathan Ward
This webinar will be a tutorial on the Getty Vocabulary Program’s new OpenRefine reconciliation service for aligning data records to ULAN, AAT, and TGN, including matching terms, cleaning data, and grabbing URIs from the Vocabularies’ Linked Open Data. The first half of the presentation will be a general overview of how the service works and its benefits, and the second will focus on advanced techniques.
Wikiproject COVID-19

Tiago Lubiana
The Wikiproject COVID-19 is working to curate and organization information about COVID-19 on Wikidata and Wikipedia. The scale and pace of a global pandemic have highlighted issues around the consistent structuring of information, such as the scope of geographically-bound statements and the time period of outbreaks, along with challenges for quickly updating information on pages across multiple languages. The Wikiproject collaboration has created data models for relevant concepts, using ShEx entity schemas that can be validated via SPARQL queries. This webinar will introduce the Wikiproject and show examples of the data models at work.
SKOS - visión general y modelado de vocabularios controlados


Antoine Isaac, Juan Antonio Pastor Sánchez
SKOS (Simple Knowledge Organization Systems) es la recomendación del W3C para representar y publicar conjuntos de datos de clasificaciones, tesauros, encabezamientos de materia, glosarios y otros tipos de vocabularios controlados y sistemas de organización del conocimiento. La primera parte del webinar incluye una visión general de las tecnologías de la web semántica y muestra detalladamente los diferentes elementos del modelo SKOS. La segunda parte aborda diferentes aproximaciones para la aplicación de SKOS para representar vocabularios controlados.
SKOS - Overview and Modeling of Controlled Vocabularies


Antoine Isaac, Juan Antonio Pastor Sánchez
SKOS (Simple Knowledge Organization Systems) is the recommendation of the W3C to represent and publish datasets for classifications, thesauri, subject headings, glossaries and other types of controlled vocabularies and knowledge organization systems in general. The first part of the webinar includes an overview of the technologies of the semantic web and shows in detail the different elements of the SKOS model. The second part shows different approaches for the application of SKOS to represent controlled vocabularies.
The Current State of Automated Content Tagging: Dangers and Opportunities

Joseph Busch
There are real opportunities to use technology to automate content tagging, and there are real dangers that automated content tagging will sometimes inappropriately promote and obscure content. We’ve all heard talks about AI, but little detail about how these applications actually work. Recently I’ve been working with clients to explore the current state of the art of so-called AI technology, and to trial several of these tools with research, policy and general news content. In addition to framing the debate about whether this is AI or automation, this talk will describe how to run trials with these tools, and will show some results from actual trials.
The Role of Dublin Core™ Metadata in the Expanding Digital and Analytical Skill Set Required by Data-Driven Organisations

Steve Brewer
Many areas of our world are being subject to digitalisation as leaders and policymakers embrace the possibilities that can be harnessed through the capturing and exploiting of data. New business models are being developed, and new revenue streams are being uncovered that require a solid and recognised data competence capacity. This process involves bringing together a range of traditional disciplines from computing and engineering to business management and data science. Facilitating successful collaboration amongst these participants in order to create new cyber-physical systems can be achieved through a range of tools, but chief among them will be the application of robust, trustworthy, and reusable data. The Dublin Core™ Metadata Initiative provides a well-established schema of terms that can be used to describe such data resources. As more organisations in diverse fields awaken to the benefits of digitalisation they will need to embrace data capture. Acquiring data science skills and competences at all levels of the organisation, and as an ongoing process over time, will be critical for their future. Whilst some elements will be specific, other skills will be common across sectors. Metadata foundations can obviously help with these commonalities. Applying a similar structured approach to understanding and supporting skills acquisition will contribute significantly to the future success of data-driven organisations.
Introduction to Metadata Application Profiles

Karen Coyle
Successful data sharing requires that users of your data understand the data format, the data semantics, and the rules that govern your particular use of terms and values. Sharing often means the creation of “cross-walks” that transfer data from one schema to another using some or all of this information. However, cross-walks are time-consuming because the information that is provided is neither standardized nor machine-readable. Application profiles aim to make sharing data more efficient and more effective. They can also do much more than facilitate sharable data: APs can help metadata developers clarify and express design options; they can be a focus for consensus within a community; they can drive user interfaces; and they can be the basis for quality control. Machine-actionable APs could become a vital tool in the metadata toolbox and there is a clear need for standardization. Communities such as Dublin Core™ and the World Wide Web Consortium are among those working in this area.
Understanding and Testing Models with ShEx


Eric Prud’hommeaux, Tom Baker
Every structured exchange requires consensus about the structure. The Shape Expressions (ShEx) language captures these structures in an intuitive and powerful syntax. From metadata description (e.g. DDI) to data description (e.g. FHIR), ShEx provides a powerful schema language to develop, test and deploy shared models for RDF data. This tutorial will explore the utility and expressivity of ShEx.Presented with side-by-side examples of schema and data, the audience will see how to use ShEx to solve every-day problems. The presentation will use multiple implementations of ShEx in order to leave the participants with enough familiarity to get started using ShEx on their own.
A Linked Data Competency Framework for Educators and Learners

Marcia Zeng
Linked Data is recognized as one of the underpinnings for open data, open science, and data-driven research and learning in the Semantic Web era. Questions still exist, however, about what should be expected as Linked Data related knowledge, skills, and learning outcomes, and where to find relevant learning materials. This webinar will introduce a competency framework that defines the knowledge and skills necessary for professional practice in the area of Linked Data, developed by the Linked Data for Professional Educators (LD4PE) project and funded by the Institute of Museum and Library Services (IMLS).
Save the Children Resource Libraries: Aligning Internal Technical Resource Libraries with a Public Distribution Website


Joseph Busch, Branka Kosovac
Save the Children (STC) is an international NGO that promotes children’s rights, provides relief and helps support children across the globe. With international headquarters in London, STC has 30 national members and supports local partners operating in over 100 countries worldwide. STC International maintains technical infrastructures that are available to members and local partners including SharePoint, Drupal and other information management applications. An effort to specify and implement a common resource library for curating and sharing internal technical resources has been underway since November 2015. This has included an inventory of existing (but heterogeneous) resource libraries on Save the Children’s work in the thematic area of Health and Nutrition, and agreement on a common metadata specification and some controlled vocabularies to be used going forward. This internal technical resource library has been aligned with Save the Children’s Resource Centre, a public web-accessible library that hosts comprehensive, reliable and up-to-date information on Save the Children’s work in the thematic areas of Child Protection, Child Rights Governance and Child Poverty.
How to Design Build Semantic Applications with Linked Data

Dave Clarke
This webinar will demonstrate how to design and build rich end-user search and discovery applications using Linked Data. The Linked Open Data cloud is a rapidly growing collection of publicly accessible resources, which can be adopted and reused to enrich both internal enterprise projects and public-facing information systems.
The webinar will use the Linked Canvas application as its primary use-case. Linked Canvas is an application designed by Synaptica for the cultural heritage community. It enables high-resolution images of artworks and artifacts to be catalogued and subject indexed using Linked Data. The talk will demonstrate how property fields and relational predicates can be adopted from open data ontologies and metadata schemes, such as DCMI, SKOS, IIIF and the Web Annotation Model. Selections of properties and predicates can then be recombined to create Knowledge Organization Systems (KOS) customized for business applications. The demonstration will also illustrate how very-large-scale subject taxonomies and name authority files, such as the Library of Congress Name Authority File, DBpedia, and the Getty Linked Open Data Vocabularies collection, can be used for content enrichment and indexing.
There will be a brief discussion of the general principles of graph databases, RDF triple stores, and the SPARQL query language. This technical segment will discuss the pros and cons of accessing remote server endpoints versus cached copies of external Linked Data resources, as well as the challenge of providing high-performance full text search against graph databases.
The webinar will conclude by providing a demonstration of Linked Canvas to illustrate various end-user experiences that can be created with Linked Data technology: faceted search across data collections; pinch and zoom navigation inside images; the exploration of concepts and ideas associated with specific points of interest; the discovery of conceptually related images; and the creation of guided tours with embedded audio-visual commentary.
Me4MAP: Um método para o desenvolvimento de perfis de aplicação de metadados

Mariana Curado Malta
Um perfil de aplicação de metadados (MAP) é um constructo que fornece um modelo semântico para publicação de dados na Web de Dados. Este modelo semântico não é mais do que um modelo de dados com a definição de propriedades e restrições às propriedades. Cada propriedade é apresentada com um termo de um vocabulário RDF associado, com a definição do domínio e contra-domínio e ainda a sua cardinalidade. De acordo com o documento da DCMI “Os níveis de interoperabilidade para os metadados Dublin Core” um MAP é um constructo que potencia a interoperabilidade semântica dos dados. O que isto realmente significa é que quando uma comunidade de prática decide pôr-se de acordo em seguir um MAP, isto é, em seguir um conjunto de regras para publicar os seus dados como Linked Open Data, isso permite que os dados publicados na LOD cloud sejam processados automaticamente por agentes de software.
Um MAP é portanto um constructo de muita importância, e por isso é essencial a existência de um método para o seu desenvolvimento. É muito importante fornecer aos modeladores de MAPs uma base comum de entendimento para que o desenvolvimento de um MAP deixe de ser um conjunto não sistemático de actividades e passe a ser algo mais organizado de forma a resultar em MAPs de melhor qualidade.
Este Webinar apresenta o Me4MAP, um método para o desenvolvimento de perfis de aplicação de metadados. O Me4MAP é uma proposta que foi desenvolvida no âmbito de um projecto de doutoramento e que está ainda a ser testada e aperfeiçoada. É uma proposta que parte de uma perspectiva de engenharia de software. E, acima de tudo, é um um ponto de partida para o estudo e o desenho de métodos para o desenvolvimento de MAPs.
Me4MAP - A method for the development of metadata application profiles

Mariana Curado Malta
A metadata application profile (MAP) is a construct that provides a semantic model for enhancing interoperability when publishing to the Web of Data. With a MAP, each property is defined as an RDF vocabulary term with the definition of domain, range, and cardinality. According to the DCMI document "Interoperability Levels for Dublin Core™ Metadata", a MAP is a construct that enhances semantic interoperability. Therefore, when a community of practice agrees to follow a MAP's set of rules for publishing data as Linked Open Data, it makes it possible for such data published the LOD cloud to be processed automatically by software agents.
Nailing Jello to a Wall: Metrics, Frameworks, and Existing Work for Metadata Assessment

Christina Harlow
With the increasing number of repositories, standards and resources we manage for digital libraries, there is a growing need to assess, validate and analyze our metadata - beyond our traditional approaches such as writing XSD or generating CSVs for manual review. Being able to further analyze and determine measures of metadata quality helps us better manage our data and data-driven development, particularly with the shift to Linked Open Data leading many institutions to large-scale migrations. Yet, the semantically-rich metadata desired by many Cultural Heritage Institutions, and the granular expectations of some of our data models, makes performing assessment, much less going on to determine quality or performing validation, that much trickier. How do we handle analysis of the rich understandings we have built into our Cultural Heritage Institutions' metadata and enable ourselves to perform this analysis with the systems and resources we have?
This webinar sets up this question and proposes some guidelines, best practices, tools and workflows around the evaluation of metadata used by and for digital libraries and Cultural Heritage Institution repositories. What metrics have other researchers or practitioners applied to measure their definition of quality? How do these metrics or definitions for quality compare across examples – from the large and aggregation-focused, like Europeana, to the relatively small and project-focused, like Cornell University Library's own SharedShelf instance? Do any metadata assessment frameworks exist, and how do they compare to the proposed approaches in core literature in this area, such as Thomas Bruce and Diane Hillmann's 2004 article, "The Continuum of Metadata Quality"? The Digital Library Federation Assessment Interest Group (or DLF AIG) has a Metadata Working Group that has been attempting to build a framework that can be used broadly for digital repository metadata assessment - the state of this work, and the issues it has raised, will be discussed in this webinar as well. Finally, how does one begin to approach this metadata assessment – what tools, applications, or efforts for performing assessment exist for common digital repository applications or data publication mechanisms?
This webinar hopes to provide some solutions to these questions within existing literature, work, and examples of metadata assessment happening 'on the ground'. The goal is for webinar participants to walk away prepared to handle their own metadata assessment needs by using the existing work outlined and being better aware of the open questions in this domain.
Data on the Web Best Practices: Challenges and Benefits



Bernadette Farias Lóscio, Caroline Burle dos Santos Guimarães, Newton Calegari
There is a growing interest in the publication and consumption of data on the Web. Government and non-governmental organizations already provide a variety of data on the Web, some open, others with access restrictions, covering a variety of domains such as education, economics, ECommerce and scientific data. Developers, journalists, and others manipulate this data to create visualizations and perform data analysis. Experience in this area reveals that a number of important issues need to be addressed in order to meet the requirements of both publishers and data consumers.
Boas Práticas para Dados na Web: Desafios e Benefícios



Bernadette Farias Lóscio, Caroline Burle dos Santos Guimarães, Newton Calegari
Existe um interesse crescente na publicação e consumo de dados na Web. Organizações governamentais e não-governamentais já disponibilizam uma variedade de dados na Web, alguns abertos, outros com restrições de acesso, abrangendo diversos domínios como educação, economia, segurança, patrimônio cultural, eCommerce e dados científicos. Desenvolvedores, jornalistas e outras pessoas manipulam esses dados para criar visualizações e realizar análises de dados. A experiência neste tema revela que é necessário abordar várias questões importantes a fim de satisfazer os requisitos tanto dos publicadores como dos consumidores de dados.
Neste webinar discutiremos os principais desafios enfrentados pelos publicadores e consumidores de dados ao compartilharem dados na Web. Também introduziremos o conjunto de Boas Práticas, propostas pelo W3C, para enfrentar esses desafios. Finalmente, discutiremos os benefícios de envolver os publicadores de dados na utilização das Boas Práticas, bem como melhorar a forma que os conjuntos de dados são disponibilizados na Web.
From MARC silos to Linked Data silos? Data models for bibliographic Linked Data

Osma Suominen
Many libraries are experimenting with publishing their metadata as Linked Data to open up bibliographic silos, usually based on MARC records, to the Web. The libraries who have published Linked Data have all used different data models for structuring their bibliographic data. Some are using a FRBR-based model where Works, Expressions and Manifestations are represented separately. Others have chosen basic Dublin Core, dumbing down their data into a lowest common denominator format. The proliferation of data models limits the reusability of bibliographic data. In effect, libraries have moved from MARC silos to Linked Data silos of incompatible data models. There is currently no universal model for how to represent bibliographic metadata as Linked Data, even though many attempts for such a model have been made.
Modelado y publicación de los vocabularios controlados del proyecto UNESKOS

Juan Antonio Pastor Sánchez
Se presentan los procesos de modelado y publicación de los vocabularios del proyecto UNESKOS aplicando tecnologías de la Web Semántica. Más específicamente, los vocabularios representados son el Tesauro de la UNESCO y la Nomenclatura de Ciencia y Tecnología. Ambos vocabularios están publicados como conjuntos de datos RDF con una estructura para facilitar su consulta y reutilización según los principios Linked Open Data. También se muestra como se ha aplicado la norma ISO-25964 para representar el tesauro de la UNESCO utilizando conjuntamente SKOS y la ontología ISO-THES. Asímismo se analizarán las soluciones tecnológicas empleadas para el proceso de publicación y consulta de ambos vocabularios
This webinar presents the modeling and publishing process of the vocabularies for the UNESKOS project by applying Semantic Web technologies. More specifically, the vocabularies represented are the UNESCO Thesaurus and the Nomenclature for fields of Science and Technology. Both vocabularies are published as RDF datasets with a structure that allows its query and reuse according to the principles of Linked Open Data. The webinar will demonstrate the application of ISO-25964 standard to represent the UNESCO thesaurus using SKOS and the ISO-THES ontology. Technological solutions used for the project will also be discussed.
SKOS in Two Parts - Part 2: Publishing SKOS concept schemes with Skosmos

Joachim Neubert, Osma Suominen
In the past seven years, SKOS has become a widely recognized and used common interchange format for thesauri, classifications, and other types of vocabularies. This has opened a huge opportunity for the development of generic tools and methods that should apply to all vocabularies that can be expressed in SKOS. While expensive, proprietary or custom-developed solutions aimed at one particular thesaurus or classification have been dominant, now more and more open source tools are being created to deal with various aspects of vocabulary management.
SKOS in Two Parts - Part 1: Change Tracking in Knowledge Organization Systems with skos-history

Joachim Neubert, Osma Suominen
In the past seven years, SKOS has become a widely recognized and used common interchange format for thesauri, classifications, and other types of vocabularies. This has opened a huge opportunity for the development of generic tools and methods that should apply to all vocabularies that can be expressed in SKOS. While expensive, proprietary or custom-developed solutions aimed at one particular thesaurus or classification have been dominant, now more and more open source tools are being created to deal with various aspects of vocabulary management.
Linked Data Fragments: Querying multiple Linked Data sources on the Web

Ruben Verbourg
The dream of Linked Data: if we just get our data online, the promised Semantic Web will eventually rise. Everybody will be able to query our data with minimal effort. We will be able to integrate data from multiple sources on the fly. Everything will just work and data will flow freely ever after.
Creating Content Intelligence: Harmonized Taxonomy & Metadata in the Enterprise Context

Stephanie Lemieux
Many organizations have content dispersed across multiple independent repositories, often with a real lack of metadata consistency. The attention given to enterprise data is often not extended to unstructured content, widening the gap between the two worlds and making it near impossible to provide accurate business intelligence, good user experience, or even basic findability.
Schema.org: Part 2: Extending Potential and Possibilities

Richard Wallis
In this second more technical webinar in the series, Independent Consultant, Richard Wallis, explains the Schema.org extension mechanism, for external and reviewed/hosted extensions, and their relationship to the core Schema.org vocabulary. He will take an in-depth look at, demonstrate, and share experiences in designing, and creating a potential extension to the vocabulary. He will step through the process of creating the required vocabulary definition and examples files on a local system using a few simple tools then sharing them on a publicly visible temporary cloud instance before proposing to the Schema.org group.
Schema.org: Part 1 - Fit For a Bibliographic Purpose

Richard Wallis
In this first webinar in the series, Independent Consultant, Richard Wallis, traces the history of the Schema.org vocabulary, plus its applicability to the bibliographic domain. He will share the background to, and activities of, the Schema Bib Extend W3C Community Group he chairs; why it was set up; how it approached the creation of bibliographic extension proposals; and how those proposals were shaped. He will then review the current status of the vocabulary and the recent introduction of the bib.schema.org and auto.schema.org extensions. Although Richard will be using bibliographic examples, the content of this webinar will be of interest and relevance to those in other domains, and/or considering other extensions.
Implementing Linked Data in Low-Resource Conditions


Johannes Keizer, Caterina Caracciolo
Opening up and linking data is becoming a priority for many data producers because of institutional requirements, or to consume data in newer applications, or simply to keep pace with current development. Since 2014, this priority has gaining momentum with the Global Open Data in Agriculture and Nutrition initiative (GODAN). However, typical small and medium-size institutions have to deal with constrained resources, which often hamper their possibilities for making their data publicly available. This webinar will be of interest to any institution seeking ways to publish and curate data in the Linked Data World.
Keizer and Caracciolo will provide an overview of bottlenecks that institutions typically face when entering the world of open and linked data, and will provide recommendations on how to proceed. They will also discuss the use of standard and linked vocabularies to produce linked data, especially in the area of agriculture. They will describe AGRISAs, a web-based resource linking agricultural datasets as an example of linked data application resulting from the collaboration of small institutions. They will also mention AgriDrupal, a Drupal distribution that supports the production and consumption of linked datasets.
Redux: An update of a webinar first presented in 2013.
OpenAIRE Guidelines: Promoting Repositories Interoperability and Supporting Open Access Funder Mandates


Pedro Príncipe, Jochen Schirrwagen
The OpenAIRE Guidelines for Data Source Managers provide recommendations and best practices for encoding of bibliographic information in OAI metadata. They have adopted established standards for different classes of content providers: (1) Dublin Core™ for textual publications in institutional and thematic repositories; (2) DataCite Metadata Kernel for research data repositories; and (3) CERIF-XML for Current Research Information Systems.
The principle of these guidelines is to improve interoperability of bibliographic information exchange between repositories, e-journals, CRIS and research infrastructures. They are a means to help content providers to comply with funders Open Access policies, e.g. the European Commission Open Access mandate in Horizon2020, and to standardize the syntax and semantics of funder/project information, open access status, links between publications and datasets. The presenters will provide an overview of the guidelines, implementation support in major platforms and tools for validation.
Digital Preservation Metadata and Improvements to PREMIS in Version 3.0

Angela Dappert
The PREMIS Data Dictionary for Preservation Metadata is the international standard for metadata to support the preservation of digital objects and ensure their long-term usability. Developed by an international team of experts, PREMIS is implemented in digital preservation projects around the world, and support for PREMIS is incorporated into a number of commercial and open-source digital preservation tools and systems. The PREMIS Editorial Committee coordinates revisions and implementation of the standard, which consists of the Data Dictionary, an XML schema, and supporting documentation.
The PREMIS Data Dictionary is currently in version 2.2. A new major release 3.0 is due out this summer. This webinar gives a brief overview of why digital preservation metadata is needed, shows examples of digital preservation metadata, shows how PREMIS can be used to capture this metadata, and illustrates some of the changes that will be available in version 3.0.
From 0 to 60 on SPARQL queries in 50 minutes (Redux)

Ethan Gruber
This webinar provides an introduction to SPARQL, a query language for RDF. Users will gain hands on experience crafting queries, starting simply, but evolving in complexity. These queries will focus on coinage data in the SPARQL endpoint hosted by http://nomisma.org: numismatic concepts defined in a SKOS-based thesaurus and physical specimens from three major museum collections (American Numismatic Society, British Museum, and Münzkabinett of the Staatliche Museen zu Berlin) linked to these concepts. Results generated from these queries in the form of CSV may be imported directly into Google Fusion Tables for immediate visualization in the form of charts and maps.
Approaches to Making Dynamic Data Citable: Recommendations of the RDA Working Group

Andreas Rauber
Being able to reliably and efficiently identify entire or subsets of data in large and dynamically growing or changing datasets constitutes a significant challenge for a range of research domains. In order to repeat an earlier study, to apply data from an earlier study to a new model, we need to be able to precisely identify the very subset of data used. While verbal descriptions of how the subset was created (e.g. by providing selected attribute ranges and time intervals) are hardly precise enough and do not support automated handling, keeping redundant copies of the data in question does not scale up to the big data settings encountered in many disciplines today. Furthermore, we need to be able to handle situations where new data gets added or existing data gets corrected or otherwise modified over time. Conventional approaches, such as assigning persistent identifiers to entire data sets or individual subsets or data items, are thus not sufficient.
In this webinar, Andreas Rauber will review the challenges identified above and discuss solutions that are currently elaborated within the context of the working group of the Research Data Alliance (RDA) on Data Citation: Making Dynamic Data Citeable. The approach is based on versioned and time-stamped data sources, with persistent identifiers being assigned to the time-stamped queries/expressions that are used for creating the subset of data. We will further review results from the first pilots evaluating the approach.
VocBench 2.0: A Web Application for Collaborative Development of Multilingual Thesauri


Caterina Caracciolo, Armando Stellato
VocBench is a web-based platform for the collaborative maintenance of multilingual thesauri. VocBench is an open source project, developed in the context of a collaboration between the Food and Agricultural Organization of the UN (FAO) and the University of Rome Tor Vergata. VocBench is currently used for the maintenance of AGROVOC, EUROVOC, GEMET, the thesaurus of the Italian Senate, the Unified Astronomy Thesaurus of Harvard University, as well as other thesauri.
VocBench has a strong focus on collaboration, supported by workflow management for content validation and publication. Dedicated user roles provide a clean separation of competencies, addressing different specificities ranging from management aspects to vertical competencies in content editing, such as conceptualization versus terminology editing. Extensive support for scheme management allows editors to fully exploit the possibilities of the SKOS model, as well as to fulfill its integrity constraints.
Since version 2, VocBench is open source software, open to a large community of users and institutions supporting its development with their feedback and ideas. During the webinar we will demonstrate the main features of VocBench from the point of view of users and system administrators, and explain in what way you may join the project.
The Libhub Initiative: Increasing the Web Visibility of Libraries

Eric Miller
As a founding sponsor, Zepheira's introduction of the Libhub Initiative creates an industry-wide focus on the collective visibility of libraries and their resources on the Web. Libraries and memory organizations have rich content and resources that the Web can not see or use. The Libhub Initiative aims to find common ground for libraries, providers, and partners to publish and use data with non-proprietary, web standards. Libraries can then communicate in a way Web applications understand and Web users can see through the use of enabling technology like Linked Data and shared vocabularies such as schema.org and BIBFRAME. The Libhub Initiative uniquely prioritizes the linking of these newly exposed library resources to each other and to other resources across the Web, a critical requirement of increased Web visibility.
In this webinar, Eric will talk about the transition libraries must make to achieve Web visibility, explain recent trends that support these efforts, and introduce the Libhub Initiative — an active exploration of what can happen when libraries begin to speak the language of the Web.
The Learning Resource Metadata Initiative, describing learning resources with schema.org, and more?


Lorna Campbell, Phil Barker
The Learning Resource Metadata Initiative (LRMI) is a collaborative initiative that aims to make it easier for teachers and learners to find educational materials through major search engines and specialized resource discovery services. The approach taken by LRMI is to extend the schema.org ontology so that educationally significant characteristics and relationships can be expressed. In this webinar, Phil Barker and Lorna M. Campbell of Cetis will introduce schema.org and present the background to LRMI, its aims and objectives, and who is involved in achieving them. The webinar will outline the technical aspects of the LRMI specification, describe some example implementations and demonstrate how the discoverability of learning resources may be enhanced. Phil and Lorna will present the latest developments in LRMI implementation, drawing on an analysis of its use by a range of open educational resource repositories and aggregators, and will report on the potential of LRMI to enhance education search and discovery services. Whereas the development of LRMI has been inspired by schema.org, the webinar will also include discussion of whether LRMI has applications beyond those of schema.org.
How to pick the low hanging fruits of Linked Data


Seth Van Hooland, Ruben Verbourg
The concept of Linked Data has gained momentum over the past few years, but the understanding and the application of its principles often remain problematic. This webinar offers a short critical introduction to Linked Data by positioning this approach within the global evolution of data modeling, allowing an understanding of the advantages but also of the limits of RDF. After this conceptual introduction, the fundamental importance of data quality in the context of Linked Data is underlined by applying data profiling techniques with the help of OpenRefine. Methods and tools for metadata reconciliation and enrichment, such as Named-Entity Recognition (NER), are illustrated with the help of the same software. This webinar will refer to case-studies with real-life data which can be re-used by participants to continue to explore OpenRefine at their own pace after the webinar. The case-studies have been developed in the context of the handbook "Linked Data for Libraries, Archives and Museums", which will be published by Facet Publishing in June 2014.
Schema.org and Linked Data: Complementary Approaches to Publishing Data

Dan Brickley
Schema.org - a collaboration of the Google, Yahoo!, and Bing search engines -- provides a way to include structured data in Web pages. Since its introduction in June 2011, the Schema.org vocabulary has grown to cover descriptive terms for content such as movies, music, organizations, TV shows, products, locations, news items, and job listings. The goal of Schema.org is "to improve the display of search results, making it easier for people to find the right web pages". The Schema.org initiative has emerged as a focal point for publishers of structured data in Web pages, especially but not exclusively in the commercial sector.
Taking Library Data From Here to There

Karen Coyle
Libraries have been creating metadata for resources for well over a century. The good news is that library metadata is rules-based and that the library cataloging community has built up a wealth of knowledge about publications, their qualities, and the users who seek them. The bad news is that library practices were fixed long before computers would be used to store and retrieve the data. Library cataloging practice continues to have elements of the era of printed catalogs and alphabetized cards, and needs to modernize to take advantage of new information technologies. This metadata, however, exists today in tens of thousands of databases and there is a large sigh heard around the world whenever a librarian considers the need to make this massive change.
As with all large problems, this one becomes more tractable when broken into smaller pieces. Karen Coyle will present her "five stars of library data," an analysis of the changes needed and some steps that libraries can begin to take immediately. She will also discuss the "open world" view of the linked data movement and how this view can increase the visibility of libraries in the global information space. This webinar will give an introduction to the types of changes that are needed as well as the value that can be realized in library services. Attendees will learn of some preparatory steps have already been taken, which should confirm that libraries have indeed begun the journey "From Here to There."