<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.0 20120330//EN" "http://jats.nlm.nih.gov/publishing/1.0/JATS-journalpublishing1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:mml="http://www.w3.org/1998/Math/MathML" article-type="research-article" xml:lang="en">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">IR</journal-id>
<journal-title-group>
<journal-title>Information Research</journal-title>
</journal-title-group>
<issn pub-type="epub">1368-1613</issn>
<publisher>
<publisher-name>University of Bor&#x00E5;s</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">ir30iConf47215</article-id>
<article-id pub-id-type="doi">10.47989/ir30iConf47215</article-id>
<article-categories>
<subj-group xml:lang="en">
<subject>Research article</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>Are we there yet? Evaluation of AI-generated metadata for online information resources</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author"><name><surname>Zavalin</surname><given-names>Vyacheslav</given-names></name>
<xref ref-type="aff" rid="aff0001"/></contrib>
<contrib contrib-type="author"><name><surname>Zavalina</surname><given-names>Oksana L.</given-names></name>
<xref ref-type="aff" rid="aff0002"/></contrib>
<aff id="aff0001"><bold>Vyacheslav Zavalin</bold> is Assistant Professor in Department of Information Science, Texas Woman&#x2019;s University of North Texas, USA. He received his Ph.D. from University of North Texas, and his research interests are in cataloguing, metadata, and data analytics. Dr. Zavalin can be contacted at <email xlink:href="vzavalin@twu.edu">vzavalin@twu.edu</email></aff>
<aff id="aff0002"><bold>Oksana L. Zavalina</bold> is Professor in Department of Information Science, University of North Texas, USA. She received her Ph.D. from University of Illinois, and her research interests are in the information organization in libraries and other repositories, assessment of metadata quality and user needs. Dr. Zavalina can be contacted at <email xlink:href="Oksana.Zavalina@unt.edu">Oksana.Zavalina@unt.edu</email></aff>
</contrib-group>
<pub-date pub-type="epub"><day>06</day><month>05</month><year>2025</year></pub-date>
<pub-date pub-type="collection"><year>2025</year></pub-date>
<volume>30</volume>
<issue>i</issue>
<fpage>732</fpage>
<lpage>740</lpage>
<permissions>
<copyright-year>2025</copyright-year>
<copyright-holder>&#x00A9; 2025 The Author(s).</copyright-holder>
<license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by-nc/4.0/">
<license-p>This is an Open Access article distributed under the terms of the Creative Commons Attribution-NonCommercial 4.0 International License (<ext-link ext-link-type="uri" xlink:href="http://creativecommons.org/licenses/by-nc/4.0/">http://creativecommons.org/licenses/by-nc/4.0/</ext-link>), permitting all non-commercial use, distribution, and reproduction in any medium, provided the original work is properly cited.</license-p>
</license>
</permissions>
<abstract xml:lang="en">
<title>Abstract</title>
<p><bold>Introduction.</bold> Generative AI tools are increasingly used in creating descriptive metadata the quality of which is key for information discovery and support of information user tasks. Machine-readable online information resources such as websites naturally lend themselves to automatic metadata creation. Yet, assessments of AI-generated metadata for them are lacking. AI metadata quality research to date is limited to 2 metadata standards.</p>
<p><bold>Method.</bold> This experimental study assessed the quality of AI-generated descriptive metadata in 4 most widely used standards: Dublin core, MODS, MARC, and BIBFRAME. Three generative AI tools &#x2013; Gemini, Gemini advanced, and ChatGPT4 &#x2013; were used to create metadata for an educational website.</p>
<p><bold>Analysis.</bold> Zero-shot queries prompting AI tools to generate metadata followed the same structure and included the link to metadata scheme&#x2019;s openly accessible documentation. Comparative in-depth analysis of accuracy and completeness of entire resulting AI-generated metadata records was performed.</p>
<p><bold>Results.</bold> Overall, AI-generated metadata does not meet the quality threshold. ChatGPT performs somewhat better than 2 other tools on completeness, but accuracy is similarly low in all 3 tools.</p>
<p><bold>Conclusions.</bold> Current metadata-generating effectiveness of AI tools does not allow to conclude that involvement of human metadata experts in creation of quality (and therefore functional) metadata can be significantly reduced without strong negative impact on information discovery. </p>
</abstract>
</article-meta>
</front>
<body>
<sec id="sec1">
<title>Introduction</title>
<p>Artificial intelligence (AI) generative tools are increasingly experimented with to assess AI usability in assisting humans in descriptive metadata creation. Metadata is crucial for discoverability of all kinds of information resources. Major metadata standards designed with the goal of describing (with relatively small-scale sets of metadata elements) and providing access to online information resources include 2 versions of Dublin Core (simple (https://www.dublincore.org/specifications/dublin-core/dces/)and extended (https://www.dublincore.org/specifications/dublin-core/dcmi-terms)), and metadata object description schema (MODS) (https://loc.gov/standards/mods/) introduced in late 1990s and evolving since then. In addition, the key metadata standard in use by libraries worldwide (with hundreds of millions of metadata records currently aggregated in WorldCat database (https://search.worldcat.org/)) since late 1970s, machine readable cataloguing (MARC) (https://www.loc.gov/marc/bibliographic/), originally intended for representing analogue (text and audio-visual) information resources, is a much more comprehensive metadata element set that has been continuously evolving into a standard that supports discovery of any type of information resources, including those available online in any file format (e.g., HTML, streaming audio and video, PDF, etc.). BIBFRAME (https://www.loc.gov/bibframe/) &#x2013; launched in 2010s and intended as an alternative to MARC &#x2013; is both an ontology and a rich metadata element set designed for enabling seamless resource discovery by emphasizing relationships representation and use of Linked Data approaches in describing information resources and entities related to them: their creators, contributors, subjects, etc. Regardless of the metadata element set used, metadata creators consult guidelines designed iteratively by teams of experts and approved by metadata community of practice to support decision-making by those tasked with describing information resources to facilitate discovery and access.</p>
</sec>
<sec id="sec2">
<title>Literature Review</title>
<p>Metadata&#x2019;s effectiveness in supporting information resource discovery, as well as specific information user tasks formulated in the IFLA library reference model (LRM) is the defining feature of quality metadata (<xref rid="R19" ref-type="bibr">Riva, Lebeouf, &#x0026; &#x017D;umer, 2017</xref>). Metadata quality evaluations&#x2019; results inform improvements in metadata education and practice which translate into higher quality, more functional metadata. Evaluations are guided by metadata quality frameworks that define quality criteria and propose measures for assessing metadata against these criteria. The most influential framework developed by Bruce and Hillman (<xref rid="R4" ref-type="bibr">2004</xref>) consists of 7 criteria. Over the years, these criteria have been adopted and adapted for analysis in many metadata studies, and informed development of other frameworks, most notably the information quality framework by Stvilia and colleagues (2007), which consists of 22 criteria. Three metadata quality criteria defined in these frameworks &#x2013; <italic>accuracy</italic>, <italic>completeness</italic>, and <italic>consistency</italic> &#x2013; are used in studies the most often, and according to the survey of metadata managers (<xref rid="R18" ref-type="bibr">Park &#x0026; Tosaka, 2010</xref>) are considered the most important in quality assurance.</p>
<p>Most metadata evaluations so far assessed human-created metadata because the lion&#x2019;s share of metadata that supports discovery is created this way (even when automation tools are used to assist in this job, and even despite the recent advances of AI generative tools that offer a potential of partially automating the process) and will likely remain largely human-generated. It is important to remember that ideas behind using machines to automate the complex tasks of metadata creation are not new and certainly not unique to the current AI revolution. Decades before the generative AI emergence, machine-generated metadata was widely researched, developed, tested, and implemented (as shown, for example, in overview of automatic metadata generation projects in <xref rid="R9" ref-type="bibr">Greenberg, 2005</xref>). For instance, various tools designed for automating some of the relatively small-scale (in terms of the space, measured in number of characters occupied in the metadata record) but important and intellectually challenging metadata creation tasks (e.g., generating a classification number based on the included subject headings, etc.) have been in existence for decades and used with acceptable accuracy &#x2013; by many professional metadata creators to reduce the amount of time needed for metadata record creation (e.g., <xref rid="R8" ref-type="bibr">Golub, 2021</xref>). Other popular uses of such tools were to automatically add to the metadata records certain codes (e.g., those representing languages used in the information resource and geographic areas covered by the content of the resource). These simple tools often take the form of macros that are shared in the professional community of metadata creators and are applied to represent any kind of information resource using the subscription based OCLC Connexion professional software (e.g., <xref rid="R16" ref-type="bibr">OCLC, 2023</xref>).</p>
<p>Other electronic tools designed to assist metadata creation are openly available to anyone interested and take the shape of online forms, to be completed, based on guidelines, by a human. The information entered by trained metadata creators into such form is then processed and converted into a certain encoding (e.g., JSON-LD, RDFXML, etc.). The most widely known such tool is the BIBFRAME Editor (https://bibframe.org/marva/editor/) designed to support human expert work with BIBFRAME metadata element set and guidelines. This tool has several application profiles built into it for representing monographs, notated music, serials, cartographic resources, sound recordings, moving images, rare materials, prints and photographs. Another notable openly accessible tool that is a set of online forms (for representing datasets, e-books, government documents, maps, microfilms, monographs, scores, serials, and theses and dissertations) is Metadata Maker (https://metadatamaker.library.illinois.edu/) developed by the University of Illinois at Urbana-Champaign Library based on BIBFRAME 2.0 (<xref rid="R10" ref-type="bibr">Heng &#x0026; Han, 2023</xref>; <xref rid="R15" ref-type="bibr">Michael &#x0026; Han, 2020</xref>). This tool allows creating metadata in MARC, MARCXML, MODS, and ONIX metadata schemes.</p>
<p>Due to their very nature, machine-readability, and internet indexability, online information resources such as webpages are presumably much easier to automate the metadata creation processes for and substantially reduce the amount of human effort needed. With that goal, long before the emergence of AI generative tools, the DC-dot (https://www.ukoln.ac.uk/metadata/dcdot/) automation tool was developed &#x2013; using Java Script &#x2013; and made available in the 1990s by the Innovation Support Centre at The United Kingdom Office for Library and Information Networking (UKOLN). The DC-dot tool generated metadata records encoded in either HTML or RDFXML and was widely used by metadata creators for generating an entire metadata record (albeit basic, and in need of verification and augmentation by human experts) in simple version of Dublin Core standard that includes 15 metadata elements. Some versions of this popular tool were integrated in professional workflow in OCLC Connexion &#x2013; the metadata management utility that feeds WorldCat database. However, due to <italic>&#x2018;the cessation of core funding&#x2019;</italic> in 2013, this popular tool that consistently generated reliable metadata results for webpages was discontinued. Ten years later, the AI generative tools became available, with the promise of offering not only the functionality of DC-dot but more robust automatic metadata generation use, capable of creating metadata records for various kinds of information resources and following various metadata schemes.</p>
<p>Most published studies which assess descriptive metadata quality focus on Dublin Core, MODS, or MARC (e.g., <xref rid="R12" ref-type="bibr">Jackson et al., 2008</xref>; <xref rid="R13" ref-type="bibr">Kurtz, 2010</xref>; <xref rid="R17" ref-type="bibr">Park &#x0026; Maszaros, 2009</xref>; Weagley, Gelches &#x0026; <xref rid="R18" ref-type="bibr">Park, 2010</xref>; <xref rid="R25" ref-type="bibr">Zavalin, Zavalina &#x0026; Safa, 2021</xref>; <xref rid="R24" ref-type="bibr">Zavalin, Zavalina &#x0026; Miksa, 2021</xref>). Likely due to the much younger age of BIBFRAME metadata element set and the vast complexity and length of BIBFRAME metadata records &#x2013; which, when of a reasonable quality, are at least three times longer than MARC records for the same information resource represented in XML syntax (https://id.loc.gov/tools/bibframe/comparebf-lccn/2018958785.xml, ) - no published studies have examined the quality of BIBFRAME metadata.</p>
<p>AI&#x2019;s potential uses in metadata generation have been explored by multiple authors who focused on individual specific metadata tasks such as automatic assignment of subject headings (e.g., <xref rid="R6" ref-type="bibr">Chou &#x0026; Chu, 2022</xref>; Chow, Kao, &#x0026; <xref rid="R20" ref-type="bibr">Li, 2024</xref>; <xref rid="R11" ref-type="bibr">Ganadi et al., 2023</xref>) or classification numbers (<xref rid="R3" ref-type="bibr">Bodenhamer, 2023</xref>; Martorana, et al., 2024; Zhang, Wu, &#x0026; <xref rid="R27" ref-type="bibr">Zhang, 2023</xref>). Two studies evaluating AI-generated descriptive metadata records have been published at the time of writing this paper, to the end of our knowledge. Taniguchi (<xref rid="R21" ref-type="bibr">2024</xref>) compared ChatGPT-created and human-expert-created MARC records for maps, sheet music, sound recordings, etc. Brzustowitz (2023) used in-depth comparative analysis of small samples of MARC and Dublin Core records created by human metadata experts and ChatGPT to represent books and audio recordings. As of the time of writing our paper, no studies have explored the quality of AI-generated metadata representing online resources and no study looked beyond MARC and simple Dublin Core in AI-generated analysis. Our project preliminary results of which are reported below addresses this research gap.</p>
</sec>
<sec id="sec3">
<title>Method</title>
<p>In this study, we used basic and advanced versions of Google&#x2019;s Gemini AI tool, as well as ChatGPT 4, to generate AI metadata for the same resource (an English-language educational webpage (https://informationscience.unt.edu/)) in 4 most widely used metadata schemes / element sets (in the order of simplest-to-most-complex):
<list list-type="bullet">
<list-item><p>Extended version of Dublin Core: Dublin Core Metadata Initiative (DCMI) Metadata Terms</p></list-item>
<list-item><p>Metadata Object Description Schema (MODS)</p></list-item>
<list-item><p>MARC 21 Bibliographic Format, and</p></list-item>
<list-item><p>BIBFRAME.</p></list-item>
</list></p>
<p>Google&#x2019;s two AI generative tools &#x2013; Gemini and Gemini Advanced &#x2013; were selected for our study, in addition to ChatGPT that is the most used AI-generative tool in existing research on AI usability in metadata generation. The reason for Gemini and Gemini Advanced selection was that this study&#x2019;s target information resource type is a website indexed by Google. Thus, the assumption was that these Google AI tools can create good quality metadata by harnessing the power of Google- indexed webpages.</p>
<p>We analysed the quality of the results, relying on our team&#x2019;s expert knowledge of these metadata elements sets and existing guidelines and documentation for these standards, as well as on expert-human-created metadata for the target information resource used in this experiment, with the goal of answering the following research questions:
<list list-type="bullet">
<list-item><p>How <italic>accurate</italic> is the AI-generated metadata?</p></list-item>
<list-item><p>How <italic>complete</italic> is the AI-generated metadata?</p></list-item>
<list-item><p>How does the quality of AI-generated metadata compare across the 4 metadata element sets and 3 AI generative tools?</p></list-item>
</list></p>
<p>We evaluated completeness and accuracy based on the measures detailed in the recent studies analysing quality of metadata in Dublin Core and other metadata schemes (e.g., Aljalahmah &#x0026; Zavalina, 2024; Zavalin &#x0026; Zavalina, 2023; Zavalina &#x0026; <xref rid="R4" ref-type="bibr">Burke, 2021</xref>). The measures included:
<list list-type="bullet">
<list-item><p>for <italic>completeness:</italic> the total number of metadata fields and their instances in the record, the number of missing applicable metadata fields/instances, etc.</p></list-item>
<list-item><p>for <italic>accuracy:</italic> the use of appropriate encoding, well-formedness, and validity of XML; the number of metadata fields and instances containing errors: misrepresentation of information resource in the date value, using non-authorized term from a controlled vocabulary, misusing the field that is intended for other kinds of data, etc.</p></list-item>
</list></p>
<p>The official documentation for the examined metadata standards and human-created metadata records for the target information resource served as a source of information on the number of applicable metadata fields, etc. that were used in analysis. The prompts used in all 3 generative AI tools followed this pattern: &#x2018;<italic>Generate a metadata record for this webpage: https://informationscience.unt.edu/. Follow the [NAME &#x0026; VERSION] standard: [LINK TO STANDARD]&#x2019;.</italic> For all 4 metadata schemes, the official documentation is freely available online in machine-readable HTML or XML format.</p>
</sec>
<sec id="sec4">
<title>Preliminary Results and Discussion</title>
<p><xref ref-type="fig" rid="F1">Figure 1</xref> presents the prompts that were used in this experiment and summarizes the results of the quality assessment of the outputs generated by Gemini, Gemini Advanced, and ChatGPT. Overall, metadata records generated by ChatGPT were more detailed. Those generated by Gemini Advanced for 3 metadata schemes &#x2013; Dublin Core, MODS, and MARC &#x2013; tended to contain more details than the ones generated by Gemini. This is in line with our assumptions based on Google&#x2019;s positioning Gemini Advanced as a more powerful AI tool.</p>
<p>Dublin Core metadata generated by ChatGPT and Gemini Advanced, while incomplete, was comparable in the number of metadata fields included to OAIster-harvested Dublin Core records that previous research found to have 8 metadata fields in most cases (<xref rid="R12" ref-type="bibr">Jackson et al., 2008</xref>). However, it is important to note that OAIster aggregation metadata follows the simple version of Dublin Core, while this experiment tasked AI generative tools with creating metadata that follows the extended version of Dublin Core standard (DCMI Metadata Terms) which has over 40 metadata elements as opposed to 15. For more robust MODS and MARC metadata schemes, completeness was moderate (with a few important omissions) in ChatGPT-generated record, and very low in both versions of the Gemini tool, although somewhat more complete in Gemini Advanced (<xref ref-type="fig" rid="F1">Figure 1</xref>).</p>
<p>While ChatGPT and Gemini Advanced generated BIBFRAME records (albeit incomplete, and with basic JSON encoding as opposed to JSON-LD or RDF serialization that support Linked Data functionality), Gemini returned this statement,</p>
<disp-quote>
<p><italic>I don&#x2019;t have a tool that can directly generate bibframe metadata for webpages, but I can search on the web using google tools. Here&#x2019;s what I found on the web: I couldn&#x2019;t find the bibframe metadata for this webpage, but there are some resources that discuss bibframe [&#x2026;].</italic></p>
</disp-quote>
<p>The accuracy was low in all 3 generative-AI tools&#x2019; outputs for Dublin Core, MODS, and MARC. The proportion of metadata fields with mistakes was similarly high across tools for the same metadata scheme. Also, all 3 tools failed to generate the DCMI Metadata Terms records and generated records that follow the simpler version of this same metadata scheme: DCMES 1.1. However, unexpectedly, for Dublin Core and MODS, the simplest of the 3 tools, Gemini was more successful than Gemini Advanced in generating the metadata record properly encoded (in XML) and structured (with both top-level elements and sub elements) for MODS. ChatGPT was successful in structuring the metadata record in this hierarchical metadata scheme. ChatGPT-generated metadata was encoded in machine-readable syntax for all 4 metadata schemes (XML for Dublin Core and MODS, MARC for MARC Bibliographic Format, and JSON for BIBFRAME). However, XML encoding was missing key components that negatively affected its functionality in information discovery and wrong version of JSON syntax was used (<xref ref-type="fig" rid="F1">Figure 1</xref>). </p>
<fig id="F1">
<label>Figure 1.</label>
<caption><p>Metadata quality evaluation results summary</p></caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="images\c62-fig1.jpg"><alt-text>none</alt-text></graphic>
</fig>
</sec>
<sec id="sec5">
<title>Conclusions and next steps</title>
<p>The overall quality of AI-generated metadata in this experiment does NOT meet the basic expectations for functional metadata. Low-quality metadata &#x2013; such as observed for AI-generated metadata in this study - fails to support Find, Identify, Select, Obtain, &#x0026; Explore user tasks defined in IFLA LRM (<xref rid="R19" ref-type="bibr">Riva, Lebeouf, &#x0026; &#x017D;umer, 2017</xref>). At this point in the development of AI tools, relying solely on AI-generated metadata creation would result in significantly compromised information access. However, AI-generated metadata could be used as a starting point in the human metadata creation for Dublin Core, MARC, and MODS metadata standards, with the need for thorough checking and editing by human metadata creators. AI-generated BIBFRAME metadata with Gemini and Gemini Advanced Google AI tools is currently non-existent, while ChatGPT-produced BIBFRAME metadata although very incomplete and not supporting Linked Data functionality, could be used by expert metadata professionals as a starting point, with significant manual revisions.</p>
<p>This exploratory research focused on one type of information resource previously excluded from experimenting with AI metadata generation: online educational website. Future analysis will compare accuracy and completeness of metadata generated by these same AI tools for different types of information resources. Another important direction for further research is training the AI-generative tools by refining the metadata generation prompts in different ways and by providing the tool (where applicable) with feedback on the quality of generated metadata.</p>
<p>Findings of studies such as this one, as well as experimenting with AI-generated metadata itself, need to be integrated in metadata education for information professionals. For example, instructors of cataloguing and digital library metadata courses can demonstrate to their students how these tools work for metadata generation, use AI-generated metadata in teaching students the topics of metadata quality evaluation, ask students to improve AI-generated records as part of their metadata-creation assignments.</p>
<p>The need for research into the quality of AI-generated descriptive metadata is time-sensitive because <italic>&#x2018;AI has been increasingly shaping the library management landscape in recent</italic> years&#x2019; (<xref rid="R2" ref-type="bibr">Bisht et al., 2023</xref>). While we agree that <italic>&#x2018;integration of AI technology has the potential to greatly enhance the efficiency, accuracy, and user experience in library cataloguing&#x2019;</italic> (<xref rid="R2" ref-type="bibr">Bisht et al., 2023</xref>) by assisting the human metadata creators, this potential needs to be thoroughly and objectively evaluated before taking any related management steps. It is important to conduct such examinations to collect the data that should be used in evidence-based decisions regarding management of information organizations (including libraries) and such crucial functions as metadata creation.</p>
</sec>
</body>
<back>
<ref-list>
<title>References</title>
<ref id="R1"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Aljalahmah</surname><given-names>S.H.</given-names></name><name><surname>Zavalina</surname><given-names>O. L.</given-names></name></person-group><year>2024</year><article-title>Student-created Dublin Core metadata representing Arabic language eBooks: Comparison of individual and group work outcomes</article-title><source>Journal of Education for Library and Information Science (JELIS)</source><volume>65</volume><issue>3</issue><fpage>325</fpage><lpage>344</lpage><ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3138/jelis-2023-0016">https://doi.org/10.3138/jelis-2023-0016</ext-link></element-citation></ref>
<ref id="R2"><element-citation publication-type="book"><person-group person-group-type="author"><name><surname>Bisht</surname><given-names>S.</given-names></name><name><surname>Nutiyal</surname><given-names>A.P.</given-names></name><name><surname>Sharma</surname><given-names>S.</given-names></name><name><surname>Sai</surname><given-names>M.</given-names></name><name><surname>Bathla</surname><given-names>N.</given-names></name><name><surname>Singh</surname><given-names>P.</given-names></name></person-group><year>2023</year><chapter-title>The role of Artificial Intelligence in shaping library management and its utilization</chapter-title><source>2023 International Conference on Disruptive Technologies (ICDT)</source><publisher-loc>Greater Noida, India</publisher-loc><fpage>467</fpage><lpage>472</lpage><ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1109/ICDT57929.2023.10150520.">https://doi.org/10.1109/ICDT57929.2023.10150520.</ext-link></element-citation></ref>
<ref id="R3"><element-citation publication-type="other"><person-group person-group-type="author"><name><surname>Bodenhamer</surname><given-names>J.</given-names></name></person-group><year>2023</year><source>The reliability and usability of ChatGPT for library metadata</source><ext-link ext-link-type="uri" xlink:href="https://hdl.handle.net/11244/339626 (accessed 12 August 2024)">https://hdl.handle.net/11244/339626 (accessed 12 August 2024)</ext-link></element-citation></ref>
<ref id="R4"><element-citation publication-type="book"><person-group person-group-type="author"><name><surname>Bruce</surname><given-names>T.R.</given-names></name><name><surname>Hillman</surname><given-names>D.I.</given-names></name></person-group><year>2004</year><chapter-title>The continuum of metadata quality: defining, expressing, exploiting</chapter-title><person-group person-group-type="editor"><name><surname>Hillman</surname><given-names>D.</given-names></name></person-group><person-group person-group-type="editor"><name><surname>Westbrook</surname><given-names>L.</given-names></name></person-group><source>Metadata in Practice</source><fpage>238</fpage><lpage>256</lpage><publisher-name>American Library Association</publisher-name><publisher-loc>Chicago</publisher-loc></element-citation></ref>
<ref id="R5"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Brzustowicz</surname><given-names>R.</given-names></name></person-group><year>2023</year><article-title>From ChatGPT to CatGPT: The implications of Artificial Intelligence on library cataloguing</article-title><source>Information Technology and Libraries</source><volume>42</volume><issue>3</issue><ext-link ext-link-type="uri" xlink:href="https://doi.org/10.5860/ital.v42i3.16295">https://doi.org/10.5860/ital.v42i3.16295</ext-link></element-citation></ref>
<ref id="R6"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Chou</surname><given-names>C.</given-names></name><name><surname>Chu</surname><given-names>T.</given-names></name></person-group><year>2022</year><article-title>An analysis of BERT (NLP) for assisted subject indexing for project Gutenberg</article-title><source>Cataloging &#x0026; Classification Quarterly</source><volume>68</volume><issue>8</issue><fpage>807</fpage><lpage>835</lpage><ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1080/01639374.2022.2138666">https://doi.org/10.1080/01639374.2022.2138666</ext-link></element-citation></ref>
<ref id="R7"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Chow</surname><given-names>E. H. C.</given-names></name><name><surname>Kao</surname><given-names>T. J.</given-names></name><name><surname>Li</surname><given-names>X.</given-names></name></person-group><year>2024</year><article-title>An Experiment with the Use of ChatGPT for LCSH Subject Assignment on Electronic Theses and Dissertations</article-title><source>Cataloging &#x0026; Classification Quarterly</source><volume>62</volume><issue>5</issue><fpage>574</fpage><lpage>588</lpage><ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1080/01639374.2024.2394516">https://doi.org/10.1080/01639374.2024.2394516</ext-link></element-citation></ref>
<ref id="R8"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Golub</surname><given-names>K.</given-names></name></person-group><year>2021</year><article-title>Automated subject indexing: An overview</article-title><source>Cataloging &#x0026; Classification Quarterly</source><volume>59</volume><issue>8</issue><fpage>702</fpage><lpage>719</lpage><ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1080/01639374.2021.2012311">https://doi.org/10.1080/01639374.2021.2012311</ext-link></element-citation></ref>
<ref id="R9"><element-citation publication-type="other"><person-group person-group-type="author"><name><surname>Greenberg</surname><given-names>J.</given-names></name></person-group><year>2005</year><article-title>Metadata generation: Processes, people, and tools</article-title><source>Bulletin of the American Society for Information Science and Technology</source><ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1002/bult.269">https://doi.org/10.1002/bult.269</ext-link></element-citation></ref>
<ref id="R10"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Heng</surname><given-names>G.</given-names></name><name><surname>Han</surname><given-names>M.J.</given-names></name></person-group><year>2023</year><article-title>Revamping Metadata Maker for &#x2018;Linked Data Editor&#x2019;: Thinking Out Loud</article-title><source>Code4Lib Journal</source><volume>55</volume><ext-link ext-link-type="uri" xlink:href="https://journal.code4lib.org/articles/16925">https://journal.code4lib.org/articles/16925</ext-link></element-citation></ref>
<ref id="R11"><element-citation publication-type="book"><person-group person-group-type="author"><name><surname>Ganadi</surname><given-names>A.E.</given-names></name><name><surname>Vigliermo</surname><given-names>R.A.</given-names></name><name><surname>Sala</surname><given-names>L.</given-names></name><name><surname>Vanzini</surname><given-names>M.</given-names></name><name><surname>Ruozzi</surname><given-names>F.</given-names></name><name><surname>Bergamaschi</surname><given-names>S.</given-names></name></person-group><year>2023</year><chapter-title>Bridging Islamic knowledge and AI: inquiring ChatGPT on possible categorizations for an Islamic digital library</chapter-title><source>Proceedings of the 2nd Workshop on Artificial Intelligence for Cultural Heritage (IAI4CH 2023) co-located with the 22nd International Conference of the Italian Association for Artificial Intelligence (AIxIA 2023)</source><comment>November 6, 2023</comment><publisher-loc>Roma, Italy</publisher-loc><ext-link ext-link-type="uri" xlink:href="https://ceurws.org/Vol-3536/03_paper.pdf (accessed 12 August 2024)">https://ceurws.org/Vol-3536/03_paper.pdf (accessed 12 August 2024)</ext-link></element-citation></ref>
<ref id="R12"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Jackson</surname><given-names>A.S.</given-names></name><name><surname>Han</surname><given-names>M.J.</given-names></name><name><surname>Groetsch</surname><given-names>K.</given-names></name><name><surname>Mustafoff</surname><given-names>M.</given-names></name><name><surname>Cole</surname><given-names>T.W.</given-names></name></person-group><year>2008</year><article-title>Dublin Core metadata harvested through OAI-PMH</article-title><source>Journal of Library Metadata</source><volume>8</volume><issue>1</issue><fpage>5</fpage><lpage>21</lpage><ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1300/J517v08n01_02">https://doi.org/10.1300/J517v08n01_02</ext-link></element-citation></ref>
<ref id="R13"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Kurtz</surname><given-names>M.</given-names></name></person-group><year>2010</year><article-title>Dublin Core, DSpace, and a brief analysis of three university repositories</article-title><source>Information Technology and Libraries</source><volume>29</volume><issue>1</issue><fpage>40</fpage><lpage>46</lpage><ext-link ext-link-type="uri" xlink:href="https://doi.org/10.6017/ital.v29i1.3157">https://doi.org/10.6017/ital.v29i1.3157</ext-link></element-citation></ref>
<ref id="R14"><element-citation publication-type="other"><person-group person-group-type="author"><name><surname>Martorana</surname><given-names>M.</given-names></name><name><surname>Kuhn</surname><given-names>T.</given-names></name><name><surname>Stork</surname><given-names>L.</given-names></name><name><surname>Ossenbruggen</surname><given-names>J.V.</given-names></name></person-group><year>2024</year><article-title>Zero-Shot Topic Classification of Column Headers: Leveraging LLMs for Metadata Enrichment</article-title><source>International Conference on Semantic Systems</source><comment>arXiv-Computer Science-Databases</comment><ext-link ext-link-type="uri" xlink:href="https://doi.org/10.48550/arXiv.2403.00884">https://doi.org/10.48550/arXiv.2403.00884</ext-link></element-citation></ref>
<ref id="R15"><element-citation publication-type="other"><person-group person-group-type="author"><name><surname>Michael</surname><given-names>B.</given-names></name><name><surname>Han</surname><given-names>M.J.</given-names></name></person-group><year>2020</year><article-title>Assessing BIBFRAME 2.0: Exploratory Implementation in Metadata Maker</article-title><source>International Conference on Dublin Core and Metadata Applications</source><comment>2020</comment><ext-link ext-link-type="uri" xlink:href="https://doi.org/10.23106/dcmi.952141614">https://doi.org/10.23106/dcmi.952141614</ext-link></element-citation></ref>
<ref id="R16"><element-citation publication-type="other"><person-group person-group-type="author"><collab>OCLC</collab></person-group><year>2023</year><article-title>OCLC-supplied macros</article-title><ext-link ext-link-type="uri" xlink:href="https://help.oclc.org/Metadata_Services/Connexion/Connexion_client/Connexion_client_basics/Use_macros/Get_started/50OCLC_supplied_macros">https://help.oclc.org/Metadata_Services/Connexion/Connexion_client/Connexion_client_basics/Use_macros/Get_started/50OCLC_supplied_macros</ext-link></element-citation></ref>
<ref id="R17"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Park</surname><given-names>J.R.</given-names></name><name><surname>Maszaros</surname><given-names>S.</given-names></name></person-group><year>2009</year><article-title>Metadata Object Description Schema (MODS) in digital repositories: An exploratory study of metadata use and quality</article-title><source>Knowledge Organization</source><volume>36</volume><issue>1</issue><fpage>46</fpage><lpage>59</lpage><ext-link ext-link-type="uri" xlink:href="https://doi.org/10.5771/0943-7444-2009-1-46">https://doi.org/10.5771/0943-7444-2009-1-46</ext-link></element-citation></ref>
<ref id="R18"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Park</surname><given-names>J.R.</given-names></name><name><surname>Tosaka</surname><given-names>Y.</given-names></name></person-group><year>2010</year><article-title>Metadata quality control in digital repositories and collections: criteria, semantics, and mechanisms</article-title><source>Cataloging &#x0026; Classification Quarterly</source><volume>48</volume><issue>8</issue><fpage>696</fpage><lpage>715</lpage><ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1080/01639374.2010.508711">https://doi.org/10.1080/01639374.2010.508711</ext-link></element-citation></ref>
<ref id="R19"><element-citation publication-type="other"><person-group person-group-type="author"><name><surname>Riva</surname><given-names>P.</given-names></name><name><surname>Lebeouf</surname><given-names>P.</given-names></name><name><surname>&#x017D;umer</surname><given-names>M.</given-names></name></person-group><year>2017</year><article-title>IFLA Library Reference Model: A Conceptual Model for Bibliographic Information</article-title><comment>Retrieved August 12, 2024</comment><ext-link ext-link-type="uri" xlink:href="https://www.ifla.org/wp- content/uploads/2019/05/assets/cataloguing/frbr-lrm/ifla-lrm-august-2017_rev201712.pdf">https://www.ifla.org/wp- content/uploads/2019/05/assets/cataloguing/frbr-lrm/ifla-lrm-august-2017_rev201712.pdf</ext-link></element-citation></ref>
<ref id="R20"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Stvilia</surname><given-names>B.</given-names></name><name><surname>Gasser</surname><given-names>L.</given-names></name><name><surname>Twidale</surname><given-names>M. B.</given-names></name><name><surname>Smith</surname><given-names>L. C.</given-names></name></person-group><year>2007</year><article-title>A framework for information quality assessment</article-title><source>Journal of the American Society of Information Science</source><volume>58</volume><fpage>1720</fpage><lpage>1733</lpage><ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1002/asi.20652">https://doi.org/10.1002/asi.20652</ext-link></element-citation></ref>
<ref id="R21"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Taniguchi</surname><given-names>S.</given-names></name></person-group><year>2024</year><article-title>Creating and evaluating MARC 21 bibliographic records using ChatGPT</article-title><source>Cataloging &#x0026; Classification Quarterly</source><volume>62</volume><issue>5</issue><fpage>527</fpage><lpage>546</lpage><ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1080/01639374.2024.2394513">https://doi.org/10.1080/01639374.2024.2394513</ext-link></element-citation></ref>
<ref id="R22"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Weagley</surname><given-names>J.</given-names></name><name><surname>Gelches</surname><given-names>E.</given-names></name><name><surname>Park</surname><given-names>J.</given-names></name></person-group><year>2010</year><article-title>Interoperability and metadata quality in digital video repositories: a study of Dublin Core</article-title><source>Journal of Library Metadata</source><volume>10</volume><issue>1</issue><fpage>37</fpage><lpage>57</lpage><ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1080/19386380903546984">https://doi.org/10.1080/19386380903546984</ext-link></element-citation></ref>
<ref id="R23"><element-citation publication-type="other"><person-group person-group-type="author"><name><surname>Zavalin</surname><given-names>V.I.</given-names></name><name><surname>Zavalina</surname><given-names>O. L.</given-names></name></person-group><year>2023</year><article-title>Exploration of accuracy, completeness, and consistency in metadata for physical objects in museum collections</article-title><source>Information for a Better World: Normality, Virtuality, Physicality, Inclusivity: 18th International Conference</source><comment>iConference 2023, Proceedings</comment><fpage>83</fpage><lpage>90</lpage><ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1007/978-3-031-28032-0_7">https://doi.org/10.1007/978-3-031-28032-0_7</ext-link></element-citation></ref>
<ref id="R24"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Zavalin</surname><given-names>V.</given-names></name><name><surname>Zavalina</surname><given-names>O.L.</given-names></name><name><surname>Miksa</surname><given-names>S.D.</given-names></name></person-group><year>2021</year><article-title>Exploration of subject representation and support of Linked Data in recently created library metadata: Examination of most widely held WorldCat bibliographic records</article-title><source>Library Resources and Technical Services</source><volume>65</volume><issue>4</issue><fpage>154</fpage><lpage>165</lpage><ext-link ext-link-type="uri" xlink:href="https://doi.org/10.5860/lrts.65n4.1544">https://doi.org/10.5860/lrts.65n4.1544</ext-link></element-citation></ref>
<ref id="R25"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Zavalin</surname><given-names>V.</given-names></name><name><surname>Zavalina</surname><given-names>O.L.</given-names></name><name><surname>Safa</surname><given-names>R.</given-names></name></person-group><year>2021</year><article-title>Patterns of subject metadata change in MARC21 bibliographic records representing video recordings</article-title><source>Proceedings of the Association for Information Science and Technology</source><volume>58</volume><issue>1</issue><ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1002/pra2.494">https://doi.org/10.1002/pra2.494</ext-link></element-citation></ref>
<ref id="R26"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Zavalina</surname><given-names>O.L.</given-names></name><name><surname>Burke</surname><given-names>M.</given-names></name></person-group><year>2021</year><article-title>Assessing skill-building in metadata instruction: Quality evaluation of Dublin Core metadata records created by graduate students</article-title><source>Journal of Education for Library and Information Science</source><volume>62</volume><issue>4</issue><fpage>423</fpage><lpage>442</lpage><ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3138/jelis.62-4-2020-0083">https://doi.org/10.3138/jelis.62-4-2020-0083</ext-link></element-citation></ref>
<ref id="R27"><element-citation publication-type="other"><person-group person-group-type="author"><name><surname>Zhang</surname><given-names>S.</given-names></name><name><surname>Wu</surname><given-names>M.</given-names></name><name><surname>Zhang</surname><given-names>X.</given-names></name></person-group><year>2023</year><article-title>Utilising a large language model to annotate subject metadata: a case study in an Australian national research data catalogue&#x201D;</article-title><source>arXiv-CS-Computation and Language</source><ext-link ext-link-type="uri" xlink:href="https://doi.org/10.48550/arXiv.2310.11318">https://doi.org/10.48550/arXiv.2310.11318</ext-link></element-citation></ref>
</ref-list>
</back>
</article>