<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.0 20120330//EN" "http://jats.nlm.nih.gov/publishing/1.0/JATS-journalpublishing1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:mml="http://www.w3.org/1998/Math/MathML" article-type="research-article" xml:lang="en">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">IR</journal-id>
<journal-title-group>
<journal-title>Information Research</journal-title>
</journal-title-group>
<issn pub-type="epub">1368-1613</issn>
<publisher>
<publisher-name>University of Bor&#x00E5;s</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">ir30iConf47335</article-id>
<article-id pub-id-type="doi">10.47989/ir30iConf47335</article-id>
<article-categories>
<subj-group xml:lang="en">
<subject>Research article</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>Innovative practice of archival data development workflow in the AGI era: a case study of scientist archives project</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author"><name><surname>Fu</surname><given-names>Yaming</given-names></name>
<xref ref-type="aff" rid="aff0001"/></contrib>
<contrib contrib-type="author"><name><surname>Song</surname><given-names>Jie</given-names></name>
<xref ref-type="aff" rid="aff0002"/></contrib>
<contrib contrib-type="author"><name><surname>Zhang</surname><given-names>Xinran</given-names></name>
<xref ref-type="aff" rid="aff0003"/></contrib>
<contrib contrib-type="author"><name><surname>Bi</surname><given-names>Jingyun</given-names></name>
<xref ref-type="aff" rid="aff0004"/></contrib>
<aff id="aff0001"><bold>Dr. Yaming Fu</bold> is a Lecturer at School of Cultural Heritage and Information Management, Shanghai University, Shanghai, China. Her research interests include digital humanities, knowledge organization and archives informatization. E-mail: <email xlink:href="ymingfu@126.com">ymingfu@126.com</email></aff>
<aff id="aff0002"><bold>Jie Song</bold> is the Founder and CEO of Intellijourney Co., Ltd., Shanghai, China. His research interests include IIIF and artificial intelligence. E-mail: <email xlink:href="joe.song@intellijourney.com">joe.song@intellijourney.com</email></aff>
<aff id="aff0003"><bold>Xinran Zhang</bold> is an undergraduate at School of Cultural Heritage and Information Management, Shanghai University, Shanghai, China. Her research interests include basic theory of archival science, digital humanities. E-mail: <email xlink:href="1654770712@qq.com">1654770712@qq.com</email></aff>
<aff id="aff0004"><bold>Jingyun Bi</bold> is an undergraduate at Department of Electrical and Electronic Engineering, The Hong Kong Polytechnic University, Hong Kong SAR, China. His research interests include application of generative AI in social data, and radio frequency surrogate models for automatic design of active and passive circuits. E-mail: <email xlink:href="2537503782@qq.com">2537503782@qq.com</email></aff>
</contrib-group>
<pub-date pub-type="epub"><day>06</day><month>05</month><year>2025</year></pub-date>
<pub-date pub-type="collection"><year>2025</year></pub-date>
<volume>30</volume>
<issue>i</issue>
<fpage>349</fpage>
<lpage>360</lpage>
<permissions>
<copyright-year>2025</copyright-year>
<copyright-holder>&#x00A9; 2025 The Author(s).</copyright-holder>
<license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by-nc/4.0/">
<license-p>This is an Open Access article distributed under the terms of the Creative Commons Attribution-NonCommercial 4.0 International License (<ext-link ext-link-type="uri" xlink:href="http://creativecommons.org/licenses/by-nc/4.0/">http://creativecommons.org/licenses/by-nc/4.0/</ext-link>), permitting all non-commercial use, distribution, and reproduction in any medium, provided the original work is properly cited.</license-p>
</license>
</permissions>
<abstract xml:lang="en">
<title>Abstract</title>
<p><bold>Introduction.</bold> The advent of large language models (LLMs) presents a transformative opportunity for the field of archival science, offering advanced capabilities in intelligent information processing, semantic search, and more. These innovations address critical challenges posed by the exponential growth in archival materials and the increasing demand for efficient data analysis. Traditional archival workflows, often rely on manual description and optical character recognition (OCR), struggle with the complexities of unstructured digital data, especially in the context of digitized historical archives and manuscripts.</p>
<p><bold>Method.</bold> This paper explores a novel archival workflow through the case study of the scientist archives project, integrating human-machine collaboration and leveraging technologies such as open-sourced archival databases, IIIF-supported OCR environments, and advanced LLMs, including classic retrieval-augmented generation (Classic RAG) and graph-based retrieval-augmented generation (GraphRAG).</p>
<p><bold>Analysis.</bold> A Scientist Archives project is analysed, which utilises AGI era technologies to mine and manage the archives.</p>
<p><bold>Results.</bold> By embracing these technologies, the proposed approach seeks to revolutionize archival management, enhancing both efficiency and the depth of content revelation in the AGI era.</p>
<p><bold>Conclusion(s).</bold> This study contributes to the ongoing discourse on the intelligent transformation of archival practices, providing a roadmap for future archival data mining and management.</p>
</abstract>
</article-meta>
</front>
<body>
<sec id="sec1">
<title>Introduction</title>
<p>Since 2018, the archival community has actively discussed the impact of artificial intelligence and digital technologies, and research on the digitalization and intelligent transformation of archival management has attracted much attention (Rolan, 2021). By late 2022, the release of ChatGPT marked a significant milestone in the development of large language models (LLMs), fueling growing discussions and heightened expectations for the era of artificial general intelligence (AGI) (<xref rid="R2" ref-type="bibr">Emmert-Streib, 2024</xref>). As a core and controversial concept in the field of computer research, AGI refers to the AI systems that exhibit comparable intelligence to human intelligence in a wide range of tasks and situations, and such systems usually have general task-solving capabilities and continuous self-learning capabilities and <italic>&#x2018;exhibit brain-like intelligence characteristics in cognition and decision-making&#x2019;</italic> (<xref rid="R3" ref-type="bibr">Morris, 2023</xref>). LLM shows versatility in natural language processing, provides a window for the public to approach AGI, and serves as an important milestone on the road to AGI. As a generative natural language processing model for large-scale corpus training, &#x2018;<italic>LLM has the functions of automatic text generation, intelligent information processing, semantic search and recognition, intelligent image generation, etc.&#x2019;</italic> (<xref rid="R4" ref-type="bibr">Naveed, 2023</xref>), which provides a promising research direction for archival data mining and archival management, and triggers deep thinking on a more <italic>&#x2018;intelligent&#x2019;</italic> transformation of archival work.</p>
<p>The surge in archival materials, the growing demand for efficient methods of analysing archival data and in-depth archival mining have brought unprecedented challenges to traditional archival workflow. Manual description processes are time-consuming, labour-intensive, and prone to errors, particularly when dealing with large volumes of archival materials. Optical character recognition (OCR) software is widely used to extract text from digital images, facilitating the creation of searchable text files (<xref rid="R5" ref-type="bibr">Bukhari, 2017</xref>). However, digital archival data is often unstructured, complicating retrieval and utilization. This is particularly true for digitized historical archives and manuscripts, which are difficult to recognize, analyse, and describe. Moreover, traditional workflows fall short in granular content revelation, failing to meet users&#x2019; diverse information needs in the new era.</p>
<p>This project uses the archival materials of Chen Taosheng (In Chinese: &#x9648;&#x9A0A;&#x58F0;), the founder and pioneer of modern industrial microbiology in China, as a case study to explore a novel archival workflow that integrates human-machine collaborative models in managing scientist archives. The proposed workflow leverages advanced technologies, including an open-sourced archival database, the international image interoperability framework (IIIF) supported work environment, and LLMs, particularly classic retrieval-augmented generation (Classic RAG) and graph-based retrieval-augmented generation (GraphRAG). The project aims to revolutionize archival workflows, enhancing the overall efficiency and quality of archival management in the era of AGI.</p>
</sec>
<sec id="sec2">
<title>Problem statement</title>
<p>Traditional archival workflows are insufficient to meet the demands of modern archives management. This is not only due to the unreliable manual processing and cataloguing but also the fragmented use of independent tools for different archival tasks. These tools include but are not limited to &#x2018;<italic>metadata management tool, OCR tool, ontology design tool and knowledge graph tool&#x2019;</italic> (<xref rid="R6" ref-type="bibr">Koch, 2019</xref>; <xref rid="R7" ref-type="bibr">Philips, 2020</xref>). In addition, traditional workflows lack the granularity needed to reveal the full depth of archival content, but in the meantime, the authenticity and security of the archival data remain, which is key to archival management. Therefore, the research problems of this project are:</p>
<p>First, design and implement an IIIF-based OCR environment to support archival transcription, recognition, comparison, and description. IIIF, as <italic>&#x2018;a set of open standards for delivering high-quality, attributed digital objects online at scale&#x2019;</italic> (Link:  <ext-link ext-link-type="uri" xlink:href="https://iiif.io">https://iiif.io</ext-link>), could provide standardized image processing protocols, enhancing image-formatted archival data management, improving the efficiency, and ensuring accuracy of archival data.</p>
<p>Second, formulate a workflow for managing scientists&#x2019; archives, organize scientists&#x2019; archives, photos, literary works, monographs, patents, and relevant documents, index them in the open-sourced archival workspace ArchivesSpace, and integrate them with LLM tools to facilitate efficient access and utilization.</p>
<p>Third, build Classic RAG and GraphRAG to deeply mine and analyse archival data, improve the interpretation ability of archival data with scanned archival images, and design the &#x2018;chat&#x2019; function with AIGC, forming a new paradigm of archival utilization.</p>
<p>With the above three research goals, the project aims to develop an AGI-Era Archival workflow solution, an efficient and reliable framework to promote the intelligent processing of archival description, analysis, development, and utilization.</p>
</sec>
<sec id="sec3">
<title>Methodology</title>
<sec id="sec3_1">
<title>Research process</title>
<p>This work originated in early 2023 with the call to collect, organize and preserve the archival materials of scientists, especially the ones that were kept by individuals and the ones that were in bad condition. The aim was to collect, digitise, analyse, and make use of the scattered and deteriorating archival materials related to the lives and work of scientists, as well as to excavate the values hidden in the archives with available technologies. In October 2023, a project member interviewed a close friend of Chen Taosheng, a founder and pioneer of modern industrial microbiology in China and a key figure in establishing the department of bioengineering at the predecessor of Shanghai University. During this interview, the project member obtained valuable family archival materials. Subsequently, project members visited the Shanghai municipal archives, and the Shanghai University archives, acquiring more materials related to Chen Taosheng. With the number and types of archival materials growing, an urgent need has emerged to design an efficient workflow to manage, describe, interpret, and analyse those archival materials safely without damaging the original files&#x2019; quality and meaning. Thus, the project takes Chen Taosheng&#x2019;s archival materials as a case study to explore the development of an archival workflow for archival data in the AGI era.</p>
<p>An innovative archival development workflow in the AGI era should enhance efficiency while maintaining data authenticity, accuracy, and security. This project, an AI-integrated working environment based on the IIIF framework and embedded with OCR tools, was designed to facilitate processing image-based archival data. It also utilized LLM technology to assist in analysing massive archival raw images and their corresponding texts. Classic RAG and GraphRAG were built to realize the <italic>&#x2018;chat&#x2019;</italic> function with the scientist and form scientists&#x2019; knowledge graphs, scientists&#x2019; footprint maps, and scientists&#x2019; chronology with the help of artificial intelligence, thereby preserving and extending their research contributions. Through this integrated platform, tasks such as archival recognition, classification, transcription, description, comparison, and analysis were completed in a unified environment, significantly improving the efficiency of archival development while ensuring data accuracy.</p>
</sec>
<sec id="sec3_2">
<title>Data and ethical statement</title>
<p>All archival materials utilized in this project were obtained with permission from four sources, including:
<list list-type="order">
<list-item><p>the close friend of Chen Taosheng: The materials are mainly published magazines and newspapers related to the scientist, essays published or mentioned by Chen Taosheng himself, life photos of Chen Taosheng, and his unpublished poetry collection <italic>&#x2018;Guanwei collection&#x2019;</italic> (in Chinese: &#x89C2;&#x5FAE;&#x96C6;).</p></list-item>
<list-item><p>Shanghai municipal archives: the materials contain his views on scientific research, his career and work proposal and its reply, his work introduction letters, and materials for selecting advanced individuals in the science field.</p></list-item>
<list-item><p>Archives of Shanghai University: the materials include the forms filled in by him during his work at the predecessor of Shanghai University, his thoughts and opinions, the relevant school records, and his relevant letter of appointment.</p></list-item>
<list-item><p>Open data from Shanghai Library: records of other celebrities and scientists related to Chen Taosheng; records of the institutions and places that relate to him.</p></list-item>
</list></p>
<p>The types of archival data include original archival files, old photographs, manuscripts, newspaper clippings, journal articles clippings. These materials were acquired with explicit agreements permitting their digitization and use only for academic purposes.</p>
<p>The subsequent data processing and experiments were conducted within the university laboratory under the oversight and management of the university&#x2019;s office of scientific research management, safeguarding the integrity and authenticity of the archival materials throughout the project.</p>
</sec>
<sec id="sec3_3">
<title>Research process and methods</title>
<p>In the scientist archival data development workflow project, the main research path is <italic>&#x2018;physical form of archives&#x2014;digital archival images&#x2014;textual archival data&#x2014;archival knowledge&#x2019;</italic> (see <xref ref-type="fig" rid="F1">Fig. 1</xref>).</p>
<fig id="F1">
<label>Figure 1.</label>
<caption><p>Research process and methods</p></caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="images\c29-fig1.jpg"><alt-text>none</alt-text></graphic>
</fig>
<p>The workflow began with digitizing physical archives and records, utilizing high-resolution scanners to convert archival documents, photographs, and other archival materials into digital formats. These digitalized files were then organized and catalogued using the open-sourced archival management system, ArchivesSpace, where structured metadata and standard archival descriptions to enhance data retrieval and operability (see <xref ref-type="fig" rid="F2">Figure 2</xref>). During this process, raw images were integrated through IIIF objects and recorded as digital object instances corresponding to physical items.</p>
<p><italic>&#x2018;ArchivesSpace is an open-source, browser-based archival information management software application&#x2019;</italic> (Link: <ext-link ext-link-type="uri" xlink:href="https://archivesspace.org">https://archivesspace.org</ext-link>). The underlying descriptive standard it uses is DACS (Describing Archives: A Content Standard). ArchivesSpace allowed the users to add, arrange, describe, evaluate materials, manage locations, control permissions, and link subject tags when storing archive metadata. DACS is a foundational statement of content standards and principles for archival organization and description within the United States archival community. It includes descriptions of archival material and authoritative records representing the people and organizations that created them (<xref rid="R10" ref-type="bibr">SAA, 2022</xref>). DACS contains 25 archival elements, many encoded with multiple EAD tags.</p>
<fig id="F2">
<label>Figure 2.</label>
<caption><p>ArchivesSpace archival management system</p></caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="images\c29-fig2.jpg"><alt-text>none</alt-text></graphic>
</fig>
<p>Then IIIF was integrated with the OCR text editing plugin in the Office Word environment (see <xref ref-type="fig" rid="F3">Figure 3</xref>) to realize a commonly used archival work environment and the seamless linking of raw images with their corresponding OCR texts, ensuring comprehensive documentation and accessibility.</p>
<fig id="F3">
<label>Figure 3.</label>
<caption><p>IIIF integrated word environment with OCR plugin</p></caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="images\c29-fig3.jpg"><alt-text>none</alt-text></graphic>
</fig>
<p>Archival images (unstructured data) are transformed into structured and semi-structured data through the IIIF-based OCR tool. Structured data refers to tabular data that can be directly stored in a structured database, while semi-structured data refers to data in a loose key value format such as JSONL format. Traditional manual description tasks are thus converted into data mapping tasks that can be automatically completed by LLMs. Once a high-quality electronic archival database is constructed, it can serve as a private AIGC database for Classic RAG or GraphRAG needed by LLMs, targeting at content mining, content analysis, and other tasks that can reveal the value of archives.</p>
</sec>
</sec>
<sec id="sec4">
<title>Findings</title>
<sec id="sec4_1">
<title>Large archival digital images solutions</title>
<p>Using scanned archival images as archival data posed several technical challenges, including handling large quantities of images, managing complex image-text interactions, ensuring description accuracy, and maintaining archival security when employing LLMs. To address these challenges, the project adopted the IIIF framework as a standard, breaking down image development into three independent yet progressive workflows.</p>
<p>First, the project implemented an IIIF Image API as a compliant image server base (see <xref ref-type="fig" rid="F4">Figure 4</xref> and <xref ref-type="fig" rid="F5">Figure 5</xref>). The Image API defined pixel-level operations for images, forming the foundation for dynamically presenting archival images, particularly high-definition images. It also enabled real-time acquisition and manipulation of finely indexed images. The project incorporated a series of engineering technologies, including container orchestration and serverless computing, to ensure system performance meets business needs when handling large quantities of high-definition images.</p>
<fig id="F4">
<label>Figure 4.</label>
<caption><p>IIIF Image API</p></caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="images\c29-fig4.jpg"><alt-text>none</alt-text></graphic>
</fig>
<fig id="F5">
<label>Figure 5.</label>
<caption><p>An example of IIIF Image API</p></caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="images\c29-fig5.jpg"><alt-text>none</alt-text></graphic>
</fig>
<p>Second, the project utilized the IIIF presentation API for object-based encapsulation and OCR Text integration. The project leveraged the flexible definition methods of the presentation API to combine individual images into logically related image objects and described metadata in multiple languages at different levels (see <xref ref-type="fig" rid="F6">Figure 6</xref>). Given the diverse layout of images, the project incorporated the ability to select appropriate areas freely and to apply different OCR models for handwritten and printed texts. The project developed an OCR text editing plugin that combines IIIF and office word, enhancing annotation flexibility and solving the problem of standardized storage of annotation results, enabling quick associations between texts and images during retrieval.</p>
<fig id="F6">
<label>Figure 6.</label>
<caption><p>Flexible definition methods of the presentation API</p></caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="images\c29-fig6.jpg"><alt-text>none</alt-text></graphic>
</fig>
<p>This solution preserves the originality of archival materials, respects the existing form of archival materials, gives them the possibility of being used as <italic>&#x2018;data&#x2019;</italic> and <italic>&#x2018;evidence&#x2019;,</italic> and ensures that archival materials can be preserved for a long time.</p>
</sec>
<sec id="sec4_2">
<title>Archival data AIGC knowledge base and LLM solutions</title>
<p>By leveraging LLMs&#x2019; language reconstruction and mining capabilities, the project significantly enhanced the explanatory power of archival data, establishing a new paradigm for archival utilization within the new technical environment. <italic>&#x2018;RAG (retrieval enhanced generation) technology is evolving, including the Classic RAG and Graph RAG, representing a significant advance in information retrieval and knowledge representation&#x2019;</italic> (<xref rid="R11" ref-type="bibr">Omrani,2024</xref>). This project explored the two approaches and built a chat function allowing users to &#x2019;chat&#x2019; with the scientist (see <xref ref-type="fig" rid="F9">Figure 9</xref>).</p>
<p>The Classic RAG system (see <xref ref-type="fig" rid="F7">Figure 7</xref>) first converted documents into Markdown format through OCR. Using the hierarchical structure of this format, the system employed an overlapping sliding window method to segment text while maintaining semantic integrity. Each segmented text block was then converted into a JSONL object containing metadata such as text content and URLs. These objects are then converted into embedding vectors stored in a vector database. During retrieval, the system augmented the context window by locating similar text in the database through nearest neighbour search.</p>
<fig id="F7">
<label>Figure 7.</label>
<caption><p>Classic RAG</p></caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="images\c29-fig7.jpg"><alt-text>none</alt-text></graphic>
</fig>
<p>GraphRAG followed this process but introduced an additional layer of inference. The LLM first performed inference operations on each sentence in the dataset, extracting entities and relationships. Named entity recognition (NER) was used to identify entities within the text and determine the relationships between entities, as well as the strength of those relationships. Then, the extracted relationships were formed into a knowledge graph, which consisted of nodes (entities) and edges (relationships) (see <xref ref-type="fig" rid="F8">Figure 8</xref>).</p>
<fig id="F8">
<label>Figure 8.</label>
<caption><p>Knowledge graph based on GraphRAG</p></caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="images\c29-fig8.jpg"><alt-text>none</alt-text></graphic>
</fig>
<p>While the Classic RAG primarily relied on vector databases to process unstructured text, the Graph RAG established a super-large-scale vocabulary by introducing knowledge graphs that equated entities and relationships with words. This enabled a more accurate understanding of content and more accurate search results.</p>
<fig id="F9">
<label>Figure 9.</label>
<caption><p>Chat interface of the project</p></caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="images\c29-fig9.jpg"><alt-text>none</alt-text></graphic>
</fig>
</sec>
<sec id="sec4_3">
<title>Towards a human-machine archival workflow</title>
<p>The Chen Taosheng archival data development project successfully implemented a systematic workflow by utilizing the open-source archival management system ArchivesSpace, along with an IIIF integrated word environment with OCR plugin, for comprehensive archival data cataloguing, describing, recognition, comparison, and management. Additionally, by extending the LLM with Classic RAG and GraphRAG, the project successfully enabled AIGC-generated texts to refer to original images. This technique not only effectively addressed potential hallucination issues in LLMs but also significantly enhanced the credibility of LLM responses by presenting the original images, thereby maximizing the utility and reliability of archival resources.</p>
<p>The integrated IIIF and OCR tool-based work environment ensured the automated generation of full-text data from archival materials and enhanced the efficiency and accuracy of the process through human-machine collaboration. The automation of repetitive tasks reduced the workload on archivists, allowing them to focus on more complex aspects of archival work. This hybrid approach not only improved efficiency but also reduced the risk of human error. The adoption of LLM supported the archival workflow in a good way by enabling the explanation ability of archival materials and enriching users&#x2019; interaction with archives.</p>
</sec>
</sec>
<sec id="sec5">
<title>Lessons and future research</title>
<p>The integration of AIGC texts with original images through IIIF objects effectively addressed the hallucination issue in LLMs, significantly enhancing the credibility of LLM responses. However, several challenges were encountered during the project experiment when interacting with the LLM. Firstly, when answering questions requiring large chunks of content, the original large chunks of content were often truncated (see <xref ref-type="fig" rid="F10">Figure 10</xref> as an example). This truncation hindered the LLM&#x2019;s ability to identify relationships between keywords, leading to responses that, although traceable, were often speculative and lacked credibility. To solve this issue, the size of the text chunks and the length of overlapping segments were typically increased to make the semantic relationships more complete. However, this approach proved insufficient for large-scale overviews unless the corpus was rich enough to provide summative content. This limitation underscored the need for GraphRAG.</p>
<fig id="F10">
<label>Figure 10.</label>
<caption><p>Large chunks of content may be truncated in the answering the question.</p></caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="images\c29-fig10.jpg"><alt-text>none</alt-text></graphic>
</fig>
<p>Secondly, ambiguity in naming entities posed another significant issue. For example, <italic>&#x2018;Mr. Chen&#x2019;</italic> and <italic>&#x2018;Professor Chen&#x2019;</italic> in the literature both referred to Chen Taosheng, however, these designations were sometimes incorrectly identified as different entities when constructing knowledge graphs. This problem was further exacerbated by OCR errors (see <xref ref-type="fig" rid="F11">Figure 11</xref> as an example). To solve this problem, cluster-based entity disambiguation is required, i.e., identifying the real-world entity to which ambiguous referents were directed by leveraging the principle that the same referent typically appeared in approximate contexts. Clustering algorithms were then applied to disambiguate these entities.</p>
<fig id="F11">
<label>Figure 11.</label>
<caption><p>Ambiguity in naming entities.</p></caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="images\c29-fig11.jpg"><alt-text>none</alt-text></graphic>
</fig>
<p>Future directions for the project include considering the integration of multi-modal large language models (MLLMs) to deeply mine semantic information from raw archival images beyond text, creating innovative modes for multi-modal archival management and development. The project team also plan to continue to optimize the human-machine collaborative archival development models and technological applications. These efforts aim to explore more efficient data processing and archival management solutions, further enhancing the intelligence and processing capabilities of archival work. Such advancements were expected to not only increase the value and utilization efficiency of archival data but also provide new theoretical support and practical guidance for the development of the archival industry.</p>
</sec>
<sec id="sec6">
<title>Conclusion</title>
<p>This project demonstrated the potential of modern technology in the AGI era in archival management through a case study of the scientist Chen Taosheng&#x2019;s archival materials. By leveraging IIIF, OCR and LLMs technologies, and implementing human-machine collaborative models, an innovative workflow for native image archival data has been developed in this project. These advancements significantly enhanced the efficiency and accuracy of archival development and utilization, providing theoretical and practical guidance for the field.</p>
<p>As the AGI era unfolded, the workflow for managing archives and records underwent a transformative evolution. The collaboration between humans and machines proved crucial in navigating the complexities of the AGI era. Human expertise in archival science ensures the contextual accuracy and ethical handling of records, while machine capabilities significantly expedite the processing and organization of vast amounts of data. This symbiotic relationship enhances the overall workflow, leveraging the strengths of both human intuition and machine precision. Ensuring the accuracy and trust of archive documents requires robust validation mechanisms, where human oversight plays a critical role in verifying the outputs generated by automated systems. By fostering a collaborative environment, archivists are able to release the full potential of AGI technologies, driving innovation in archival practices while maintaining the integrity and authenticity of archival records. This balanced approach not only preserves the historical and cultural significance of archives but also paves the way for more advanced and reliable archival management in the future.</p>
</sec>
</body>
<back>
<ref-list>
<title>References</title>
<ref id="R1"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Rolan</surname><given-names>Gregory</given-names></name><etal/></person-group><article-title>&#x2018;More human than human? Artificial intelligence in the archive.&#x2019;</article-title><source>Archives and Manuscripts</source><volume>47</volume><issue>2</issue><year>2019</year><fpage>179</fpage><lpage>203</lpage></element-citation></ref>
<ref id="R2"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Emmert-Streib</surname><given-names>Frank</given-names></name></person-group><article-title>&#x2018;Is ChatGPT the way toward artificial general intelligence.&#x2019;</article-title><source>Discover Artificial Intelligence</source><volume>4</volume><issue>1</issue><year>2024</year><fpage>1</fpage><lpage>8</lpage></element-citation></ref>
<ref id="R3"><element-citation publication-type="other"><person-group person-group-type="author"><name><surname>Morris</surname><given-names>Meredith Ringel</given-names></name><etal/></person-group><article-title>&#x2018;Levels of AGI: Operationalizing Progress on the Path to AGI.&#x2019;</article-title><source>arXiv preprint arXiv:2311.02462</source><year>2023</year></element-citation></ref>
<ref id="R4"><element-citation publication-type="other"><person-group person-group-type="author"><name><surname>Naveed</surname><given-names>Humza</given-names></name><etal/></person-group><article-title>&#x2018;A comprehensive overview of large language models.&#x2019;</article-title><source>arXiv preprint arXiv:2307.06435</source><year>2023</year></element-citation></ref>
<ref id="R5"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Bukhari</surname><given-names>Syed Saqib</given-names></name><etal/></person-group><article-title>&#x2018;anyocr: An open-source ocr system for historical archives.&#x2019;</article-title><source>2017 14th IAPR international conference on document analysis and recognition (ICDAR)</source><volume>1</volume><comment>IEEE</comment><year>2017</year></element-citation></ref>
<ref id="R6"><element-citation publication-type="book"><person-group person-group-type="author"><name><surname>Koch</surname><given-names>In&#x00EA;s</given-names></name><etal/></person-group><chapter-title>&#x2018;Knowledge graph implementation of archival descriptions through CIDOC- CRM.&#x2019;</chapter-title><source>International conference on theory and practice of digital libraries</source><publisher-loc>Cham</publisher-loc><publisher-name>Springer International Publishing</publisher-name><year>2019</year></element-citation></ref>
<ref id="R7"><element-citation publication-type="other"><person-group person-group-type="author"><name><surname>Philips</surname><given-names>James P.</given-names></name><name><surname>Nasseh</surname><given-names>Tabrizi</given-names></name></person-group><article-title>&#x2018;Historical document processing: historical document processing: a survey of techniques, tools, and trends.&#x2019;</article-title><source>arXiv preprint arXiv:2002.06300</source><year>2020</year></element-citation></ref>
<ref id="R8"><element-citation publication-type="other"><comment>Link</comment><ext-link ext-link-type="uri" xlink:href="https://iiif.io">https://iiif.io</ext-link></element-citation></ref>
<ref id="R9"><element-citation publication-type="other"><comment>Link</comment><ext-link ext-link-type="uri" xlink:href="https://archivesspace.org">https://archivesspace.org</ext-link></element-citation></ref>
<ref id="R10"><element-citation publication-type="other"><person-group person-group-type="author"><collab>SAA</collab></person-group><article-title>Describing Archives: A Content Standard.Society of American Archivists&#x2019; Technical Subcommittee</article-title><ext-link ext-link-type="uri" xlink:href="https://saa-ts-dacs.github.io/dacs/06_part_I/02_chapter_01.html">https://saa-ts-dacs.github.io/dacs/06_part_I/02_chapter_01.html</ext-link><year>2022</year></element-citation></ref>
<ref id="R11"><element-citation publication-type="other"><person-group person-group-type="author"><name><surname>Omrani</surname><given-names>Pouria</given-names></name><etal/></person-group><article-title>&#x2018;Hybrid Retrieval-Augmented Generation Approach for LLMs Query Response Enhancement.&#x2019;</article-title><year>2024</year><source>10th International Conference on Web Research (ICWR)</source><comment>IEEE, 2024</comment></element-citation></ref>
</ref-list>
</back>
</article>