<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.0 20120330//EN" "http://jats.nlm.nih.gov/publishing/1.0/JATS-journalpublishing1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:mml="http://www.w3.org/1998/Math/MathML" article-type="research-article" xml:lang="en">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">IR</journal-id>
<journal-title-group>
<journal-title>Information Research</journal-title>
</journal-title-group>
<issn pub-type="epub">1368-1613</issn>
<publisher>
<publisher-name>University of Bor&#x00E5;s</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">ir30iConf47110</article-id>
<article-id pub-id-type="doi">10.47989/ir30iConf47110</article-id>
<article-categories>
<subj-group xml:lang="en">
<subject>Research article</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>Research on the influencing mechanism of blind or visually impaired persons&#x2019; evaluation on generative AI in visual tasks</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author"><name><surname>Chen</surname><given-names>Huitong</given-names></name>
<xref ref-type="aff" rid="aff0001"/></contrib>
<contrib contrib-type="author"><name><surname>Pan</surname><given-names>Yuting</given-names></name>
<xref ref-type="aff" rid="aff0002"/></contrib>
<contrib contrib-type="author"><name><surname>Yan</surname><given-names>Hui</given-names></name>
<xref ref-type="aff" rid="aff0003"/></contrib>
<aff id="aff0001"><bold>Huitong Chen</bold> is a PhD student in Information Science at the School of Information Resource Management, Renmin University of China. Her research interests include the social impact of Artificial Intelligence, information behavior, and digital inclusion. She can be contacted at <email xlink:href="chenhuitong@ruc.edu.cn">chenhuitong@ruc.edu.cn</email></aff>
<aff id="aff0002"><bold>Yuting Pan</bold> is a PhD student in Information Science at the School of Information Resource Management, Renmin University of China. Her research interests focus on community informatics. She can be contacted at <email xlink:href="sophie_pyt@163.com">sophie_pyt@163.com</email></aff>
<aff id="aff0003"><bold>Hui Yan</bold> is Professor and Doctoral Supervisor at the School of Information Resource Management, Renmin University of China. His research interests include the social impact of Artificial Intelligence, community informatics and digital inequality. He can be contacted at <email xlink:href="hyanpku@ruc.edu.cn">hyanpku@ruc.edu.cn</email></aff>
</contrib-group>
<pub-date pub-type="epub"><day>06</day><month>05</month><year>2025</year></pub-date>
<pub-date pub-type="collection"><year>2025</year></pub-date>
<volume>30</volume>
<issue>i</issue>
<fpage>1064</fpage>
<lpage>1072</lpage>
<permissions>
<copyright-year>2025</copyright-year>
<copyright-holder>&#x00A9; 2025 The Author(s).</copyright-holder>
<license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by-nc/4.0/">
<license-p>This is an Open Access article distributed under the terms of the Creative Commons Attribution-NonCommercial 4.0 International License (<ext-link ext-link-type="uri" xlink:href="http://creativecommons.org/licenses/by-nc/4.0/">http://creativecommons.org/licenses/by-nc/4.0/</ext-link>), permitting all non-commercial use, distribution, and reproduction in any medium, provided the original work is properly cited.</license-p>
</license>
</permissions>
<abstract xml:lang="en">
<title>Abstract</title>
<p><bold>Introduction.</bold> Generative AI (GAI) has shown significant potential in assisting blind or visually impaired (BVI) Persons in visual tasks. However, existing evaluations of GAI tend to focus on technical performance, overlooking the specific usage contexts and experiences of BVI users.</p>
<p><bold>Method.</bold> This study conducted action research and semi-structured interviews with 19 BVI persons, to explore their evaluations of GAI in visual tasks and the influencing mechanism of their evaluations.</p>
<p><bold>Analysis.</bold> Following grounded theory, we identified 16 categories, and corresponding 5 core categories, as well as their relationships.</p>
<p><bold>Results.</bold> The findings indicate that BVI persons primarily evaluate GAI based on three criteria: accessibility, credibility, and interactivity. Their evaluation is influenced by four main factors: system, information, BVI user, and context. Notably, both BVI user and contextual factors moderate the influence of the system and information on user evaluation.</p>
<p><bold>Conclusions.</bold>This study develops a model that explains the influence mechanism behind the evaluation on GAI by BVI persons in visual tasks. It not only broadens the scope of human-AI interaction research by incorporating diverse user types and task contexts, but also provides an empirical foundation for developing human-centered GAI.</p>
</abstract>
</article-meta>
</front>
<body>
<sec id="sec1">
<title>Introduction</title>
<p>There have been numerous assistive technologies designed for enhancing the accessibility of BVI persons. The booming development of GAI has ushered in new possibilities for advancements in assistive technology. In 2023, Be My Eyes announced the successful deployment of Be My AI, a visual assistance tool powered by OpenAI&#x2019;s GPT-4 vision model (<xref rid="R1" ref-type="bibr">Be My Eyes, 2023</xref>). Beyond the traditional way of resorting to volunteers, BVI persons can now utilize Be My AI to process visual information, which empowers them with greater independence, and improving well-being in individual and social terms (<xref rid="R2" ref-type="bibr">Bendel, 2024</xref>).</p>
<p>However, despite the proliferation of assistive technology for BVI persons in the market, their effectiveness and satisfaction in the actual lives of BVI persons remain limited (<xref rid="R12" ref-type="bibr">Khan &#x0026; Khusro, 2021</xref>). While benchmarks in computer science effectively evaluate the performance of GAI in visual tasks, these tests tend to focus on the accuracy and efficiency of the models while ignoring the specific usage contexts and experiences of BVI persons. This disparity can lead to the development of technologies that are difficult to meet the real needs of BVI persons.</p>
<p>Therefore, this study examines the visual task encountered by BVI persons to investigate their evaluations on GAI in these tasks and explore the influencing mechanism of their evaluations. This not only provides a new theoretical perspective for research in human-intelligent interaction, but also introduces a BVI persons-centred approach to technology-driven model evaluation, offering valuable practical insights for the advancement of human-centred GAI.</p>
</sec>
<sec id="sec2">
<title>Literature review</title>
<sec id="sec2_1">
<title>Evaluation of assistive technologies for BVI persons</title>
<p>Currently, an increasing number of assistive technologies based on artificial intelligence (AI) with computer vision (CV) are under development. There have been accumulated specific research findings about the evaluation of these technologies, which can be categorized into two orientations: technology-driven and user demand-driven. Among them, the former dominates the mainstream, focusing on the performance and practical implementation of assistive technologies. Scholars in the field of computer science have constructed many benchmark datasets for vision-language models (<xref rid="R6" ref-type="bibr">Dai et al., 2023</xref>), and conducted the benchmark evaluation. Two key dimensions are involved in these evaluation metrics: automatic evaluation metrics, which principally include generated responses, Recall-Oriented Understudy for Gisting Evaluation (ROUGE), and BERT- based semantic textual similarity; human evaluation metrics which focus on the three main dimensions of correctness, actionability, and fluency (<xref rid="R28" ref-type="bibr">Yang et al., 2024</xref>; <xref rid="R18" ref-type="bibr">Zhao et al., 2024</xref>).</p>
<p>The user demand-driven evaluation approach focuses on the feedback and user experience from BVI persons. This feedback includes a rich set of metrics, including satisfaction (<xref rid="R21" ref-type="bibr">Rattanaphinyowanich &#x0026; Nunta, 2021</xref>), accuracy, reliability, accessibility, privacy, security, compatibility, energy efficiency, usability (<xref rid="R4" ref-type="bibr">Bhagat et al., 2024</xref>), functionality, aesthetics characteristics and social acceptability (Hamilton et al., 2016; <xref rid="R19" ref-type="bibr">Phillips et al., 2018</xref>). In the field of library and information science, the information needs and information behavior of BVI persons have received particular attention (<xref rid="R3" ref-type="bibr">Berget &#x0026; MacFarlane, 2020</xref>). Research has covered various topics, including the selection of information sources (<xref rid="R20" ref-type="bibr">Rahman et al., 2017</xref>; <xref rid="R5" ref-type="bibr">Chen et al., 2024</xref>), information-seeking behavior (<xref rid="R24" ref-type="bibr">Williamson et al., 2000</xref>), and interactions with information retrieval systems (Xie et al., 2021; <xref rid="R3" ref-type="bibr">Berget &#x0026; MacFarlane, 2020</xref>). Among them, the accessibility and usability issues faced by BVI persons are the primary focus (<xref rid="R25" ref-type="bibr">Xie et al., 2020</xref>). These studies can be classified into two categories according to their research approaches: for one thing, examining established evaluation criteria, such as examining whether WCAG 2.2 effectively supports BVI persons&#x2019; access to digital libraries from the perspectives of stakeholders including users, experts, and developers (<xref rid="R27" ref-type="bibr">Xie et al., 2022</xref>). For another, conducting empirical research on the challenges and perspectives faced by BVI persons in interacting with information retrieval systems. The research findings show that while the library websites are accessible according to the extracted indices of W3C, empirical data from BVI users (i.e. successful task completion, working time, satisfaction level) suggests that the websites are not easy to use (<xref rid="R17" ref-type="bibr">Najafgholinejad, 2024</xref>).</p>
</sec>
<sec id="sec2_2">
<title>Research on human-AI interaction experience</title>
<p>With the development of AI technology, the focus of HCI research is shifting from human interaction with non-AI computing systems to human interaction with AI systems, thus giving rise to the cutting-edge topic of human-AI interaction (HAII) (<xref rid="R11" ref-type="bibr">Jiang et al., 2024</xref>). Researchers have conducted extensive and thorough explorations around human interactions with AI systems such as chatbots, voice assistants, virtual humans, autonomous vehicles, etc., with application scenarios covering shopping, healthcare, and transport sector, and more. As the two main elements in human-intelligent interaction, user characteristics and AI characteristics are the focus in the research of human-intelligent interaction experience. Among them, the inherent characteristics of individual users, such as gender identity, political ideology (<xref rid="R16" ref-type="bibr">Molina &#x0026; Sundar, 2024</xref>), social role, user autonomy (<xref rid="R10" ref-type="bibr">Huh et al., 2023</xref>), motivation and social presence (<xref rid="R22" ref-type="bibr">Shao &#x0026; Kwon, 2021</xref>), health condition (<xref rid="R7" ref-type="bibr">Esmaeilzadeh et al., 2021</xref>) can affect users&#x2019; perception and evaluation of AI, as well as their willingness to adopt. Users&#x2019; evaluation or perception of AI can be expressed in terms of ease of use (<xref rid="R15" ref-type="bibr">Loske &#x0026; Klumpp, 2021</xref>), anthropomorphism (<xref rid="R18" ref-type="bibr">Pelau et al., 2021</xref>), etc. Characteristics of AI such as effects of explanation and synchronization (<xref rid="R8" ref-type="bibr">Fan et al., 2022</xref>), roles (<xref rid="R14" ref-type="bibr">Liao &#x0026; Sundar, 2021</xref>) and communication models (<xref rid="R13" ref-type="bibr">Lew &#x0026; Walther, 2023</xref>) also affect the experience of human-AI interaction. Furthermore, few studies have explored the role of task as a component in human-computer interaction. For example, people are less supportive of AI and its creators when AI performs highly (vs low) hedonic tasks (<xref rid="R29" ref-type="bibr">Yanit et al., 2023</xref>).</p>
<p>In summary, there is a relative scarcity of studies focusing on BVI users&#x2019; evaluation on assistive technologies based on AI with CV. Research in the HAII domain has focused on the influence of both humans and AI on the human-AI interaction experience, and has conducted preliminary exploration into the mechanisms between each element. However, there is still a relative lack of research on the interaction between BVI persons and AI, especially on HAII in visual tasks. To fill this research gap, this paper aims to address two specific research questions: <italic><bold>What are the evaluation criteria of BVI persons towards GAI in visual tasks? What are the underlying mechanisms influencing their evaluation?</bold></italic></p>
</sec>
</sec>
<sec id="sec3">
<title>Research methodology</title>
<sec id="sec3_1">
<title>Data collection</title>
<p>With the help of the Capital Library and China Association of persons with visual disabilities, we recruited 19 participants, and their characteristics are summarized in <xref ref-type="table" rid="T1">Table 1</xref>. Among them, minor participants will take part in the interviews accompanied by their parents. We ensured that all the participants gave their voluntary consent based on full understanding.</p>
<p>For the 12 participants who had never used AI before, we conducted one-on-one <italic>&#x2018;Be My Eyes&#x2019;</italic> operation training for them at the Capital Library during March 2024. In the training, we collected behavioral data through participatory observation. After the training, we conducted interviews to gather participants&#x2019; perceptions and evaluations of Gen AI, as well as the factors influencing their views. For the 7 participants who have experience in using GAI, we employed a semi-structured interview approach for data collection, questions centred around: visual challenges in daily life; experiences using GAI for visual tasks; evaluations of GAI and the influencing factors.</p>
<table-wrap id="T1">
<label>Table 1.</label>
<caption><p>Basic characteristics of interviewees</p></caption>
<table>
<thead>
<tr>
<th align="left" valign="top"><bold>Categories</bold></th>
<th align="left" valign="top"><bold>Subcategories</bold></th>
<th align="left" valign="top"><bold>Quantity</bold></th>
<th align="left" valign="top"><bold>Categories</bold></th>
<th align="left" valign="top"><bold>Subcategories</bold></th>
<th align="left" valign="top"><bold>Quantity</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top" rowspan="2">Gender</td>
<td align="left" valign="top">Male</td>
<td align="left" valign="top">9</td>
<td align="left" valign="top" rowspan="2">Blindness Severity</td>
<td align="left" valign="top">Totally Blind</td>
<td align="left" valign="top">12</td>
</tr>
<tr>
<td align="left" valign="top">Female</td>
<td align="left" valign="top">10</td>
<td align="left" valign="top">Low Vision</td>
<td align="left" valign="top">7</td>
</tr>
<tr>
<td align="left" valign="top" rowspan="4">Age Group</td>
<td align="left" valign="top">9-18 Years</td>
<td align="left" valign="top">7</td>
<td align="left" valign="top" rowspan="3">Experience in Using GAI</td>
<td align="left" valign="top">Never Use</td>
<td align="left" valign="top">12</td>
</tr>
<tr>
<td align="left" valign="top">19-60 Years</td>
<td align="left" valign="top">5</td>
<td align="left" valign="top">Occasional Use</td>
<td align="left" valign="top">3</td>
</tr>
<tr>
<td align="left" valign="top">41-40 Years</td>
<td align="left" valign="top">2</td>
<td align="left" valign="top">Regular Use</td>
<td align="left" valign="top">4</td>
</tr>
<tr>
<td align="left" valign="top">Over 60 Years</td>
<td align="left" valign="top">5</td>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="sec3_2">
<title>Data analysis</title>
<p>This study follows the principle and requirement of grounded theory. Data collection and analysis were alternately reciprocal and continuously comparative. Data were organized and coded immediately after each technical training or interview. The participant recruitment ended when we reached theoretical saturation. After three coding stage-open coding, axial coding, and selective coding (<xref rid="R23" ref-type="bibr">Strauss &#x0026; Corbin, 1990</xref>), we identified 16 categories, and corresponding 5 core categories. The 5 core categories include system, information, BVI user, context and evaluation. Based on the coding results, a theoretical model is proposed (<xref ref-type="fig" rid="F1">Figure 1</xref>).</p>
<fig id="F1">
<label>Figure 1.</label>
<caption><p>Model of the influence mechanism of BVI persons evaluation on GAI in visual tasks</p></caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="images\c88-fig1.jpg"><alt-text>none</alt-text></graphic>
</fig>
</sec>
</sec>
<sec id="sec4">
<title>Findings</title>
<sec id="sec4_1">
<title>Evaluation criteria of GAI by BVI persons</title>
<p>The research found that respondents&#x2019; evaluation on GAI&#x2019;s performance in visual tasks primarily focused on accessibility, interactivity, and credibility. Specifically, this manifested as: (1) the degree of accessibility and operability of GAI at the physical level and its comprehensibility at the intellectual level. (2) The degree to which GAI can offer efficient and personalized information delivery and feedback interface. (3) The degree to which GAI is honest or truthful, and can be trusted. This is specifically manifested in GAI&#x2019;s fairness and interpretability.</p>
</sec>
<sec id="sec4_2">
<title>Influence of system quality on GAI evaluation</title>
<p>The system quality of GAI is the basic guarantee for providing visual information services to BVI persons. It is further categorized into ease of use, responsiveness and model performance.</p>
<p>Ease of use represents the technical threshold of GAI, which directly affects interviewees evaluation of accessibility. The accessibility of GAI increases as the download, registration, and operating procedures of GAI get simpler, and as the number of voice prompts it offers during the capture process increases. As Interviewee #01 noted, <italic>&#x2018;it is difficult for us to take clear photos and the operation is difficult for us.&#x2019;</italic></p>
<p>Responsiveness refers to the efficiency and effectiveness of the system in responding to BVI persons requests, which greatly influences BVI persons evaluation of GAI&#x2019;s interactivity. During training, most interviewees had a particularly negative impression of the interactivity of tools like Be My Eyes due to their long processing time for images and frequent response failures. interviewee#11, who has used multiple GAI tools, had a better interactive experience in using the Luomo Toolbox, noted <italic>&#x2018;it is instantaneous and it only requires camera alignment to achieve immediate recognition.&#x2019;</italic></p>
<p>Model performance refers to the capacity of the visual-verbal model embedded in the AI tool to process visual information. This ability influences BVI persons evaluation of GAI&#x2019;s credibility and interactivity through the quality of the generated information content.</p>
</sec>
<sec id="sec4_3">
<title>Influence of information quality on GAI evaluation</title>
<p>The accuracy, objectivity, and comprehensiveness of information are essential for building trust in GAI among BVI persons. Additionally, the relevance of information influences their evaluation of the interactivity of GAI.</p>
<p>Accurate information is crucial for the safety and autonomy of BVI persons, as it forms the foundation for their reliance on GAI in decision-making. A model&#x2019;s ability to accurately identify and describe visual content directly correlates with the level of trust it can foster among BVI persons. Most interviewees express concern over the prevalent issue of AI hallucinations, hesitant to rely on GAI for some critical and extremely accurate processing of visual data.</p>
<p>The objectivity of the information described by the model also influences users&#x2019; assessment of credibility. Interviewees #02 and #10 believe that AI is less susceptible to subjective feelings and cognitive limits when compared to humans, thus allowing for a more accurate and objective reflection of visual information. This perception leads them to view GAI as fair and inclusive.</p>
<p>Comprehensive information offers broad and complete coverage of the necessary or expected aspects of a problem or subject. Interviewees believe that the more detailed the GAI&#x2019;s description of visual information, the clearer their panoramic understanding of the environment, which in turn enhances their evaluation of the credibility of GAI. Interestingly, the comprehensive information provided by AIGC can create valuable Information Encountering. For instance, interviewee #10 casually used Be My Eyes to take a photo of the front and from AI&#x2019;s detailed description, he learned there was a bench ahead, allowing him to sit down and take a break. This unexpected help provided practical assistance and enhanced his trust in GAI.</p>
<p>The relevance of generated information content is manifested in its relevance to the situation and emotions of BVI persons. The greater the relevance of information, the better GAI can understand and respond to the personalized needs of BVI persons, resulting in higher user ratings of interactivity. For example, in a cooking scenario, interviewees only need GAI to quickly capture the names and expiration dates of spices. However, the comprehensiveness of AI-generated information can make it difficult for BVI persons to quickly access the required information, thus reducing interaction efficiency. In reading scenarios, both interviewee#7 and interviewee#2 agreed that the GAI&#x2019;s voice <italic>&#x2018;sounds like a customer service, devoid of any emotional expression in its reading&#x2019;.</italic> This mechanized presentation provides a poor interactive experience for BVI persons.</p>
</sec>
<sec id="sec4_4">
<title>Influence of user characteristics on GAI evaluation</title>
<p>User dimensions include their GAI use experience, familiarity with GAI, and personality characteristics. As users accumulate experience, BVI persons gradually become more familiar with the operation and functions of GAI, which in turn enhances their evaluation of accessibility and interactivity. In terms of personality, some interviewees consider themselves conservative and cautious, which leads to a natural distrust of new technologies such as GAI. On the other hand, interviewees with more positive and open-minded personalities tend to rate GAI with higher credibility.</p>
<p>Additionally, user dimensions can moderate the intensity and direction of the influence that system and information factors have on user evaluations. Users with prior experience, familiarity with GAI, and open-minded personality are more likely to be satisfied with the system and information provided by GAI, leading to more positive evaluations. For instance, experienced interviewees are generally accustomed to the response time of GAI, leading to higher evaluations of its interactivity.</p>
</sec>
<sec id="sec4_5">
<title>Influence of contextual factors on GAI evaluation</title>
<p>The visual task, physical environment and social relationships experienced by BVI persons directly influence their assessment of GAI. Additionally, contextual factors can moderate the intensity and direction of the influence that system and information factors have on user evaluations.</p>
<p>In visual tasks that are urgent and important related to personal safety, such as reading hospital reports, medication information, or crossing the street, BVI persons may perceive that the responsiveness of GAI and the quality of the information it provides are insufficient to meet their need, leading to a decrease in their evaluation of its credibility and interactivity. However, in non-urgent and less important visual tasks, such as ordering food, checking nail designs, or picking up keys, BVI persons believe that using GAI reduces their reliance on and disruption to others, resulting in higher evaluations of its accessibility.</p>
<p>The physical environment refers to the objective conditions under which BVI persons encounter visual challenges, including sound environment, physical layouts, and network infrastructure. BVI persons primarily rely on auditory feedback to compensate for visual deficiencies. In noisy environment, they find GAI less accessible and interactive. The layout and design of the physical environment directly affect BVI persons&#x2019; interactions with GAI. Many interviewees indicated that if commonly used items and devices in their home or workplaces are in fixed position, they would be able to take more focused photos, thus facilitating efficient information transfer and feedback with GAI. The quality of network infrastructure impacts the responsiveness of GAI systems, which in turn impacts users&#x2019; perceptions of the accessibility and interactivity of Gen AI.</p>
<p>Social relationships, particularly strong ties, have a direct bearing on BVI persons evaluation of GAI&#x2019;s accessibility. Different parenting styles exert vastly different influences. For instance, the parents of interviewee#15 hold a very open-minded attitude towards AI and even suggested creating an AI chat group for her child after training, enabling interviewee#15 to successfully access and use GAI. In contrast, parents of interviewee#16 adopted strict control over their children, prohibiting interviewee#16 from accessing and using ICT. Consequently, interviewee#16 perceives restricted access to GAI.</p>
</sec>
</sec>
<sec id="sec5">
<title>Conclusion</title>
<p>This study delves into the evaluation criteria and influencing mechanisms of GAI from the perspective of BVI persons in visual tasks. The findings indicate that BVI persons&#x2019; evaluations of GAI are primarily based on three criteria: accessibility, credibility, and interactivity. As traditional human-computer interaction transitions to human-AI interaction, these three evaluation criteria evolve to incorporate new dimensions. BVI persons&#x2019; evaluation are directly influenced by four dimensions: system, information, user, and context. Furthermore, the user and context factors moderate the intensity and direction of the influence that system and information have on user evaluations. Specifically, ease of use and responsiveness of the system influence evaluations of accessibility and interactivity. The model performance influences evaluations of credibility and interactivity through the quality of the generated content. Additionally, user&#x2019;s experience and familiarity with GAI influence their evaluations of accessibility and interactivity. Visual tasks with high urgency and importance requirements tend to decrease the perceived credibility and interactivity of GAI. The physical environment impacts all three evaluation criteria, while social relationships primarily affect the evaluation of accessibility.</p>
<p>The main contributions of this study are as follows: It focuses on the interaction between BVI persons and GAI in visual task contexts, enriching the user types and diversifying task contexts in human-AI interaction research. Moreover, in the technology-driven domain of GAI evaluation studies, this study provides a comprehensive user-centred perspective based on grounded theory, shedding light on the factors and mechanisms that influence BVI persons&#x2019; evaluations of GAI. However, the study has certain limitations. As an exploratory study, its primary focus is on identifying new variables and relationships. The stability of these relationships, however, requires further validation through larger sample sizes in future research.</p>
</sec>
</body>
<back>
<ack>
<title>Acknowledgements</title>
<p>The research is supported by National Social Science Foundation of China (Grant No. 23AZD093). This paper is one of the key projects of the National Social Science Fund of China&#x2019;s &#x2018;The Disruptive Impact of artificial intelligence on the National Economy and Research on Information Governance&#x2019;. In addition, we would like to express our gratitude to Qihao Wang for inspiring the topic selection, as well as for the assistance provided during the data analysis process. We are also deeply thankful to the librarians of the Capital Library, Xuefeng Zhao, Qiangdong Li, Jing Su, Ruoxuan Yang, and Puhang Yu, for their unwavering support and assistance during our technical training sessions.</p>
</ack>
<ref-list>
<title>References</title>
<ref id="R1"><element-citation publication-type="other"><person-group person-group-type="author"><collab>Be My Eyes</collab></person-group><year>2023</year><article-title>Be My Eyes Integrates Be My AI&#x2122; into its First Contact Center with Stunning Results</article-title><ext-link ext-link-type="uri" xlink:href="https://www.bemyeyes.com/blog/introducing-microsofts-ai-powered-disability- answer-desk-on-be-my-eyes">https://www.bemyeyes.com/blog/introducing-microsofts-ai-powered-disability- answer-desk-on-be-my-eyes</ext-link></element-citation></ref>
<ref id="R2"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Bendel</surname><given-names>O.</given-names></name></person-group><year>2024</year><article-title>How Can Generative AI Enhance the Well-being of Blind?</article-title><source>Proceedings of the AAAI Symposium Series</source><volume>3</volume><issue>1</issue><fpage>340</fpage><lpage>347</lpage></element-citation></ref>
<ref id="R3"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Berget</surname><given-names>G.</given-names></name><name><surname>MacFarlane</surname><given-names>A.</given-names></name></person-group><year>2020</year><article-title>What is known about the impact of impairments on information seeking and searching?</article-title> <source>Journal of the Association for Information Science and Technology</source><volume>71</volume><issue>5</issue><fpage>596</fpage><lpage>611</lpage></element-citation></ref>
<ref id="R4"><element-citation publication-type="other"><person-group person-group-type="author"><name><surname>Bhagat</surname><given-names>S.</given-names></name><name><surname>Joshi</surname><given-names>P.</given-names></name><name><surname>Agarwal</surname><given-names>A.</given-names></name><name><surname>Gupta</surname><given-names>S.</given-names></name></person-group><year>2024</year><article-title>Accessibility evaluation of major assistive mobile applications available for the visually impaired</article-title><source>arXiv preprint arXiv:2407.17496</source></element-citation></ref>
<ref id="R5"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Chen</surname><given-names>H.</given-names></name><name><surname>Yan</surname><given-names>H.</given-names></name><name><surname>Wu</surname><given-names>Z.</given-names></name><name><surname>Zhao</surname><given-names>X.</given-names></name></person-group><year>2024</year><article-title>Silicon-based Life or Carbon-based Life? An Exploratory Study on Visual Information Source Selection of Blind or Visually Impaired Persons</article-title><source>Proceedings of the Association for Information Science and Technology</source><volume>61</volume><issue>1</issue><fpage>493</fpage><lpage>498</lpage></element-citation></ref>
<ref id="R6"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Dai</surname><given-names>W.</given-names></name><name><surname>Li</surname><given-names>J.</given-names></name><name><surname>Li</surname><given-names>D.</given-names></name><name><surname>Tiong</surname><given-names>A.</given-names></name> <name><surname>meng</surname> <given-names>huat</given-names></name><name><surname>Zhao</surname><given-names>J.</given-names></name><name><surname>wang</surname><given-names>W.</given-names></name><name><surname>li</surname><given-names>B.</given-names></name><name><surname>Fung</surname><given-names>P.</given-names></name><name><surname>Hoi</surname><given-names>S.</given-names></name></person-group><year>2023</year><article-title>InstructBLIP: Towards General-Purpose Vision-Language Models with Instruction Tuning</article-title><source>ArXiv</source><volume>2305</volume><issue>06500v2</issue><ext-link ext-link-type="uri" xlink:href="https://doi.org/10.48550/arXiv.2305.06500">https://doi.org/10.48550/arXiv.2305.06500</ext-link></element-citation></ref>
<ref id="R7"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Esmaeilzadeh</surname><given-names>P.</given-names></name><name><surname>Mirzaei</surname><given-names>T.</given-names></name><name><surname>Dharanikota</surname><given-names>S.</given-names></name></person-group><year>2021</year><article-title>Patients perceptions toward human&#x2013; artificial intelligence interaction in health care: experimental study</article-title><source>Journal of medical Internet research</source><volume>23</volume><issue>11</issue><fpage>e25856</fpage></element-citation></ref>
<ref id="R8"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Fan</surname><given-names>M.</given-names></name><name><surname>Yang</surname><given-names>X.</given-names></name><name><surname>Yu</surname><given-names>T.</given-names></name><name><surname>Liao</surname><given-names>Q.V.</given-names></name><name><surname>Zhao</surname><given-names>J.</given-names></name></person-group><year>2022</year><article-title>Human-ai collaboration for UX evaluation: effects of explanation and synchronization</article-title><source>Proceedings of the ACM on Human-Computer Interaction</source><volume>6</volume><issue>CSCW1</issue><fpage>1</fpage><lpage>32</lpage></element-citation></ref>
<ref id="R9"><element-citation publication-type="other"><person-group person-group-type="author"><name><surname>Hamilton-Fletcher</surname><given-names>G.</given-names></name><name><surname>Obrist</surname><given-names>M.</given-names></name><name><surname>Watten</surname><given-names>P.</given-names></name><name><surname>Mengucci</surname><given-names>M.</given-names></name><name><surname>Ward</surname><given-names>J.</given-names></name></person-group><year>2016</year><article-title>&#x2018;I Always Wanted to See the Night Sky&#x2019; Blind User Preferences for Sensory Substitution Devices</article-title><source>Proceedings of the 2016 CHI Conference on Human Factors in Computing Systems</source><fpage>2162</fpage><lpage>2174</lpage></element-citation></ref>
<ref id="R10"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Huh</surname><given-names>J.</given-names></name><name><surname>Whang</surname><given-names>C.</given-names></name><name><surname>Kim</surname><given-names>H.Y.</given-names></name></person-group><year>2023</year><article-title>Building trust with voice assistants for apparel shopping: The effects of social role and user autonomy</article-title><source>Journal of Global Fashion Marketing</source><volume>14</volume><issue>1</issue><fpage>5</fpage><lpage>19</lpage></element-citation></ref>
<ref id="R11"><element-citation publication-type="other"><person-group person-group-type="author"><name><surname>Jiang</surname><given-names>T.</given-names></name><name><surname>Sun</surname><given-names>Z.</given-names></name><name><surname>Fu</surname><given-names>S.</given-names></name><name><surname>Lv</surname><given-names>Y.</given-names></name></person-group><year>2024</year><article-title>Human-AI interaction research agenda: A user-centered perspective</article-title><source>Data and Information Management</source><fpage>100078</fpage></element-citation></ref>
<ref id="R12"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Khan</surname><given-names>A.</given-names></name><name><surname>Khusro</surname><given-names>S.</given-names></name></person-group><year>2021</year><article-title>An insight into smartphone-based assistive solutions for visually impaired and blind people: issues, challenges and opportunities</article-title><source>Universal Access in the Information Society</source><volume>20</volume><issue>2</issue><fpage>265</fpage><lpage>298</lpage></element-citation></ref>
<ref id="R13"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Lew</surname><given-names>Z.</given-names></name><name><surname>Walther</surname><given-names>J.B.</given-names></name></person-group><year>2023</year><article-title>Social scripts and expectancy violations: Evaluating communication with human or AI chatbot interactants</article-title><source>Media Psychology</source><volume>26</volume><issue>1</issue><fpage>1</fpage><lpage>16</lpage></element-citation></ref>
<ref id="R14"><element-citation publication-type="other"><person-group person-group-type="author"><name><surname>Liao</surname><given-names>M.</given-names></name><name><surname>Sundar</surname><given-names>S.S.</given-names></name></person-group><year>2021</year><article-title>How should AI systems talk to users when collecting their personal information? Effects of role framing and self-referencing on human-AI interaction</article-title><source>Proceedings of the 2021 CHI conference on human factors in computing systems</source><fpage>1</fpage><lpage>14</lpage></element-citation></ref>
<ref id="R15"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Loske</surname><given-names>D.</given-names></name><name><surname>Klumpp</surname><given-names>M.</given-names></name></person-group><year>2021</year><article-title>Intelligent and efficient? An empirical analysis of human&#x2013;AI collaboration for truck drivers in retail logistics</article-title><source>The International Journal of Logistics Management</source><volume>32</volume><issue>4</issue><fpage>1356</fpage><lpage>1383</lpage></element-citation></ref>
<ref id="R16"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Molina</surname><given-names>M.D.</given-names></name><name><surname>Sundar</surname><given-names>S.S.</given-names></name></person-group><year>2024</year><article-title>Does distrust in humans predict greater trust in AI? Role of individual differences in user responses to content moderation</article-title><source>New Media &#x0026; Society</source><volume>26</volume><issue>6</issue><fpage>3638</fpage><lpage>3656</lpage></element-citation></ref>
<ref id="R17"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Najafgholinejad</surname><given-names>A.</given-names></name></person-group><year>2024</year><article-title>Accessibility and usability of user interfaces of library information retrieval systems from the perspective of visually impaired users</article-title><source>Library and Information Sciences</source><volume>26</volume><issue>4</issue><fpage>83</fpage><lpage>110</lpage></element-citation></ref>
<ref id="R18"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Pelau</surname><given-names>C.</given-names></name><name><surname>Dabija</surname><given-names>D.C.</given-names></name><name><surname>Ene</surname><given-names>I.</given-names></name></person-group><year>2021</year><article-title>What makes an AI device human-like? The role of interaction quality, empathy and perceived psychological anthropomorphic characteristics in the acceptance of artificial intelligence in the service industry</article-title><source>Computers in Human Behavior</source><volume>122</volume><fpage>106855</fpage></element-citation></ref>
<ref id="R19"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Phillips</surname><given-names>M.</given-names></name><name><surname>Proulx</surname><given-names>M.J.</given-names></name></person-group><year>2018</year><article-title>Social interaction without vision: an assessment of assistive technology for the visually impaired</article-title><source>Technology &#x0026; Innovation</source><volume>20</volume><issue>1-2</issue><fpage>85</fpage><lpage>93</lpage></element-citation></ref>
<ref id="R20"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Rahman</surname><given-names>A.L. A.</given-names></name><name><surname>Razali</surname><given-names>T.R.a. T.</given-names></name><name><surname>Ghazali</surname><given-names>A.M.</given-names></name><name><surname>Kamarudin</surname><given-names>M.H.</given-names></name></person-group><year>2017</year><article-title>BVIC&#x2019;s CIS in the technological environment</article-title><source>Advanced Science Letters</source><volume>23</volume><issue>1</issue><fpage>151</fpage><lpage>155</lpage></element-citation></ref>
<ref id="R21"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Rattanaphinyowanich</surname><given-names>T.</given-names></name><name><surname>Nunta</surname><given-names>S.</given-names></name></person-group><year>2021</year><article-title>Development of DAISY-WIBORD as computer assisted learning facilities for children with visual impairment</article-title><source>Journal of Physics</source><volume>1835</volume><issue>1</issue><fpage>012080</fpage></element-citation></ref>
<ref id="R22"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Shao</surname><given-names>C.</given-names></name><name><surname>Kwon</surname><given-names>K.H.</given-names></name></person-group><year>2021</year><article-title>Hello Alexa! Exploring effects of motivational factors and social presence on satisfaction with artificial intelligence-enabled gadgets</article-title><source>Human Behavior and Emerging Technologies</source><volume>3</volume><issue>5</issue><fpage>978</fpage><lpage>988</lpage></element-citation></ref>
<ref id="R23"><element-citation publication-type="book"><person-group person-group-type="author"><name><surname>Strauss</surname><given-names>A.</given-names></name><name><surname>Corbin</surname><given-names>J.M.</given-names></name></person-group><year>1990</year><source>Basics of qualitative research: Grounded theory procedures and techniques</source><publisher-name>Sage Publications, Inc</publisher-name></element-citation></ref>
<ref id="R24"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Williamson</surname><given-names>K.</given-names></name><name><surname>Schauder</surname><given-names>D.</given-names></name><name><surname>Bow</surname><given-names>A.</given-names></name></person-group><year>2000</year><article-title>Information seeking by blind and sight impaired citizens: an ecological study</article-title><source>Information research</source><volume>5</volume><issue>4</issue><fpage>4</fpage><lpage>5</lpage></element-citation></ref>
<ref id="R25"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Xie</surname><given-names>I.</given-names></name><name><surname>Babu</surname><given-names>R.</given-names></name><name><surname>Lee</surname><given-names>T.H.</given-names></name><name><surname>Castillo</surname><given-names>M.D.</given-names></name><name><surname>You</surname><given-names>S.</given-names></name><name><surname>Hanlon</surname><given-names>A.M.</given-names></name></person-group><year>2020</year><article-title>Enhancing usability of digital libraries: Designing help features to support blind and visually impaired users</article-title><source>Information Processing &#x0026; Management</source><volume>57</volume><issue>3</issue><fpage>102110</fpage></element-citation></ref>
<ref id="R26"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Xie</surname><given-names>I.</given-names></name><name><surname>Babu</surname><given-names>R.</given-names></name><name><surname>Lee</surname><given-names>H.S.</given-names></name><name><surname>Wang</surname><given-names>S.</given-names></name><name><surname>Lee</surname><given-names>T.H.</given-names></name></person-group><year>2021</year><article-title>Orientation tactics and associated factors in the digital library environment: Comparison between blind and sighted users</article-title><source>Journal of the Association for Information Science and Technology</source><volume>72</volume><issue>8</issue><fpage>995</fpage><lpage>1010</lpage></element-citation></ref>
<ref id="R27"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Xie</surname><given-names>I.</given-names></name><name><surname>Babu</surname><given-names>R.</given-names></name><name><surname>Wang</surname><given-names>S.</given-names></name><name><surname>Lee</surname><given-names>H.S.</given-names></name><name><surname>Lee</surname><given-names>T.H.</given-names></name></person-group><year>2022</year><article-title>Assessment of digital library design guidelines to support blind and visually impaired users: a study of key stakeholders&#x2019; perspectives</article-title><source>The Electronic Library</source><volume>40</volume><issue>6</issue><fpage>646</fpage><lpage>661</lpage></element-citation></ref>
<ref id="R28"><element-citation publication-type="other"><person-group person-group-type="author"><name><surname>Yang</surname><given-names>B.</given-names></name><name><surname>He</surname><given-names>L.</given-names></name><name><surname>Liu</surname><given-names>K.</given-names></name><name><surname>Yan</surname><given-names>Z.</given-names></name></person-group><year>2024</year><article-title>VIAssist: Adapting Multi-modal Large Language Models for Users with Visual Impairments</article-title><source>arXiv preprint arXiv:2404.02508</source></element-citation></ref>
<ref id="R29"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Yanit</surname><given-names>M.</given-names></name><name><surname>Yanit</surname><given-names>M.</given-names></name><name><surname>Wan</surname><given-names>F.</given-names></name></person-group><year>2023</year><article-title>Right agent, wrong level of hedonism: How high (vs low) hedonic values in AI-performed tasks lead to decreased perceptions of humanlikeness, warmth, and less consumer support</article-title><source>Computers in Human Behavior</source><volume>147</volume><fpage>107870</fpage></element-citation></ref>
<ref id="R30"><element-citation publication-type="other"><person-group person-group-type="author"><name><surname>Zhao</surname><given-names>Y.</given-names></name><name><surname>Zhang</surname><given-names>Y.</given-names></name><name><surname>Xiang</surname><given-names>R.</given-names></name><name><surname>Li</surname><given-names>J.</given-names></name><name><surname>Li</surname><given-names>H.</given-names></name></person-group><year>2024</year><article-title>Vialm: A survey and benchmark of visually impaired assistance with large models</article-title><source>arXiv preprint arXiv:2402.01735</source></element-citation></ref>
</ref-list>
</back>
</article>