ISSN: 2976-7466 | Open Access

Journal of Biosensors and Bioelectronics Research

Advancing Sentiment Analysis Through Emotionally-Agnostic Text Mining in Large Language Models (LLMS)

Author(s): Jeremiah Ratican and James Hutson*

Abstract

The conventional methodology for sentiment analysis within large language models (LLMs) has predominantly drawn upon human emotional frameworks, incorporating physiological cues that are inherently absent in text-only communication. This research proposes a paradigm shift towards an emotionallyagnostic approach to sentiment analysis in LLMs, which concentrates on purely textual expressions of sentiment, circumventing the confounding effects of human physiological responses. The aim is to refine sentiment analysis algorithms to discern and generate emotionally congruent responses strictly from text-based cues. This study presents a comprehensive framework for an emotionally-agnostic sentiment analysis model that systematically excludes physiological indicators whilst maintaining the analytical depth required for accurate emotion detection. A novel suite of metrics tailored to this approach is developed, facilitating a nuanced interpretation of sentiment within text data, which is paramount for enhancing user interaction across a spectrum of text mining applications, including recommendation systems and interactive AI characters. The research undertakes a critical comparative analysis, juxtaposing the newly proposed model with traditional sentiment analysis techniques, to evaluate efficacy enhancements and to substantiate its application potential. Further, the investigation delves into the short-term memory capabilities of LLMs, exploring the implications for character AI roleplaying interfaces and their ability to recall and respond to user input within a text-driven emotional framework. Findings indicate that this emotionally-agnostic sentiment analysis approach not only simplifies the sentiment assessment process within LLMs but also opens avenues for a more precise and contextually appropriate emotional response generation. This abstracted model, devoid of human physiological constraints, represents a significant advancement in text mining, fostering improved interactions and contributing to the evolution of LLMs as adept analytical tools in various domains where emotional intelligence is crucial.

Introduction

The nascent field of sentiment analysis, which endeavors to computationally discern and categorize affective states conveyed in text, is approaching a transformative epoch facilitated by the proliferation of large language models (LLMs). These computational behemoths, exemplified by OpenAI’s ChatGPT 3.5 and 4 iterations, have reached milestones in producing text with an unprecedented degree of linguistic sophistication and contextual relevance, simulating nuanced dialogues that bear striking resemblance to human exchanges [1]. These technological leaps have not only broadened the horizons of possible applications across diverse sectors such as healthcare, education, and assisted technologies but have also underscored the imperatives for advancements in sentiment analysis to complement the emotional depth of human-AI interaction [2-5].

This surging tide of interest in emulating human cognitive complexity encompasses a significant branch that critically examines the integration of emotional intelligence within AI systems. Existing research underscores that the learning algorithms undergirding LLMs possess the potential to reflect a spectrum of personalities and emotional responses fostering more profound user interactions and potentially elevating the user experience to new pinnacles of personalization [6, 7].

The present research trajectory anchors on an intellectually rigorous integration of large language models with emotionally agnostic text mining, a perspective resonating with modular mind theory. laid the initial theoretical groundwork, suggesting the compartmentalization of cognitive processes into discrete, functional modules within the brain. expanded upon this, advocating for the parallel in artificial intelligence constructs, presenting a compelling case for modular representations within AI systems [8, 9].

Against this conceptual backdrop, the current study advances the dialogue by proposing an emotionally agnostic framework for sentiment analysis. This framework seeks to refine text mining in LLMs by eliminating physiological cues inherent to human emotion, focusing purely on textual information to derive sentiment. This emotionally neutral approach envisages a more nuanced understanding of textual data, unencumbered by the complexities and biases introduced by human emotional expression. The integration of such an emotionally agnostic sentiment analysis with the robust capabilities of LLMs represents a nascent frontier with profound implications for the future of AI and human-AI interaction.

Expanding on a formidable theoretical foundation and an expanding corpus of empirical research, the current article aspires to chart a new course in the emulation of human cognitive complexity by converging LLMs with autonomous agents. This synthesis, predicated on the conceptual pillars of the modular mind theory, aspires to enhance the dialogue within the intersecting spheres of artificial intelligence, psychology, and neuroscience. Grounded in the seminal work of which underscored the capability of autonomous agents to reflect human behaviors through the integration of personal motives and preferences, the current study proposes an avant-garde model that meticulously replicates the nuanced facets of mental processing [10].

The scope of this article extends beyond the conventional frameworks of sentiment analysis, advocating for an approach that meticulously separates emotional biases from the interpretation of textual data. The endeavor here is to engender a model that aligns more closely with the principles of emotion-agnostic text mining, thereby promoting an objective stance in sentiment analysis within the realm of LLMs. Such a model is anticipated to provide a richer, more accurate analysis of text by disentangling the intricacies of human sentiment from the underlying communicative intent. Furthermore, the study argues for the potential of LLMs to advance sentiment analysis by adopting a text-centric perspective that eschews human-like emotional biases. By constructing a synthetic analog of human cognition, partitioned into distinct modules for different aspects of mental processing, the study articulates a novel methodology that is hoped to be instrumental in advancing the field.

 The aspirations of this work are manifold: to provide an innovative methodological approach that can be adopted by future research endeavors; to leverage the resultant insights for the development of more refined, emotionally intelligent AI systems; and to potentially uncover new dimensions of human cognition through the lens of artificial constructs. Thus, the proposed model seeks not only to elucidate the functionalities of LLMs in mimicking human cognitive processes but also to serve as a vanguard for subsequent technological innovations that straddle the domains of cognitive science and artificial intelligence. Through this synthesis of theoretical exploration and empirical validation, the article endeavors to pioneer an augmented pathway for understanding and replicating the vast complexities of human sentiment and cognition.

Literature Review

The scholarly investigation into the nexus of large language models (LLMs) and sentiment analysis within the literature unveils a nuanced conversation concerning the development of emotionally intelligent AI, as well as the psychosocial and ethical dimensions such developments entail. Researchers have posited that emotional acuity plays a significant role in intelligent behavior and decision-making, which traditionally has been integrated into AI systems to reflect more human-like interactions [11, 12]. For example, Mahmud illuminate the significance of emotions as a signaling system that informs intelligent decision-making, thereby integrating emotional awareness into the computational fabric of AI. The grafting of these emotional components onto AI has marked a turning point in the AI evolutionary trajectory [13].

 Frameworks advocating for AI systems to be endowed with social and emotional faculties have been presented as essential developments for fostering more intricate human-machine interactions [14, 15]. Observations by Samsonovich suggest that the embodiment of empathy within artificial entities can lead to more genuine social exchanges, supporting the notion that emotional intelligence within AI could enhance interpersonal dynamics. This is particularly relevant in the context of developing communication systems that leverage affective computing to interpret and reproduce emotional states, leading to a range of applications from empathetic chatbots to sector-specific user interfaces [16, 17].

 The interplay between ethical considerations and the practical deployment of AI with emotional dimensions is a topic of vigorous debate among scholars [18,19]. The ethical quandary regarding the simulation of human emotions by machines, as posited by Pusztahelyi underscores the urgent need for a moral framework within which emotionally intelligent AI operates. Furthermore, the pursuit of affective computing has gradually shifted towards developing more sophisticated cognitive models that aim to accurately embody emotional intelligence within AI systems, indicating a sustained commitment to refine these emergent technologies [20].

 Pivoting from these established insights, the present review explores the emerging paradigm of emotion-agnostic text mining within LLMs, positing a novel approach wherein the analysis of textual data is decoupled from human-like emotional biases. This literature review aims to synthesize current research that informs the design and implementation of LLMs capable of such emotion-agnostic sentiment analysis, framing a dialogue that interrogates the potential and limitations of current technologies, while exploring the future trajectory of emotionally intelligent AI absent of anthropomorphic emotional frameworks.

The literature indicates that the healthcare sector stands as a particularly promising arena for the application of emotionally intelligent AI systems. Tools for emotion recognition are posited as beneficial adjuncts to medical professionals, assisting in nuanced aspects of patient care and interaction [21]. Such applications, however, are not without their ethical considerations, particularly regarding the degree of autonomy these AI systems are granted in decision-making processes, a concern highlighted by [22]. Moreover, Andersson introduces a critical discourse on the potential implications for fundamental human rights, such as the right to freedom of thought, in the deployment of AI systems endowed with emotional intelligence [23].

The compendium of existing scholarship offers a comprehensive portrait of the multifarious aspects of emotional intelligence within AI: from foundational technological capabilities and beneficial integrations into cognitive systems, to frameworks designed to foster socio-emotionally adept AI. This is augmented by research into the communicative functions of AI, the progression of synthetic emotional intelligence, and the ethical considerations associated with the simulation of emotions by machines. The inclusion of healthcare-related applications and the legal intricacies that these technologies may encounter presents a panoramic overview of the advancements and future directions in the field of emotionally intelligent AI.

In addition to healthcare, the domains of video gaming and complex decision-making processes represent burgeoning fields of interest within the scholarly community. Pioneering studies by have delved into the implementation of reinforcement learning to train AI agents for adaptive gameplay [24, 25]. These foundational studies elucidate methods by which AI can learn and evolve within specific game environments. Moreover, research by supports the proposition that artificial agents can serve as proxies for human cognitive processes, particularly in the context of decision-making in games [26, 27]. Provide evidence that generative personas can yield behavioral outputs that closely mirror those of human players, underscoring the capacity of AI to effectively simulate complex human decision-making [27].

Within this research landscape, the current article seeks to explore the prospects of LLMs functioning within a framework that is agnostic to emotional biases. The aim is to dissect and understand the mechanisms by which LLMs can engage in sentiment analysis without emulating emotional responses. This exploration is rooted in the premise that such a development could address some of the ethical concerns presented by emotionally intelligent systems while maintaining their functional benefits in various sectors, including healthcare and gaming. The literature review thus sets the stage for a critical analysis of the potential for LLMs to operate within an emotionally neutral paradigm.

The burgeoning field of video game testing and evaluation has seen significant contributions from AI agents, a trend underscored by the work of [28], who have demonstrated the capability of synthetic agents to match human testers in the identification of software anomalies. Notwithstanding these advances, have identified a limitation in the form of a lack of behavioral diversity among agents, which can lead to predictability in their actions. This issue is echoed in subsequent research, with exploring the expansive possibilities for AI in video games, and addressing the intricate process of automating player character behavior [28-33].

In parallel, there has been a discernible shift towards the application of generative AI models such as ChatGPT across scientific and medical disciplines. The work of through extensive interviews has shed light on the potential for generative AI to expedite scientific breakthroughs. Similarly, propose that these models could refine statistical process control; however, they concurrently signal the risks of misuse and misinterpretation inherent in early-stage technologies. Further elaborate on the application of generative AI in spinal cord injury research, evidencing its efficacy in generating virtual models and enhancing medical procedures [34-36].

Recent breakthroughs have elucidated the potential of generative AI to embody a range of personas within virtual settings. A notable study by Park has contributed to this dialogue by engineering a roleplaying game environment inhabited by AI agents, each characterized by distinct personal narratives and social interplays. Employing the ChatGPT API, these researchers developed an advanced simulation framework featuring agents with individualized memories and life stories. When subjected to evaluations of behavior authenticity, these AI-driven simulations demonstrated a higher degree of realism compared to human participants engaging in role-play. Yet, the authors prudently highlight potential ethical dilemmas, such as the emergence of parasocial interactions and excessive dependence on AI constructs [10].

Synthesizing these diverse strands of investigation, it becomes evident that generative AI agents are increasingly integrated into a spectrum of domains, from entertainment to scientific inquiry and healthcare. Their proficiency in emulating complex human behavior is recognized, but is accompanied by explicit calls for responsible utilization. Particular attention is drawn to the ethical dimensions of such interactions, with an emphasis on maintaining a balance between AI advancements and human welfare. Thus, the scholarly narrative acknowledges the transformative nature of generative AI, while simultaneously advocating for a deliberate approach to understanding and addressing the challenges associated with their integration into human-centric domains.

Regarding the limitations of current approaches to large language models for sentiment analysis, the literature hints at several gaps [37]. These include challenges in understanding contextual nuance, detecting sarcasm, and the potential reinforcement of biases present in training data [38]. Although large language models like ChatGPT have made significant strides in text generation and comprehension, their capacity for sentiment analysis remains an area ripe for further refinement. Current models may struggle with the subtleties of human emotion, often misinterpreting or failing to detect the complex layers of sentiment embedded in natural language [39]. This indicates a need for continuous enhancement in algorithmic sophistication and a broader ethical dialogue on the deployment of such technologies in sectors where emotional intelligence is paramount.

Recommendations

The burgeoning domain of sentiment analysis within data mining is primed for transformation through the innovative application of LLMs like GPT-4. Traditional sentiment analysis methods frequently integrate affective computing, which attempts to detect and interpret human emotions through text [40]. However, these methods often rely on physiological or experiential data which may not be available or appropriate in textual analysis [41]. Therefore, a novel approach has been proposed that focuses exclusively on the textual elements of user input, eschewing the need for humanbased emotional models [35].

The proposed strategy posits that by analyzing text through a purely linguistic lens—stripping away the influence of physiological cues—one could discern a more nuanced spectrum of sentiment. LLMs, equipped with the capability to process and generate text, are uniquely positioned to leverage such a strategy. These models can be fine-tuned to identify and categorize sentiments into emotional archetypes or categories, thereby enabling a deeper understanding of textual sentiment without reliance on physical emotional indicators. The proposed methodology would entail creating a framework within LLMs that can classify sentiment across a range of finely grained emotional categories. The system would quantify these sentiments, providing a nuanced emotional profile of textual input. Such a framework could draw upon a database of text-based interactions, eschewing the need for emotional labels derived from human annotators, which often incorporate subconscious biases related to physiological responses.

Furthermore, the model would be trained to discard any physiological elements that may confound the sentiment analysis, relying solely on text-driven cues. This is a departure from models that mimic human emotional responses, which often incorporate physical responses that are irrelevant in a text-only context. As these advanced models interact with users, they would use metrics to refine sentiment analysis, focusing exclusively on language use patterns, semantic structures, and syntactic cues. The practical implications of this methodology are extensive. For instance, in customer service chatbots, this approach would enable a more accurate understanding of customer feedback by analyzing the text for sentiment without misconstruing the absence of physiological indicators as a lack of emotional content. The system would, theoretically, provide a more precise sentiment analysis by recognizing the subtleties in language that indicate sentiment, even when traditional emotional cues are not present.

The advancement of sentiment analysis through emotionally-agnostic text mining in LLMs promises a more refined understanding of textual sentiment. It provides an innovative approach that could significantly enhance the accuracy of sentiment analysis across various applications, from customer service to literary analysis. The recalibration of sentiment analysis methodologies to focus purely on textual data could pave the way for more objective and reliable analyses, free from the constraints and potential biases introduced by human emotional interpretations.

In the quest to emulate the complex modular architecture of human cognition, several key recommendations emerge for enhancing the capabilities of LLMs (Table 1). Foremost, an advanced classification system is warranted to capture the subtleties of emotional states through text-based interaction. Such a system must transcend the current dichotomous sentiment frameworks to discern finely grained emotional nuances, thereby providing a more precise analysis of sentiment from textual communication.

img

Additionally, the current sentiment analysis models must pivot towards non-physiological metrics that eschew reliance on the biological markers of emotion, focusing exclusively on the linguistic elements present in text. This shift requires a reevaluation and recalibration of algorithms to better interpret the subtleties and complexities of language as indicators of sentiment, divorced from the non-verbal cues that are imperceptible in digital communication.

An executive module, an element of the AI architecture analogous to human cognitive oversight functions, must be engineered to manage the coordination and integration of various cognitive modules within LLMs. This executive module would harmonize the disparate aspects of sentiment analysis, ensuring that the output is both coherent and contextually relevant.  Addressing the limitations inherent in the token-based memory of LLMs like ChatGPT, efforts to enhance the model’s ability to reference prior interactions would significantly augment the continuity of conversation and the consistency of sentiment tracking. Such improvements would mimic more closely the human capacity for short-term recall, thus supporting a more authentic dialogue.

 For these systems to remain relevant and accurately attuned to the evolving lexicon of human emotion, an inherent capacity for continuous adaptation and learning is imperative. This dynamic quality would ensure that models like GPT-4 remain sensitive to novel expressions of sentiment and linguistic shifts over time. The ethical deployment of such advanced sentiment analysis tools cannot be overstated. A responsible application of these technologies mandates vigilant oversight to mitigate potential biases and guarantee interactions that uphold the principles of respect and empathy.

Finally, the empirical validation of these advancements is essential. Rigorous testing, including comparative studies with human subjects, should be conducted to establish the efficacy of text-based emotion categorization frameworks. Such research would not only substantiate the theoretical underpinnings of these models but also pave the way for their practical application across various domains. Therefore, the full potential of LLMs for sentiment analysis is yet to be realized. By adopting these recommendations, researchers and practitioners can unlock new dimensions of understanding in the emotional landscape of human-machine interaction, enhancing the granularity with which machines can interpret and respond to human sentiment through text.

Conclusion

The discourse on sentiment analysis within large language models (LLMs) has reached a pivotal juncture with the introduction of emotionally-agnostic text mining techniques. These techniques represent a paradigm shift from conventional sentiment analysis approaches that often incorporate a hybrid of textual and physiological indicators to interpret human emotions. The proposed model, with its innovative reliance solely on textual information, stands to redefine the parameters of sentiment analysis, aligning it more closely with the nature of text-based communication.

The principal takeaway from the current discussion is the identification of the potential that lies in decoupling sentiment analysis from the traditional reliance on human emotion models that include physiological elements. By extracting sentiment purely from text, the proposed model promises to deliver a more nuanced and intricate understanding of textual sentiment, thus avoiding the pitfalls of misinterpretation that often accompany the inclusion of physiological responses.

The benefits of such a text-centric approach to sentiment analysis are multifold. First, it enables a more objective analysis by eliminating the subjectivity inherent in interpreting physiological cues. Second, it facilitates the handling of large datasets where emotional labeling by humans is impractical. Third, it ensures consistency and scalability in sentiment analysis, essential for applications such as market analysis, public opinion mining, and automated customer service.

 Future research should focus on developing a robust framework for implementing emotionally-agnostic sentiment analysis within LLMs. This will likely involve the creation of comprehensive linguistic models that can accurately parse and interpret the emotional valence of text without external emotional labeling. Empirical validation of the model’s efficacy compared to traditional sentiment analysis methods will be imperative.

In terms of implementation, the transition to emotionally-agnostic sentiment analysis will necessitate the refinement of LLMs to accommodate the new framework. Industry and academic partnerships could be pivotal in testing and refining these models in real-world scenarios. Finally, ethical considerations, particularly around the transparency and fairness of algorithmicallyderived sentiment analysis, must guide the deployment of these technologies.

In all, the proposed model of sentiment analysis through emotionally-agnostic text mining augurs a significant advancement in the field of artificial intelligence and computational linguistics. Its successful implementation could revolutionize the way sentiment is gauged and interpreted across vast swathes of text, offering a more reliable, consistent, and scalable solution [42-99].

Data Availability Data available upon request.

Conflicts of Interest

The author(s) declare(s) that there is no conflict of interest regarding the publication of this paper.

Funding Statement

NA

Authors’ Contributions

Conceptualization, J Ratican; Methodology, J Ratican; Validation, J Hutson; Investigation, J Hutson – Original Draft Preparation, J Hutson; Writing – Review & Editing, J Hutson.; Visualization, J Hutson.

References

  1. Gill SS, Kaur R (2023) ChatGPT: Vision and challenges. Internet of Things and Cyber-Physical Systems 3: 262-271.
  2. Albahri AS, Duhaim AM, Fadhel MA, Alnoor A, Baqer NS, et al. (2023) A Systematic Review of Trustworthy and Explainable Artificial Intelligence in Healthcare: Assessment of Quality, Bias Risk, and Data Fusion. Information Fusion.
  3. Lee H (2023) The Rise of ChatGPT: Exploring Its Potential in Medical Education. Anatomical Sciences Education. https:// pubmed.ncbi.nlm.nih.gov/36916887/.
  4. Liu Y, Han T, Ma S, Zhang J, Yang Y, et al. (2023) Summary of ChatGPT/GPT-4 Research and Perspective Towards the Future of Large Language Models 4: 21.
  5. Shahriar S, Hayawi K (2023) Let’s Have a Chat! A Conversation with ChatGPT: Technology, Applications, and Limitations arXiv preprint arXiv:2302.13817.
  6. Garon J (2023) A Practical Introduction to Generative AI, Synthetic Media, and the Messages Found in the Latest Medium. Synthetic Media, and the Messages Found in the Latest Medium. https://papers.ssrn.com/sol3/papers. cfm?abstract_id=4388437.
  7. Ray PP (2023) ChatGPT: A Comprehensive Review on Background, Applications, Key Challenges, Bias, Ethics, Limitations and Future Scope. Internet of Things and CyberPhysical Systems 3: 121-154.
  8. Fodor JA (1983) The Modularity of Mind. MIT Press. https:// mitpress.mit.edu/9780262560252/the-modularity-of-mind/.
  9. Bryson JJ (2005) Modular Representations of Cognitive Phenomena in AI, Psychology and Neuroscience. In Visions of Mind: Architectures for Cognition and Affect IGI Global 66-89.
  10. Park JS, O’Brien JC, Cai CJ, Morris MR, Liang P, et al. (2023) Generative agents: Interactive simulacra of human behavior arXiv preprint arXiv:2304.03442.
  11. Duan Y, Edwards JS, Dwivedi YK (2019) Artificial Intelligence for Decision Making in the Era of Big Data Evolution, Challenges and Research Agenda. International Journal of Information Management 48: 63-71.
  12. Mahmud H, Islam AN, Ahmed SI, Smolander K (2022) What Influences Algorithmic Decision-Making? A Systematic Literature Review on Algorithm Aversion. Technological Forecasting and Social Change 175: 121390.
  13. Zall R, Kangavari MR (2022) Comparative Analytical Survey on Cognitive Agents with Emotional Intelligence. Cognitive Computation 14: 1223-1246.
  14. Samsonovich AV (2020) Socially Emotional Brain-Inspired Cognitive Architecture Framework for Artificial Intelligence. Cognitive Systems Research 60: 57-76.
  15. Picard RW, Papert S, Bender W, Blumberg B, Breazeal C, et al. (2004) Affective Learning A Manifesto. BT Technology Journal 22: 253-269.
  16. Li Y, Jiang Y, Tian D, Hu L, Lu H, Yuan Z (2019) AI-Enabled Emotion Communication. IEEE Network 33: 15-21.
  17. Khachane MY (2017) Organ-Based Medical Image Classification Using Support Vector Machine. International Journal of Synthetic Emotions 8: 18-30.
  18. Pusztahelyi R (2020) Emotional AI and Its Challenges in the Viewpoint of Online Marketing. Curentul Juridic 81: 13-31.
  19. Cominelli L, Hoegen G, De Rossi D (2021) Abel: Integrating Humanoid Body, Emotions, and Time Perception to Investigate Social Interaction and Human Cognition. Applied Sciences 11: 1070.
  20. Wortman B, Wang JZ (2022) HICEM: A High-Coverage Emotion Model for Artificial Emotional Intelligence arXiv preprint arXiv:2206.07593.
  21. Marcos S, García Peñalvo FJ, Vázquez Ingelmo A (2021) Emotional AI in Healthcare: A p ilot architecture proposal to merge emotion recognition tools. In Ninth International Conference on Technological Ecosystems for Enhancing Multiculturality (TEEM’21) 342-349.
  22. Huh JH, Seo YS (2019) Understanding Edge Computing: Engineering Evolution with Artificial Intelligence IEEE Access 7: 164229-164245.
  23. Andersson R (2022) The Bioeconomy and the Birth of a “New Anthropology”. Cultural Anthropology 37: 37-44.
  24. Naddaf Y (2010) Game-Independent AI Agents for Playing Atari 2600 Console Games. University of Alberta, Canada. https://era.library.ualberta.ca/items/a661eb66-f2e0-4ed3b501-b6cbcd1fdd9d.
  25. Liu J, Togelius J, Pérez Liébana D, Lucas SM (2017) Evolving Game Skill-Depth Using General Video Game AI Agents. In 2017 IEEE Congress on Evolutionary Computation (CEC) IEEE 2299-2307.
  26. Holmgard C, Liapis A, Togelius J, Yannakakis GN (2014) Generative Agents for Player Decision Modeling in Games. 9th International Conference on the Foundations of Digital Games, Fort Lauderdale 1-8.
  27. Barthet M, Khalifa A, Liapis A, Yannakakis G (2022) Generative Personas that Behave and Experience Like Humans. In Proceedings of the 17th International Conference on the Foundations of Digital Games 1-10.
  28. Ariyurek S, Betin Can A, Surer E (2019) Automated Video Game Testing Using Synthetic and Humanlike Agents. IEEE Transactions on Games 13: 50-67.
  29. Fathi K, Palhang M (2018) Evaluation of Using Neural Networks on Variety of Agents and Playability of Games. In 2018 International Conference on Artificial Intelligence and Data Processing (IDAP) IEEE 1-8.
  30. Nareyek A (2000) Intelligent Agents for Computer Games. In International Conference on Computers and Games. Berlin, Heidelberg: Springer Berlin Heidelberg 414-422.
  31. Tan SC, Nareyek A (2009) Integrating Facial, Gesture, and Posture Emotion Expression for a 3D Virtual Agent. In Proceedings of the 14th International Conference on Computer Games: AI, Animation, Mobile, Interactive Multimedia, Educational & Serious Games 23-31.
  32. Miikkulainen R, Bryant BD, Cornelius R, Karpov IV, Stanley KO, et al. (2006) Computational Intelligence in Games. Computational Intelligence: Principles and Practice 155-191.
  33. Fernández S, Adarve R, Pérez M, Rybarczyk M, Borrajo D (2006) Planning for an AI Based Virtual Agents Game. In Proceedings of the Workshop AI Planning for Computer Games and Synthetic Characters in the ICAPS. https://www. semanticscholar.org/paper/Planning-for-an-AI-based-virtualagents-game-%E2%88%97-Fern%C2%B4andez-Adarve/77 39dc608abd41726cb4df7c71ed1816fa214023.
  34. Morris MR (2023) Scientists’ Perspectives on the Potential for Generative AI in their Fields arXiv preprint arXiv:2304.01420.
  35. Megahed FM, Chen YJ, Ferris JA, Knoth S, Jones Farmer LA (2023) How Generative AI Models such as ChatGPT can be (Mis) Used in Practice, Education, and Research? An Exploratory Study. Quality Engineering 1-29.
  36. Murphy C, Thomas FP (2023) Generative AI in Spinal Cord Injury Research and Care: Opportunities and Challenges Ahead. The Journal of Spinal Cord Medicine 46: 341-342.
  37. Taherdoost H, Madanchian M (2023) Artificial intelligence and sentiment analysis: A review in competitive research. Computers 12: 37.
  38. Kheiri K, Karimi H (2023) Sentimentgpt: Exploiting GPT for advanced sentiment analysis and its departure from current machine learning arXiv preprint arXiv:2307.10234.
  39. Reddy MJ, Somasekar J, Thopate K, Angadi SK (2023) 17 Exploring artificial intelligence techniques for enhanced sentiment analysis through data mining. Toward Artificial General Intelligence: Deep Learning, Neural Networks, Generative AI 345.
  40. Abirami AM, Gayathri V (2017) A survey on sentiment analysis methods and approach. In 2016 Eighth International Conference on Advanced Computing (ICoAC) IEEE 72-76.
  41. Kapo?i?t? Dzikien? J, Damaševi?ius R, Wo?niak M (2019) Sentiment analysis of lithuanian texts using traditional and deep learning approaches. Computers 8: 4.
  42. Al Kurdi OF, El Haddadeh R, Eldabi T (2020) The Role of Organisational Climate in Managing Knowledge Sharing Among Academics in Higher Education. International Journal of Information Management 50: 217-227.
  43. Biggs J, Tang C, Kennedy G (2022) Ebook: Teaching for Quality Learning at University 5e. McGraw-Hill education (UK). https://www.mheducation.co.uk/teaching-for-qualitylearning-at-university-5e-9780335250820-emea-group.
  44. Brew A, Boud D (1995) Teaching and Research: Establishing the Vital Link with Learning. Higher Education 29: 261-273.
  45. Budhwar P, Malik A, De Silva MT, Thevisuthan P (2022) Artificial Intelligence–Challenges and Opportunities for International HRM: A Review and Research Agenda. The International Journal of Human Resource Management 33: 1065-1097.
  46. Buffington A, Lange C, Bakker C, Nanney M, Roberts W, et al. (2021) The Collaborative Scholarship Intensive: A Research Intensive Course to Improve Faculty Scholarship. Family Medicine 53: 355-358.
  47. Centra JA (1983) Research Productivity and Teaching Effectiveness. Research in Higher Education 18: 379-389.
  48. Coate K, Barnett R, Williams G (2001) Relationships Between Teaching and Research in Higher Education in England. Higher Education Quarterly 55: 158-174.
  49. de Carvalho PAC (2019) Emojar: Collecting and Reliving Memorable and Emotionally Impactful Digital Content (Doctoral dissertation, Universidade de Lisboa (Portugal)). https://repositorio.ul.pt/handle/10451/41531.
  50. Damasio AR, Damasio H (1992) Brain and language.Scientific American 267: 88-109.
  51. Elshentenawy M, Ahmed M, Elalfy M, Bakr A, Heidar M, et al. (2023) Intellibot: A Personalized Behavioral Analysis Chatbot Framework Powered by GPT-3. In 2023 Intelligent Methods, Systems, and Applications (IMSA) IEEE 136-141.
  52. Finnegan DE, Gamson ZF (1996) Disciplinary Adaptations to Research Culture in Comprehensive Institutions. The Review of Higher Education 19: 141-177.
  53. Foster NF, Gibbons S (2005) Understanding Faculty to Improve Content Recruitment for Institutional Repositories. Online Submission 11: 1.
  54. Gazzaniga MS (1983) Right Hemisphere Language Following Brain Bisection: A 20-year Perspective. American Psychologist 38: 525.
  55. He D, Lv X, Zhu S, Chan S, Choo KKR (2023) A Method for Detecting Phishing Websites Based on Tiny-Bert Stacking. IEEE Internet of Things Journal 11.
  56. Hill Y, Lomas L, MacGregor J (2003) Students’ Perceptions of Quality in Higher Education. Quality Assurance in Education 11: 15-20.
  57. Hong S, Zheng X, Chen J, Cheng Y, Zhang C, et al. (2023) Metagpt: Meta programming for multi-agent collaborative framework arXiv preprint arXiv:2308.00352.
  58. Jacelon CS, Zucker DM, Staccarini JM, Henneman EA (2003) Peer Mentoring for Tenure-Track Faculty. Journal of Professional Nursing 19: 335-338.
  59. Jamali A, Bhutto A, Khaskhely M, Sethar W (2022) Impact of Leadership Styles on Faculty Performance: Moderating Role of Organizational Culture in Higher Education. Management Science Letters 12: 1-20.
  60. Jayman M, Glazzard J, Rose A (2022) Tipping Point: The Staff Wellbeing Crisis in Higher Education. In Frontiers in Education. Frontiers 7: 929335.
  61. Jeste DV, Graham SA, Nguyen TT, Depp CA, Lee EE, Kim HC (2020) Beyond Artificial Intelligence: Exploring Artificial Wisdom. International Psychogeriatrics 32: 993-1001.
  62. Khezr P, Mohan V (2022) The Vexing but Persistent Problem of Authorship Misconduct in Research. Research Policy 51: 104466.
  63. Koch A, Nafziger J, Nielsen HS (2015) Behavioral Economics of Education. Journal of Economic Behavior & Organization 115: 3-17.
  64. Kulage KM, Larson EL, Begg MD (2011) Sharing Facilities and Administrative (F&A) Cost Recovery to Facilitate Interdisciplinary Research. Academic Medicine: Journal of the Association of American Medical Colleges 86: 394.
  65. Latif E, Mai G, Nyaaba M, Wu X, Liu N, et al. (2023) Artificial General Intelligence (AGI) for Education arXiv preprint arXiv:2304.12479.
  66. Lee H (2023) The Rise of ChatGPT: Exploring Its Potential in Medical Education. Anatomical Sciences Education. https:// pubmed.ncbi.nlm.nih.gov/36916887/.
  67. Liu Thompkins Y, Okazaki S, Li H (2022) Artificial Empathy in Marketing Interactions: Bridging the Human AI Gap in Affective and Social Customer Experience. Journal of the Academy of Marketing Science 50: 1198-1218.
  68. Mahmud H, Islam AN, Ahmed SI, Smolander K (2022) What Influences Algorithmic Decision-Making? A Systematic Literature Review on Algorithm Aversion. Technological Forecasting and Social Change 175: 121390.
  69. Matos JF, Piedade J, Freitas A, Pedro N, Dorotea N, et al. (2023) Teaching and Learning Research Methodologies in Education: A Systematic Literature Review. Education Sciences 13: 173.
  70. Maxwell LD, Ball AL (2010) What is Scholarship? Faculty Knowledge and Perceptions of the Scholarship of Teaching and Learning. Journal of Agricultural Education 51: 127.
  71. McKeown JS, Bista K, Chan RY (2022) Global Higher Education During COVID-19: Policy, Society, and Technology. STAR Scholars. https://starscholars.org/product/ global-education/.
  72. McNair TB, Albertine S, McDonald N, Major Jr T, Cooper MA (2022) Becoming a Student-Ready College: A New Culture of Leadership for Student Success. John Wiley & Sons 240.
  73. Meilani YFCP, Tan JD, Murwani FD, Bernarto I, Sudibjo N (2021) Motivating and Retaining Generation Z Faculty Members in Private Universities. Journal of Educational and Social Research 11: 245-255.
  74. Mokiy V, Lukyanova T (2022) Development of Interdisciplinarity and Transdisciplinarity in Modern Russian Science and Higher Education. In Institutionalizing Interdisciplinarity and Transdisciplinarity. Routledge 124138.
  75. Moody J (2013) Faculty Diversity: Removing the Barriers. Routledge. https://www.routledge.com/Faculty-DiversityRemoving-the-Barriers/Moody/p/book/9780415878463.
  76. Norman R, Flaugher T, Chang S, Power E (2023) SelfPerception of Cognitive-Communication Functions after Mild Traumatic Brain Injury. American Journal of Speech Language Pathology 32: 883-906.
  77. Norton Meier L, Hand B, Cavagnetto A, Akkus R, Gunel M (2009) Pedagogy, Implementation, and Professional Development for Teaching Science Literacy: How Students and Teachers Know and Learn. Quality Research in Literacy and Science Education: International Perspectives and Gold Standards 169-187.
  78. Omar MS, Idrus IM, Jamal NA (2021) The Influence of Job Motivation on Job Satisfaction: A Case Study of Polytechnic Academic Staff. Malaysian Journal of Social Sciences and Humanities (MJSSH) 6: 206-213.
  79. Power MJ (2006) The Structure of Emotion: An Empirical Comparison of Six Models. Cognition & Emotion 20: 694713.
  80. Prince MJ, Felder RM, Brent R (2007) Does Faculty Research Improve Undergraduate Teaching? An Analysis of Existing and Potential Synergies. Journal of Engineering Education 96: 283-294.
  81. Rana V, Garg V, Mago Y, Bhat A (2023) Compact BERTBased Multi-Models for Efficient Fake News Detection. In 2023 3rd International Conference on Intelligent Technologies (CONIT) IEEE 1-4.
  82. Ratican J, Hutson J (2023) The Six Emotional Dimension (6DE) Model: A Multidimensional Approach to Analyzing Human Emotions and Unlocking the Potential of Emotionally Intelligent Artificial Intelligence (AI) via Large Language Models (LLM). Journal of Artificial Intelligence and Robotics 1.
  83. Sadiq Sohail M, Daud S (2009) Knowledge Sharing in Higher Education Institutions: Perspectives from Malaysia. Vine 39: 125-142.
  84. Santo SA, Engstrom ME, Reetz L, Schweinle WE, Reed K (2009) Faculty Productivity Barriers and Supports at a School of Education. Innovative Higher Education 34: 117-129.
  85. Sato M, Loewen S (2019) Do Teachers Care About Research? The Research Pedagogy Dialogue. ELT Journal 73: 1-10.
  86. Selvaraj C, Chandra I, Singh SK (2021) Artificial Intelligence and Machine Learning Approaches for Drug Design:Challenges and Opportunities for the Pharmaceutical Industries. Molecular Diversity 1-21.
  87. Shavit Y (2007) Stratification in Higher Education: A Comparative Study. Stanford University Press. https://www.sup.org/books/title/?id=8735.
  88. Singh A, Chouhan T (2023) Artificial Intelligence in HRM: Role of Emotional–Social Intelligence and Future Work Skill. In The Adoption and Effect of Artificial Intelligence on Human Resources Management, Part A Emerald Publishing Limited 175-196.
  89. Sohrabi C, Mathew G, Franchi T, Kerwan A, Griffin M, et al. (2021) Impact of the Coronavirus (COVID-19) Pandemic on Scientific Research and Implications for Clinical Academic Training A Review. International Journal of Surgery 86: 5763.
  90. Strich F, Mayer AS, Fiedler M (2021) What Do I Do in a World of Artificial Intelligence? Investigating the Impact of Substitutive Decision-Making AI Systems on Employees’ Professional Role Identity. Journal of the Association for Information Systems 22: 9.
  91. Tavares O, Sin C, Sá C, Bugla S, Amaral A (2022) Inbreeding and Research Collaborations in Portuguese Higher Education. Higher Education Quarterly 76: 102-115.
  92. Terenzini PT, Pascarella ET (1998) Studying College Students in the 21st century: Meeting New Challenges. The Review of Higher Education 21: 151-165.
  93. Trott CD, Sample McMeeking LB, Weinberg AE (2020) Participatory Action Research Experiences for Undergraduates: Forging Critical Connections Through Community Engagement. Studies in Higher Education 45: 2260-2273.
  94. Väljataga T, Fiedler SH, Laanpere M (2015) Re Thinking Digital Textbooks: Students as Co-Authors. In Advances in Web-Based Learning--ICWL 2015: 14th International Conference, Guangzhou, China, November 5-8, Proceedings Springer International Publishing 14: 143-151.
  95. Werth E, Williams K (2021) Exploring Student Perceptions as Co-Authors of Course Material. Open Praxis 13: 53-67.
  96. Williams R, Andreassi S, Moselli M, Fantini F, Tanzilli A, et al. (2023) Relationship between Executive Functions, Social Cognition, and Attachment State of Mind in Adolescence: An Explorative Study. International Journal of Environmental Research and Public Health 20: 28-36.
  97. Xi Z, Chen W, Guo X, He W, Ding Y, et al. (2023) The Rise and Potential of Large Language Model Based Agents: A Survey arXiv preprint arXiv:2309.07864.
  98. Zhang L, Carter Jr RA, Qian X, Yang S, Rujimora J, et al. (2022) Academia’s Responses to Crisis: A Bibliometric Analysis of Literature on Online Learning in Higher Education During COVID-19. British Journal of Educational Technology 53: 620-646.
  99. Zhu Y (2023) The Summarization of Split Brain Study of Last Few Decades. Highlights in Science, Engineering and Technology 30: 117-122.
View PDF