The final, formatted version of the article will be published soon.
TECHNOLOGY AND CODE article
Front. Psychol.
Sec. Quantitative Psychology and Measurement
Volume 16 - 2025 |
doi: 10.3389/fpsyg.2025.1488102
This article is part of the Research Topic Data Science and Machine Learning for Psychological Research View all 5 articles
Leveraging on Large Language Model (LLM) to classify sentences: a case study applying STAGES scoring methodology for sentence completion test on Ego Development
Provisionally accepted- Integral Transpersonal Institute, Milan, Italy
The emergence of Artificial Intelligence and widespread availability of Large Language Model open the door to text analysis at scale leveraging on complex classification instructions. This case study explores the usage of Large Language Models to measure Ego Development at scale and establish a methodology that can be applied to other classification instructions. Ego consists of the traits that influence how a person perceives and engages with the world, while Ego development is a crucial aspect of adult personality growth, influencing behaviors and decisions in both personal and professional contexts. This case study investigates the agreement between expert and automated classifications of Ego Development STAGES, aiming to evaluate the potential of automation in this domain leveraging on Large Language Models. Cohen's Kappa statistic has been used to measure the agreement between classifications made by experts and the automated classiification. The inter-rater agreement yielded a weighted Kappa value of 0.779, indicating a substantial level of agreement that is statistically meaningful and unlikely to be due to chance. Notably, we observed low variability in aggregated values, demonstrating that the automated process functions effectively at scale. The robustness in aggregated data is particularly evident when calculating ego development scores for individuals, groups, corporate units, and entire corporations. This capability underscores the utility of the automated system for high-level evaluations and decision-making leveraging on a solid indicator.Our findings emphasize the importance of continuous improvement in the automated classification process through algorithm refinement. Future research should address this area and focus on enhancing the contextual adaptability of the system to better capture nuanced judgments. The rapid evolution of Large Language Models may support the refinement process. While the classification system developed in this case study shows promise, targeted enhancements may help to achieve a level of accuracy and reliability that aligns with expert evaluations for single sentences. The methodology used in this case study appears to be useful to support other evaluation at scale that leverage on Large Language Models using other classifications' maps.
Keywords: Large Language Model, automated classification, Ego development, Cohen's kappa, methods
Received: 29 Aug 2024; Accepted: 10 Jan 2025.
Copyright: © 2025 Bronlet. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
* Correspondence:
Xavier Bronlet, Integral Transpersonal Institute, Milan, Italy
Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.