An Experiment of AI-Based Assessment: Perspectives of Learning Preferences, Benefits, Intention, Technology Affinity, and Trust
<p>Diagram of regression models and related hypotheses. Four separate regression models were fitted with outcomes preferring AI over teacher, intention to use AI tools, perceived benefits, and trust in AI-based essay tool. Variables marked with red were associated with the essay-grading tool.</p> "> Figure 2
<p>Marginal means for variables (<b>a</b>) preferring AI over teacher, and (<b>b</b>) preferred study style and preferred study location. Error bars indicate HDI 95%.</p> "> Figure 3
<p>Marginal means for variable preferring AI over teacher. Error bars indicate HDI 95%.</p> "> Figure 4
<p>Marginal means for variables (<b>a</b>) preferring AI over teacher, and (<b>b</b>) preferred study style and preferred study location. Error bars indicate HDI 95%.</p> ">
Abstract
:1. Introduction
2. Related Work
2.1. AI-Based and Automated Essay Scoring Systems
2.2. Individual Characteristics and Technology Affinity of Students
3. Theoretical Framework and Hypotheses
3.1. Demographic Influence
3.1.1. Age and Trust in AI-Based Essay Grading
3.1.2. Gender and Trust in AI-Based Essay Grading
3.1.3. Demographics and Perceived Benefits of AI-Based Essay Grading
3.1.4. Demographics and Behavioral Intention to Use AI-Based Educational Tools
3.1.5. Educational Programme and Trust in the AI-Based Essay Grading Tool
3.2. Study Preferences
3.3. Preferences Regarding Teacher-Led vs. AI-Led Instruction
3.4. Affinity for Technology Interaction (ATI)
3.5. The Relationship of Trust, Perceived Benefits, and Intention to Use AI
4. Methods and Materials
4.1. Participants
4.2. AI-Based Essay Scoring Setup
4.3. Procedure of Research Experiment
- Please read Section 5 of this article (note: please open it in a separate browser tab) https://hal.inria.fr/hal-02191186/document (accessed on 11 December 2024).
- Please answer the following question: “In which different ways may big data and artificial intelligence (e.g., cognitive systems) change businesses and industries?”
- Please answer this question by writing a short essay (50–100 words). Start your answering process by clicking on “Open answer form”. On submission of your answer, you will receive feedback and a respective assessment score from an AI grading tool. Note: The feedback and score have NO IMPACT on any of your grades or assessments in class! [Open answer form]
- After having received your feedback and score from the AI tool, please complete the following survey: [Link to questionnaire]
4.4. Design of the Questionnaire
4.5. Methods
5. Results
5.1. Factor Constructs for Trust, Benefit, Intention, and Interaction Depth
5.2. Regression Analysis
5.2.1. Preference of AI over Teacher
for technology interaction
5.2.2. Benefits of Using an AI-Based Essay Grading Tool
study location + preferring AI over teacher + trust in AI applications
+ affinity for technoloyg interaction
5.2.3. Trust for AI-Based Essay Grading Tool
in AI applications + affinity for technology interaction
5.2.4. Behavioral Intention to Use AI Tools in Education
program + preferred study style × preferred study location + preferring AI over
teacher + trust in AI applications + affinity for technology interaction
6. Discussion
7. Conclusions and Future Work
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Chiu, T.K.F.; Xia, Q.; Zhou, X.; Chai, C.S.; Cheng, M. Systematic Literature Review on Opportunities, Challenges, and Future Research Recommendations of Artificial Intelligence in Education. Comput. Educ. Artif. Intell. 2023, 4, 100118. [Google Scholar] [CrossRef]
- Zhai, X.; Chu, X.; Chai, C.S.; Jong, M.S.Y.; Istenic, A.; Spector, M.; Liu, J.-B.; Yuan, J.; Li, Y. A Review of Artificial Intelligence (AI) in Education from 2010 to 2020. Complexity 2021, 2021, 8812542. [Google Scholar] [CrossRef]
- Nemorin, S.; Vlachidis, A.; Ayerakwa, H.M.; Andriotis, P. AI Hyped? A Horizon Scan of Discourse on Artificial Intelligence in Education (AIED) and Development. Learn. Media Technol. 2023, 48, 38–51. [Google Scholar] [CrossRef]
- Delcker, J.; Heil, J.; Ifenthaler, D.; Seufert, S.; Spirgi, L. First-Year Students’ AI Competence as a Predictor for Intended and de Facto Use of AI Tools for Supporting Learning Processes in Higher Education. Int. J. Educ. Technol. High. Educ. 2024, 21, 18. [Google Scholar] [CrossRef]
- Gao, R.; Merzdorf, H.E.; Anwar, S.; Hipwell, M.C.; Srinivasa, A. Automatic Assessment of Text-Based Responses in Post-Secondary Education: A Systematic Review. Comput. Educ. Artif. Intell. 2024, 6, 100206. [Google Scholar] [CrossRef]
- Ramesh, D.; Sanampudi, S.K. An Automated Essay Scoring Systems: A Systematic Literature Review. Artif. Intell. Rev. 2022, 55, 2495–2527. [Google Scholar] [CrossRef]
- Minn, S. AI-Assisted Knowledge Assessment Techniques for Adaptive Learning Environments. Comput. Educ. Artif. Intell. 2022, 3, 100050. [Google Scholar] [CrossRef]
- Mononen, A.; Alamäki, A.; Kauttonen, J.; Klemetti, A.; Passi-Rauste, A.; Ketamo, H. Forecasted Self: AI-Based Careerbot-Service Helping Students with Job Market Dynamics. Eng. Proc. 2023, 39, 99. [Google Scholar] [CrossRef]
- Stanja, J.; Gritz, W.; Krugel, J.; Hoppe, A.; Dannemann, S. Formative Assessment Strategies for Students’ Conceptions—The Potential of Learning Analytics. Br. J. Educ. Technol. 2023, 54, 58–75. [Google Scholar] [CrossRef]
- Chai, C.S.; Wang, X.; Xu, C. An Extended Theory of Planned Behavior for the Modelling of Chinese Secondary School Students’ Intention to Learn Artificial Intelligence. Mathematics 2020, 8, 2089. [Google Scholar] [CrossRef]
- Bakhadirov, M.; Alasgarova, R. Factors Influencing Teachers’ Use of Artificial Intelligence for Instructional Purposes. IAFOR J. Educ. 2024, 12, 9–32. [Google Scholar] [CrossRef]
- Viberg, O.; Cukurova, M.; Feldman-Maggor, Y.; Alexandron, G.; Shirai, S.; Kanemune, S.; Wasson, B.; Tømte, C.; Spikol, D.; Milrad, M.; et al. What Explains Teachers’ Trust in AI in Education across Six Countries? Int. J. Artif. Intell. Educ. 2024. [Google Scholar] [CrossRef]
- Almaiah, M.A.; Alfaisal, R.; Salloum, S.A.; Hajjej, F.; Thabit, S.; El-Qirem, F.A.; Lutfi, A.; Alrawad, M.; Al Mulhem, A.; Alkhdour, T.; et al. Examining the Impact of Artificial Intelligence and Social and Computer Anxiety in E-Learning Settings: Students’ Perceptions at the University Level. Electronics 2022, 11, 3662. [Google Scholar] [CrossRef]
- Ke, Z.; Inamdar, H.; Lin, H.; Ng, V. Give Me More Feedback II: Annotating Thesis Strength and Related Attributes in Student Essays. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, Florence, Italy, 28 July–2 August 2019; pp. 3994–4004. [Google Scholar]
- Gikandi, J.W.; Morrow, D.; Davis, N.E. Online Formative Assessment in Higher Education: A Review of the Literature. Comput. Educ. 2011, 57, 2333–2351. [Google Scholar] [CrossRef]
- Wiliam, D.; Black, P. Meanings and Consequences: A Basis for Distinguishing Formative and Summative Functions of Assessment? Br. Educ. Res. J. 1996, 22, 537–548. [Google Scholar] [CrossRef]
- Dirin, A.; Alamäki, A.; Suomala, J. Gender Differences in Perceptions of Conventional Video, Virtual Reality and Augmented Reality. 2019. Available online: https://www.learntechlib.org/p/216491/ (accessed on 11 December 2024).
- Yang, A.C.M.; Chen, I.Y.L.; Flanagan, B.; Ogata, H. How Students’ Self-Assessment Behavior Affects Their Online Learning Performance. Comput. Educ. Artif. Intell. 2022, 3, 100058. [Google Scholar] [CrossRef]
- Clark, R.E. Media Will Never Influence Learning. Educ. Technol. Res. Dev. 1994, 42, 21–29. [Google Scholar] [CrossRef]
- Alamäki, A.; Nyberg, C.; Kimberley, A.; Salonen, A.O. Artificial Intelligence Literacy in Sustainable Development: A Learning Experiment in Higher Education. Front. Educ. 2024, 9, 1343406. [Google Scholar] [CrossRef]
- Kamii, C. The Equilibration of Cognitive Structures: The Central Problem of Intellectual Development. Am. J. Educ. 1986, 94, 574–577. [Google Scholar] [CrossRef]
- Vygotsky, L.S.; Cole, M. Mind in Society: Development of Higher Psychological Processes; Harvard University Press: Cambridge, MA, USA, 1978. [Google Scholar]
- Kaya, F.; Aydin, F.; Schepman, A.; Rodway, P.; Yetişensoy, O.; Demir Kaya, M. The Roles of Personality Traits, AI Anxiety, and Demographic Factors in Attitudes toward Artificial Intelligence. Int. J. Hum. Comput. Interact. 2024, 40, 497–514. [Google Scholar] [CrossRef]
- Rogers, E.M. Diffusion of Innovations, 5th ed.; Free Press: New York, NY, USA, 2003. [Google Scholar]
- Dunnhumby Understanding Consumer Trust in Artificial Intelligence 2023. Available online: https://www.dunnhumby.com/ (accessed on 11 December 2024).
- Gillespie, N.; Lockey, S.; Curtis, C.; Pool, J.; Akbari, A. Trust in Artificial Intelligence: A Global Study; The University of Queensland: Brisbane, Australia; KPMG Australia: Sydney, Australia, 2023. [Google Scholar]
- Bem, S.L. Gender Schema Theory: A Cognitive Account of Sex Typing. Psychol. Rev. 1981, 88, 354–364. [Google Scholar] [CrossRef]
- Lebow, S. Men More Likely than Women to Trust Generative AI. Morning Consult. 2023. Available online: https://www.insiderintelligence.com/content/men-more-likely-than-women-trust-generative-ai (accessed on 11 December 2024).
- Mearian, L. With AI, There’s a Trust Gap Based on Gender, Age. Computerworld. 2023. Available online: https://www.computerworld.com/article/3707795/ai-trust-gap-based-on-gender-age.html (accessed on 11 December 2024).
- Gefen, D.; Straub, D.W. Gender Differences in the Perception and Use of E-Mail: An Extension to the Technology Acceptance Model. MIS Q. 1997, 21, 389–400. [Google Scholar] [CrossRef]
- McKnight, D.H.; Carter, M.; Thatcher, J.B.; Clay, P.F. Trust in a Specific Technology: An Investigation of Its Components and Measures. ACM Trans. Manag. Inf. Syst. TMIS 2011, 2, 1–25. [Google Scholar] [CrossRef]
- Omrani, N.; Rivieccio, G.; Fiore, U.; Schiavone, F.; Garcia Agreda, S. To Trust or Not to Trust? An Assessment of Trust in AI-Based Systems: Concerns, Ethics and Contexts. Technol. Forecast. Soc. Chang. 2022, 181, 121763. [Google Scholar] [CrossRef]
- Castillo-Acobo, R.; Tiza, D.; Orellana, M.; Cajigas, B.; Huayta-Meza, F.; Sota, C.; Muñoz, I.; Acevedo, J.; Sernaqué, A.; Carranza, C.; et al. Artificial Intelligence Application in Education. J. Namib. Stud. Hist. Politics Cult. 2023, 33, 792–807. [Google Scholar] [CrossRef]
- Davis, F.D. Perceived Usefulness, Perceived Ease of Use, and User Acceptance of Information Technology. MIS Q. 1989, 13, 319–340. [Google Scholar] [CrossRef]
- Hornberger, M.; Bewersdorff, A.; Nerdel, C. What Do University Students Know about Artificial Intelligence? Development and Validation of an AI Literacy Test. Comput. Educ. Artif. Intell. 2023, 5, 100165. [Google Scholar] [CrossRef]
- Marrone, R.; Taddeo, V.; Hill, G. Creativity and Artificial Intelligence—A Student Perspective. J. Intell. 2022, 10, 65. [Google Scholar] [CrossRef] [PubMed]
- Long, D.; Magerko, B. What Is AI Literacy? Competencies and Design Considerations. In Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems, Honolulu, HI, USA, 25–30 April 2020; pp. 1–16. [Google Scholar]
- Venkatesh, V.; Morris, M.G.; Davis, G.B.; Davis, F.D. User Acceptance of Information Technology: Toward a Unified View. MIS Q. 2003, 27, 425–478. [Google Scholar] [CrossRef]
- Jia, K.; Wang, P.; Li, Y.; Chen, Z.; Jiang, X.; Lin, C.-L.; Chin, T. Research Landscape of Artificial Intelligence and E-Learning: A Bibliometric Research. Front. Psychol. 2022, 13, 795039. [Google Scholar] [CrossRef] [PubMed]
- Seo, K.; Tang, J.; Roll, I.; Fels, S.; Yoon, D. The Impact of Artificial Intelligence on Learner--Instructor Interaction in Online Learning. Int. J. Educ. Technol. High. Educ. 2021, 18, 54. [Google Scholar] [CrossRef] [PubMed]
- Jin, S.-H.; Im, K.; Yoo, M.; Roll, I.; Seo, K. Supporting Students’ Self-Regulated Learning in Online Learning Using Artificial Intelligence Applications. Int. J. Educ. Technol. High. Educ. 2023, 20, 37. [Google Scholar] [CrossRef]
- Zimmerman, B.J. Becoming a Self-Regulated Learner: An Overview. Theory Pract. 2002, 41, 64–70. [Google Scholar] [CrossRef]
- Wei, X.; Sun, S.; Wu, D.; Zhou, L. Personalized Online Learning Resource Recommendation Based on Artificial Intelligence and Educational Psychology. Front. Psychol. 2021, 12, 767837. [Google Scholar] [CrossRef] [PubMed]
- Colonna, L. Teachers in the Loop? An Analysis of Automatic Assessment Systems under Article 22 GDPR. Int. Data Priv. Law 2023, 14, 3–18. [Google Scholar] [CrossRef]
- Escalante, J.; Pack, A.; Barrett, A. AI-Generated Feedback on Writing: Insights into Efficacy and ENL Student Preference. Int. J. Educ. Technol. High. Educ. 2023, 20, 57. [Google Scholar] [CrossRef]
- Celik, I.; Dindar, M.; Muukkonen, H.; Järvelä, S. The Promises and Challenges of Artificial Intelligence for Teachers: A Systematic Review of Research. TechTrends 2022, 66, 616–630. [Google Scholar] [CrossRef]
- Stoica, E. A Student’s Take on Challenges of AI-Driven Grading in Higher Education. Bachelor’s Thesis, University of Twente, Enschede, The Netherlands, 2022. [Google Scholar]
- Nazaretsky, T.; Cukurova, M.; Alexandron, G. An Instrument for Measuring Teachers’ Trust in AI-Based Educational Technology. In Proceedings of the LAK22: 12th International Learning Analytics and Knowledge Conference, Online, 21–25 March 2021; pp. 1–6. [Google Scholar]
- Nazaretsky, T.; Ariely, M.; Cukurova, M.; Alexandron, G. Teachers’ Trust in AI-Powered Educational Technology and a Professional Development Program to Improve It. Br. J. Educ. Technol. 2022, 53, 914–931. [Google Scholar] [CrossRef]
- Nazaretsky, T.; Cukurova, M.; Ariely, M.; Alexandron, G. Confirmation Bias and Trust: Human Factors That Influence Teachers’ Attitudes Towards AI-Based Educational Technology. In Proceedings of the CEUR Workshop Proceedings, Uzhhorod, Ukrain, 28–30 September 2021; Volume 3042. [Google Scholar]
- Franke, T.; Attig, C.; Wessel, D. A Personal Resource for Technology Interaction: Development and Validation of the Affinity for Technology Interaction (ATI) Scale. Int. J. Hum. Comput. Interact. 2019, 35, 456–467. [Google Scholar] [CrossRef]
- Brauner, P.; Hick, A.; Philipsen, R.; Ziefle, M. What Does the Public Think about Artificial Intelligence?—A Criticality Map to Understand Bias in the Public Perception of AI. Front. Comput. Sci. 2023, 5, 1113903. [Google Scholar] [CrossRef]
- Kahr, P.K.; Rooks, G.; Willemsen, M.C.; Snijders, C.C.P. It Seems Smart, but It Acts Stupid: Development of Trust in AI Advice in a Repeated Legal Decision-Making Task. In Proceedings of the 28th International Conference on Intelligent User Interfaces, Sydney, NSW, Australia, 27–31 March 2023; pp. 528–539. [Google Scholar]
- Schadelbauer, L.; Schlögl, S.; Groth, A. Linking Personality and Trust in Intelligent Virtual Assistants. Multimodal Technol. Interact. 2023, 7, 54. [Google Scholar] [CrossRef]
- Buck, C.; Doctor, E.; Hennrich, J.; Jöhnk, J.; Eymann, T. General Practitioners’ Attitudes toward Artificial Intelligence–Enabled Systems: Interview Study. J. Med. Internet Res. 2022, 24, e28916. [Google Scholar] [CrossRef]
- Chen, X.; Zou, D.; Xie, H.; Cheng, G.; Liu, C. Two Decades of Artificial Intelligence in Education. Educ. Technol. Soc. 2022, 25, 28–47. [Google Scholar]
- Mäki, M.; Kauttonen, J.; Alamäki, A. How Students’ Information Sensitivity, Privacy Trade-Offs, and Stages of Customer Journey Affect Consent to Utilize Personal Data. 2023. Available online: https://www.theseus.fi/handle/10024/793210 (accessed on 11 December 2024).
- Nguyen, A.; Ngo, H.N.; Hong, Y.; Dang, B.; Nguyen, B.-P.T. Ethical Principles for Artificial Intelligence in Education. Educ. Inf. Technol. Dordr 2023, 28, 4221–4241. [Google Scholar] [CrossRef] [PubMed]
- Roy, R.; Babakerkhell, M.D.; Mukherjee, S.; Pal, D.; Funilkul, S. Evaluating the Intention for the Adoption of Artificial Intelligence-Based Robots in the University to Educate the Students. IEEE Access 2022, 10, 125666–125678. [Google Scholar] [CrossRef]
- Bilquise, G.; Ibrahim, S.; Salhieh, S.M. Investigating Student Acceptance of an Academic Advising Chatbot in Higher Education Institutions. Educ. Inf. Technol. Dordr 2023, 29, 6357–6382. [Google Scholar] [CrossRef]
- Alamäki, A. A Conceptual Model for Knowledge Dimensions and Processes in Design and Technology Projects. Int. J. Technol. Des. Educ. 2018, 28, 667–683. [Google Scholar] [CrossRef]
- Garrison, D.R.; Kanuka, H. Blended Learning: Uncovering Its Transformative Potential in Higher Education. Internet High. Educ. 2004, 7, 95–105. [Google Scholar] [CrossRef]
- Zilka, G.C.; Cohen, R.; Rahimi, I. Teacher Presence and Social Presence in Virtual and Blended Courses. J. Inf. Technol. Educ. Res. 2018, 17, 103. [Google Scholar]
- Lan, Y.-J. Immersion, Interaction, and Experience-Oriented Learning: Bringing Virtual Reality into FL Learning. Lang. Learn. Technol. 2020, 24, 1–15. [Google Scholar]
- Heath, R.; Brandt, D.; Nairn, A. Brand Relationships: Strengthened by Emotion, Weakened by Attention. J. Advert. Res. 2006, 46, 410–419. [Google Scholar] [CrossRef]
- Song, Y.; Dai, X.-Y.; Wang, J. Not All Emotions Are Created Equal: Expressive Behavior of the Networked Public on China’s Social Media Site. Comput. Hum. Behav. 2016, 60, 525–533. [Google Scholar] [CrossRef]
- Berger, J.; Milkman, K.L. What Makes Online Content Viral? J. Mark. Res. 2012, 49, 192–205. [Google Scholar] [CrossRef]
- Lazarus, R.S. Cognition and Motivation in Emotion. Am. Psychol. 1991, 46, 352. [Google Scholar] [CrossRef] [PubMed]
- Sheeran, P.; Webb, T.L. The Intention--Behavior Gap. Soc. Pers. Psychol. Compass 2016, 10, 503–518. [Google Scholar] [CrossRef]
- Alamäki, A.; Pesonen, J.; Dirin, A. Triggering Effects of Mobile Video Marketing in Nature Tourism: Media Richness Perspective. Inf. Process. Manag. 2019, 56, 756–770. [Google Scholar] [CrossRef]
- Kavanagh, S.; Luxton-Reilly, A.; Wuensche, B.; Plimmer, B. A Systematic Review of Virtual Reality in Education. Themes Sci. Technol. Educ. 2017, 10, 85–119. [Google Scholar]
- Mikropoulos, T.A.; Natsis, A. Educational Virtual Environments: A Ten-Year Review of Empirical Research (1999–2009). Comput. Educ. 2011, 56, 769–780. [Google Scholar] [CrossRef]
- Hall, E.; Seyam, M.; Dunlap, D. Identifying Usability Challenges in AI-Based Essay Grading Tools. In Proceedings of the International Conference on Artificial Intelligence in Education, Tokyo, Japan, 3–7 July 2023; pp. 675–680. [Google Scholar]
- Valkokari, K.; Rantala, T.; Alamäki, A.; Palomäki, K. Business Impacts of Technology Disruption-a Design Science Approach to Cognitive Systems’ Adoption within Collaborative Networks. In Proceedings of the Collaborative Networks of Cognitive Systems: 19th IFIP WG 5.5 Working Conference on Virtual Enterprises, PRO-VE 2018, Proceedings 19, Cardiff, UK, 17–19 September 2018; Springer International Publishing: Cham, Switzerland, 2018; pp. 337–348. [Google Scholar]
- Sung, E.; Mayer, R.E. Affective Impact of Navigational and Signaling Aids to E-Learning. Comput. Hum. Behav. 2012, 28, 473–483. [Google Scholar] [CrossRef]
- Cheung, M.F.Y.; To, W.-M. The Influence of the Propensity to Trust on Mobile Users’ Attitudes toward in-App Advertisements: An Extension of the Theory of Planned Behavior. Comput. Hum. Behav. 2017, 76, 102–111. [Google Scholar] [CrossRef]
- Shroff, R.H.; Deneen, C.C.; Ng, E.M.W. Analysis of the Technology Acceptance Model in Examining Students’ Behavioural Intention to Use an e-Portfolio System. Australas. J. Educ. Technol. 2011, 27. [Google Scholar] [CrossRef]
- Pavlou, P.A.; Gefen, D. Building Effective Online Marketplaces with Institution-Based Trust. Inf. Syst. Res. 2004, 15, 37–59. [Google Scholar] [CrossRef]
- Rosseel, Y. lavaan: An R Package for Structural Equation Modeling. J. Stat. Softw. 2012, 48, 1–36. Available online: http://www.jstatsoft.org/v48/i02/ (accessed on 11 December 2024). [CrossRef]
- MacCallum, R.C.; Widaman, K.F.; Zhang, S.; Hong, S. Sample Size in Factor Analysis. Psychol. Methods 1999, 4, 84. [Google Scholar] [CrossRef]
- Worthington, R.L.; Whittaker, T.A. Scale Development Research: A Content Analysis and Recommendations for Best Practices. Couns. Psychol. 2006, 34, 806–838. [Google Scholar] [CrossRef]
- Mundfrom, D.J.; Shaw, D.G.; Ke, T.L. Minimum Sample Size Recommendations for Conducting Factor Analyses. Int. J. Test. 2005, 5, 159–168. [Google Scholar] [CrossRef]
- Bürkner, P.-C.; Charpentier, E. Modelling Monotonic Effects of Ordinal Predictors in Bayesian Regression Models. Br. J. Math. Stat. Psychol. 2020, 73, 420–451. [Google Scholar] [CrossRef] [PubMed]
- Silva, L.A.; Zanella, G. Robust Leave-One-out Cross-Validation for High-Dimensional Bayesian Models. J. Am. Stat. Assoc. 2024, 119, 2369–2381. [Google Scholar] [CrossRef]
- Vehtari, A.; Gelman, A.; Gabry, J. Practical Bayesian Model Evaluation Using Leave-One-out Cross-Validation and WAIC. Stat. Comput. 2017, 27, 1413–1432. [Google Scholar] [CrossRef]
- Bürkner, P.-C.; Vuorre, M. Ordinal Regression Models in Psychology: A Tutorial. Adv. Methods Pract. Psychol. Sci. 2019, 2, 77–101. [Google Scholar] [CrossRef]
- Conijn, R.; Kahr, P.; Snijders, C. The Effects of Explanations in Automated Essay Scoring Systems on Student Trust and Motivation. J. Learn. Anal. 2023, 10, 37–53. [Google Scholar] [CrossRef]
No. | Group | Data Type | Scale | Questions | Reference |
---|---|---|---|---|---|
1. | Demographic | Categorical | - | Age: 1 = Under 20, 2 = 21–30, 3 = 31–40, 4 = over 40 Gender: 1 = Female, 2 = Male, 3 = Other or skipped Educational Programme: 1 = IT or Digital, 2 = Business, 3 = Hospitality or Tourism, 4 = Others | Self-developed |
2. | Study preferences | Categorical | - | 1. How do you prefer to study? 2. Where do you prefer to study? | Self-developed |
3. | AI vs. teacher feedback | Ordinal | Scale 2 | I think that this AI-based essay tool is more effective than a teacher’s feedback | Adapted from (Sung and Mayer, 2012) [75] |
4. | Affinity for technology interaction (ATI) | Ordinal | Scale 2 |
| Adapted from (Franke et al., 2019) [51] |
5. | Trust in AI applications | Ordinal | Scale 1 |
| Adapted from (Cheung and To, 2017) [76] |
6. | Trust in AI-based essay tool | Ordinal | Scale 1 |
| Adapted from (Cheung and To, 2017) [76] |
7. | Perceived benefits | Ordinal | Scale 1 |
| Adapted from (Shroff et al., 2011) [77] |
8. | Behavioral intention to use AI-based educational tools | Ordinal | Scale 1 |
| Adapted from (Pavlou and Gefen, 2004) [78] |
Factor Construct | Item | Est. Std | Z-Val | CI. Lower | CI. Upper |
---|---|---|---|---|---|
Behavioral intention to use AI-based educational tools | Plan to use | 0.914 | 35.358 | 0.863 | 0.964 |
Possibility to use | 0.648 | 12.463 | 0.546 | 0.750 | |
Use if opportunity | 0.867 | 29.399 | 0.809 | 0.925 | |
Perceived benefits | Enhance learning | 0.769 | 19.671 | 0.693 | 0.846 |
Improved studies | 0.813 | 23.857 | 0.746 | 0.880 | |
Increased productivity | 0.779 | 20.420 | 0.704 | 0.854 | |
Accelerated studies | 0.644 | 12.194 | 0.541 | 0.748 | |
Found useful | 0.803 | 22.714 | 0.734 | 0.872 | |
Trust in AI-based essay grading tool | Believable | 0.867 | 33.661 | 0.816 | 0.917 |
Accurate | 0.859 | 31.780 | 0.806 | 0.912 | |
Confidence | 0.866 | 32.836 | 0.814 | 0.917 | |
Reliable | 0.812 | 24.974 | 0.748 | 0.876 | |
Trust in AI applications | Easy to trust | 0.885 | 35.182 | 0.836 | 0.934 |
High tendency | 0.971 | 50.471 | 0.934 | 1.009 | |
Blindly trusting | 0.748 | 19.138 | 0.671 | 0.824 | |
Affinity for technology interaction | Detail focus | 0.733 | 17.022 | 0.649 | 0.817 |
Test functions | 0.843 | 27.871 | 0.783 | 0.902 | |
Intensive trial | 0.830 | 26.949 | 0.770 | 0.890 | |
Enjoy learning | 0.885 | 35.949 | 0.837 | 0.933 | |
Full use | 0.713 | 15.792 | 0.624 | 0.801 |
Measure | Behavioral Intention | Perceived Benefits | Trust in AI-Based Essay Grading Tool | Trust in AI Applications | Affinity for Technology Interaction |
---|---|---|---|---|---|
Alpha | 0.8498 | 0.8729 | 0.9119 | 0.9002 | 0.8981 |
Omega | 0.8644 | 0.8716 | 0.9128 | 0.9007 | 0.9018 |
Avevar | 0.6878 | 0.5770 | 0.7245 | 0.7529 | 0.6506 |
Factor 1 | Factor 2 | Estimate | z-Value | p (>|z|) |
---|---|---|---|---|
Behavioral intention to use AI-based educational tools | Perceived benefits | 0.7350 | 15.2990 | 0 |
Trust in AI-based essay grading tool | 0.3720 | 4.7170 | 0 | |
Trust in AI applications | 0.5270 | 7.9870 | 0 | |
Affinity for technology interaction | 0.1990 | 2.2910 | 0.0220 | |
Perceived benefits | Trust in AI-based essay grading tool | 0.6050 | 9.8260 | 0 |
Trust in AI applications | 0.4700 | 6.6040 | 0 | |
Affinity for technology interaction | 0.1220 | 1.3710 | 0.1700 | |
Trust in AI-based essay grading tool | Trust in AI applications | 0.5100 | 7.7460 | 0 |
Affinity for technology interaction | 0.0730 | 0.8310 | 0.4060 | |
Trust in AI applications | Affinity for technology interaction | 0.1470 | 1.7390 | 0.0820 |
Term | Estimate | Est. Error | Q2.5 | Q97.5 |
---|---|---|---|---|
Trust in AI applications | 0.4368 | 0.1715 | 0.1017 | 0.7745 |
Affinity for technology interaction | 0.1000 | 0.1650 | −0.2193 | 0.4247 |
Trust in AI applications: affinity for technology interaction | 0.2015 | 0.1505 | −0.0965 | 0.4932 |
Age group | 0.1762 | 0.2633 | −0.3445 | 0.7077 |
Term | Estimate | Est. Error | Q2.5 | Q97.5 |
---|---|---|---|---|
Education program 2 | −0.2828 | 0.1645 | −0.6077 | 0.0457 |
Education program 4 | −0.7954 | 0.4317 | −1.6359 | 0.0509 |
Preferred study style 2 | 0.2656 | 0.3236 | −0.3687 | 0.9019 |
Preferred study location 2 | −0.1384 | 0.1670 | −0.4680 | 0.1848 |
Trust in AI applications | 0.3788 | 0.0742 | 0.2343 | 0.5244 |
Affinity for technology interaction | 0.0556 | 0.0721 | −0.0857 | 0.1963 |
Preferred study style 2: preferred study location 2 | −0.1124 | 0.3705 | −0.8422 | 0.6179 |
Preferring AI over teacher | 0.3132 | 0.0990 | 0.1365 | 0.5269 |
H3. There is a significant association between students’ demographics and perceived benefits of AI-based essay grading. | Not supported. Our data did not support including demographic variables in the model. |
H7. Students who prefer to study alone perceive more benefits from AI-based essay grading | Inconclusive. The perceived benefit was lower for those preferring studying alone; however, we could not rule out zero () between differences. |
H10. Students who believe that AI-based assessment is more effective than teacher feedback perceive more benefits from AI-based essay grading | Supported. Students who believe that AI-based assessment is more effective than teacher feedback perceive more benefits from AI-based essay grading. |
H13. Students’ affinity for technology interaction does not significantly influence their perceived benefits of AI-based essay grading. | Supported. We found no notable trend (0.0556) between affinity for technology interaction and perceived benefits of AI-based essay grading. |
Term | Estimate | Est. Error | Q2.5 | Q97.5 |
---|---|---|---|---|
Trust in AI applications | 0.3834 | 0.0638 | 0.2597 | 0.5108 |
Affinity for technology interaction | 0.0170 | 0.0616 | −0.105 | 0.1366 |
Preferring AI over teacher | 0.2461 | 0.0774 | 0.1053 | 0.4125 |
H1. Younger students (under 40) have higher trust in AI-based essay grading than older students | Not supported. Our data did not support including demographic variables in the model. |
H2. Male students tend to show more trust in AI-based essay grading compared to other genders | Not supported. Our data did not support including demographic variables in the model. |
H5. Students of IT or other digital education programs have higher trust in the AI-based essay grading tool than the students of other degree programs. | Not supported. Our data did not support including program variables in the model. |
H8. Students who prefer to study alone have more trust in AI-based essay grading. | Not supported. Our data did not support including the study style variable in the model. |
H12. Student’s affinity for technology interaction does not significantly influence their trust in AI-based essay grading. | Supported. We found no notable trend (0.0170) between affinity for technology interaction and perceived benefits of AI-based essay grading. |
H9. Students who believe that AI-based assessment is more effective than teacher feedback have more trust in AI-based essay grading | Supported. Students who believed that AI-based assessment is more effective than teacher feedback found the tool notably more trustful (trend 0.2461). |
Term | Estimate | Est. Error | Q2.5 | Q97.5 |
---|---|---|---|---|
Education program 2 | −0.2776 | 0.1515 | −0.5821 | 0.0132 |
Education program 4 | −0.6489 | 0.4126 | −1.4085 | 0.2095 |
Preferred study style 2 | 0.3026 | 0.2947 | −0.2493 | 0.8974 |
Preferred study location 2 | −0.2992 | 0.1564 | −0.6014 | 0.0070 |
Trust in AI applications | 0.4485 | 0.0684 | 0.3184 | 0.5881 |
Affinity for technology interaction | 0.0999 | 0.0643 | −0.0289 | 0.2230 |
Preferred study style 2: preferred study location 2 | −0.3147 | 0.3363 | −0.9873 | 0.3290 |
Preferring AI over teacher | 0.1421 | 0.0861 | −0.0107 | 0.3253 |
Contrast | Estimate | Lower. HP | Upper. HP | |
---|---|---|---|---|
Preferred study style = 2 and preferred study location = 1 - preferred study style = 1 and preferred study location = 2 | 0.5920 | 0.0183 | 1.1946 | * |
Preferred study style = 2 and preferred study location = 1 - preferred study style = 2 and preferred study location = 2 | 0.6068 | 0.0435 | 1.2294 | * |
Preferred study location = 1 - preferred study location = 2 | 0.4543 | 0.0125 | 0.9100 | ** |
H4. There is a significant association between students’ demographics and behavioral intention to use AI-based educational tools in the future. | Not supported. Our data did not support including demographic variables in the model. |
H6. Students who prefer to study remotely have a higher behavioral intention to use AI-based educational tools | Supported. Intention to use AI-based educational tools was higher for those who prefer to study remotely (difference 0.45). |
H11. Students who believe that AI-based assessment is less effective than teacher feedback have a lower behavioral intention to use AI-based educational tools. | Inconclusive. We noticed a positive trend (0.1421) but could not rule out zero (). |
H14. Students with a higher affinity for technology interaction have higher behavioral intention to use AI-based educational tools. | Inconclusive. Although there was a clear positive trend, it was not notably different from zero. |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Alamäki, A.; Khan, U.A.; Kauttonen, J.; Schlögl, S. An Experiment of AI-Based Assessment: Perspectives of Learning Preferences, Benefits, Intention, Technology Affinity, and Trust. Educ. Sci. 2024, 14, 1386. https://doi.org/10.3390/educsci14121386
Alamäki A, Khan UA, Kauttonen J, Schlögl S. An Experiment of AI-Based Assessment: Perspectives of Learning Preferences, Benefits, Intention, Technology Affinity, and Trust. Education Sciences. 2024; 14(12):1386. https://doi.org/10.3390/educsci14121386
Chicago/Turabian StyleAlamäki, Ari, Umair Ali Khan, Janne Kauttonen, and Stephan Schlögl. 2024. "An Experiment of AI-Based Assessment: Perspectives of Learning Preferences, Benefits, Intention, Technology Affinity, and Trust" Education Sciences 14, no. 12: 1386. https://doi.org/10.3390/educsci14121386
APA StyleAlamäki, A., Khan, U. A., Kauttonen, J., & Schlögl, S. (2024). An Experiment of AI-Based Assessment: Perspectives of Learning Preferences, Benefits, Intention, Technology Affinity, and Trust. Education Sciences, 14(12), 1386. https://doi.org/10.3390/educsci14121386