Agent4Edu: Generating Learner Response Data by Generative Agents
for Intelligent Education Systems
Agent4Edu: 通过生成式代理为智能教育系统生成学习者响应数据
Abstract 摘要
Personalized learning represents a promising educational strategy within intelligent educational systems, aiming to enhance learners’ practice efficiency. However, the discrepancy between offline metrics and online performance significantly impedes their progress. To address this challenge, we introduce Agent4Edu, a novel personalized learning simulator leveraging recent advancements in human intelligence through large language models (LLMs). Agent4Edu features LLM-powered generative agents equipped with learner profile, memory, and action modules tailored to personalized learning algorithms. The learner profiles are initialized using real-world response data, capturing practice styles and cognitive factors. Inspired by human psychology theory, the memory module records practice facts and high-level summaries, integrating reflection mechanisms. The action module supports various behaviors, including exercise understanding, analysis, and response generation.
Each agent can interact with personalized learning algorithms, such as computerized adaptive testing, enabling a multifaceted evaluation and enhancement of customized services. Through a comprehensive assessment, we explore the strengths and weaknesses of Agent4Edu, emphasizing the consistency and discrepancies in responses between agents and human learners. The code, data, and appendix are publicly available at https://github.com/bigdata-ustc/Agent4Edu.
个性化学习是智能教育系统中一种有前景的教育策略,旨在提高学习者的实践效率。然而,线下指标与线上表现之间的差异显著阻碍了他们的进步。为应对这一挑战,我们引入了 Agent4Edu,这是一个利用大型语言模型(LLMs)最新进展来模拟人类智能的新型个性化学习模拟器。Agent4Edu 具有由 LLM 驱动的生成式代理,配备了针对个性化学习算法的学员档案、记忆和行动模块。学员档案使用真实世界的响应数据初始化,捕捉练习风格和认知因素。受人类心理学理论启发,记忆模块记录练习事实和高级摘要,并整合反思机制。行动模块支持多种行为,包括练习理解、分析和响应生成。每个代理可以与个性化学习算法(如计算机化自适应测试)交互,从而实现定制化服务的多维度评估和提升。 通过全面评估,我们探讨了 Agent4Edu 的优势和不足,重点分析了智能体与人类学习者之间回答的一致性与差异性。代码、数据和附录可在 https://github.com/bigdata-ustc/Agent4Edu 公开获取。
1 Introduction 1 引言
Intelligent education platforms like Coursera.com and LeetCode.com provide a rich array of learning resources, such as courses and exercises, within a flexible online environment. The accessibility and convenience of these platforms have attracted a growing number of learners. A key online learning activity is “practice”, where learners independently select and answer exercises. The platforms record their responses, such as the correctness of their answers. By analyzing response data, many personalized learning services, such as exercise recommendations, knowledge tracing, and computerized adaptive testing, can be tailored to meet each learner’s specific needs, enhancing the learning process and increasing learner satisfaction. For instance, on LeetCode, analyzing a learner’s historical programming experiences allows the platform to recommend exercises of appropriate difficulty levels, thus optimizing learning gains.
像 Coursera.com 和 LeetCode.com 这样的智能教育平台,在一个灵活的在线环境中提供了丰富的学习资源,如课程和练习。这些平台的可访问性和便利性吸引了越来越多的学习者。一个关键的在线学习活动是“练习”,学习者可以独立选择并回答练习题。平台会记录他们的回答,例如回答的正确性。通过分析回答数据,可以提供许多个性化学习服务,如练习推荐、知识追踪和计算机化自适应测试,以满足每个学习者的特定需求,从而增强学习过程并提高学习者满意度。例如,在 LeetCode 上,通过分析学习者的历史编程经验,平台可以推荐适当难度的练习题,从而优化学习收益。
The effectiveness of personalized learning services hinges on the availability of high-quality response data for the corresponding algorithm training. However, the scarcity of offline response data and potential biases in its correlation with online practice introduces a significant gap between offline metrics and actual online performance. This discrepancy impedes the integration of research with real-world applications.
To bridge this gap, a promising approach is to simulate learner response data. Imagine an online platform equipped with a configurable simulation system that faithfully captures human learners’ response patterns while seamlessly interacting with personalized learning algorithms. Such a simulator undoubtedly has the potential to revolutionize the traditional research paradigm in intelligent education, providing innovative avenues for response data collection, personalized algorithm development and evaluation.
个性化学习服务的有效性取决于相应算法训练所需的高质量响应数据。然而,线下响应数据的稀缺性及其与在线练习的潜在关联偏差,在离线指标与实际在线表现之间造成了显著差距。这种差异阻碍了研究与实际应用的结合。为弥补这一差距,一种有前景的方法是模拟学习者响应数据。想象一个配备可配置模拟系统的在线平台,该系统能够忠实地捕捉人类学习者的响应模式,并与个性化学习算法无缝交互。这样的模拟器无疑有潜力革新智能教育中的传统研究范式,为响应数据收集、个性化算法开发与评估提供创新途径。
Several approaches to simulating learner response data have been proposed and have achieved notable success (Piech et al. 2015; Zhao et al. 2023). However, two major limitations exist in current approaches:
(1) Simplified Simulations. Most existing studies predict learners’ responses (e.g., correct or incorrect answers) without considering the detailed answer processes by which humans use their knowledge to understand, analyze, and solve problems. Hence, these simulations may lack reliability and interpretability.
(2) Dependency on Real Response Data. An ideal simulator should be capable of simulating learner responses even when real-world datasets are insufficiently available, thereby enhancing its applicability. However, current methods require high-quality real-world data to train the simulation strategy. As a result, these methods can only generate learner response data similar to existing real-world datasets and struggle to generalize to more challenging scenarios, such as zero-shot simulations.
已有多种方法被提出用于模拟学习者响应数据,并取得了显著成功(Piech et al. 2015; Zhao et al. 2023)。然而,当前方法存在两大主要局限性:(1) 简化模拟。大多数现有研究仅预测学习者的响应(例如,正确或错误的答案),而未考虑人类如何运用其知识来理解、分析和解决问题所涉及的详细过程。因此,这些模拟可能缺乏可靠性和可解释性。(2) 依赖真实响应数据。理想的模拟器应能够在真实世界数据不足的情况下模拟学习者响应,从而提高其适用性。然而,当前方法需要高质量的真实现实数据来训练模拟策略。结果,这些方法只能生成与现有真实现实数据集相似的学习者响应数据,难以泛化到更具挑战性的场景,例如零样本模拟。
Recent advancements in large language models (LLMs) have demonstrated remarkable capabilities in autonomous interaction and decision-making (Brown et al. 2020; Ouyang et al. 2022; Yue et al. 2023; Jin et al. 2023; Long et al. 2024). These advancements underscore the potential of leveraging LLM-powered agents to simulate human social behaviors, such as daily life in Smallville (Park et al. 2023) and software development (Qian et al. 2023). LLM-based user simulators possess rich pre-trained knowledge and human-like intelligence, enabling them to perceive and simulate intricate human practice processes. Furthermore, their in-context learning ability allows LLMs to perform zero-shot simulations with minimal reliance on real-world data (Wang et al. 2023c). Consequently, LLM-based generative agents present a promising approach for addressing the current limitations of learner response simulators.
大型语言模型(LLMs)的最新进展已展现出在自主交互和决策方面的卓越能力(Brown 等人 2020;Ouyang 等人 2022;Yue 等人 2023;Jin 等人 2023;Long 等人 2024)。这些进展凸显了利用 LLM 驱动的智能体模拟人类社会行为(如《小爱镇》的日常生活和软件开发)的潜力(Park 等人 2023;Qian 等人 2023)。基于 LLM 的用户模拟器拥有丰富的预训练知识,具备类人智能,能够感知和模拟复杂的人类实践过程。此外,LLMs 的情境学习能力使其能够进行零样本模拟,对真实世界数据的依赖性极低(Wang 等人 2023c)。因此,基于 LLM 的生成式智能体为解决学习者响应模拟器的当前局限性提供了一种有前景的方法。
In this paper, we introduce Agent4Edu, a personalized learning simulator designed for intelligent educational systems, comprising two key components: an LLM-powered generative agent and a personalized learning environment (see the framework in Figure 1).
From a learner perspective, the LLM-powered generative agent is responsible for simulating learners’ response data by capturing their response patterns and inferring problem-solving actions. Each agent is initialized based on available learner response data and consists of three modules: a learner profile, memory, and action module. The learner profile module stores learners’ past practice styles (e.g., activity) and cognitive factors (e.g., ability), aligning with human learners’ learning status. The memory module, inspired by psychological theories (Baker 2001) and human learning mechanism (Wang et al. 2023d), records past practice experiences and summarizes learning status through reflections. This facilitates coherent observations, monitors knowledge proficiency evolution, reinforces memory, and simulates human forgetting.
The action module enables agents to choose, understand, analyze, and solve exercises recommended by personalized learning algorithms, leading to more reliable and interpretable response generation.
Our agent can also utilize tools, such as employing the psychological IRT model (Baker 2001) to assess ability within the Profile module and using DNeuralCDM (Wang et al. 2023a) to trace knowledge proficiency evolution within the Memory module.
From a personalized learning perspective, the learning environment can be configured with any personalized learning algorithm, allowing agents to interact directly and simulate a real learning environment.
Notably, despite extensive research on simulating user behavior with generative agents, we are the first to focus specifically on educational scenarios to generate response data for individual learners.
在本文中,我们介绍了 Agent4Edu,这是一个为智能教育系统设计的个性化学习模拟器,包含两个关键组件:一个由 LLM 驱动的生成式代理和一个个性化学习环境(见图 1)。从学习者的角度来看,由 LLM 驱动的生成式代理负责通过捕捉学习者的响应模式并推断问题解决行为来模拟学习者的响应数据。每个代理都基于可用的学习者响应数据进行初始化,并包含三个模块:学习者画像模块、记忆模块和行动模块。学习者画像模块存储学习者的过去练习风格(例如,活动)和认知因素(例如,能力),与人类学习者的学习状态相一致。记忆模块受心理学理论(Baker 2001)和人类学习机制(Wang 等人 2023d)的启发,记录过去的练习经验并通过反思总结学习状态。这有助于连贯的观察、监控知识熟练度的演变、强化记忆并模拟人类的遗忘。 行动模块使智能体能够选择、理解、分析并解决个性化学习算法推荐的练习,从而生成更可靠和可解释的响应。我们的智能体还可以利用工具,例如在个人资料模块中使用心理 IRT 模型(Baker 2001)评估能力,并在记忆模块中使用 DNeuralCDM(Wang 等人 2023a)追踪知识熟练度的演变。从个性化学习的角度来看,学习环境可以配置为使用任何个性化学习算法,使智能体能够直接交互并模拟真实的学习环境。值得注意的是,尽管在利用生成式智能体模拟用户行为方面已有大量研究,但我们首次专注于教育场景,为单个学习者生成响应数据。
Our main contributions are summarized as follows:
我们的主要贡献总结如下:
-
•
We develop Agent4Edu, a personalized learning simulator that leverages LLM-powered generative agents to simulate human learners’ response data as well as demonstrate the practice process. Additionally, the agent interacts with personalized learning environments to evaluate and improve intelligent tutoring algorithms.
• 我们开发了 Agent4Edu,一个个性化学习模拟器,它利用 LLM 驱动的生成式代理来模拟人类学习者的响应数据,并展示实践过程。此外,该代理与个性化学习环境交互,以评估和改进智能辅导算法。 -
•
Our generative agents, featuring profile, memory, and action modules specifically designed for “Education”, can not only generate response data but also accurately simulate human choices, understanding, analysis, and problem-solving for exercises, outperforming existing learner simulation methods.
• 我们的生成式代理,具有为“教育”专门设计的配置文件、记忆和行动模块,不仅可以生成响应数据,还可以准确模拟人类在练习中的选择、理解、分析和问题解决能力,超越了现有的学习者模拟方法。 -
•
To systematically evaluate Agent4Edu, we conduct comprehensive experiments from both the agent and personalized learning perspectives. From the agent perspective, we assess the consistency between the agents and human learners. From the learning perspective, we evaluate and improve personalized learning algorithms for computerized adaptive testing, based on generative agents and simulated data. Extensive experimental results demonstrate the effectiveness of Agent4Edu.
• 为系统地评估 Agent4Edu,我们从智能体和个性化学习的角度开展全面实验。从智能体角度,我们评估智能体与人类学习者的行为一致性。从学习角度,我们基于生成式智能体和模拟数据,评估并改进计算机化自适应测试的个性化学习算法。大量的实验结果证明了 Agent4Edu 的有效性。
2 Related Work 2 相关工作
Learner Response Data Simulation
Learner Simulation aims to address the shortage of high-quality practice data in intelligent educational systems and has been applied in numerous previous studies (Zhao et al. 2023; Yao et al. 2024). Memory-based (Reddy, Levine, and Dragan 2017) relies on manually crafted rules to predict learners’ responses or memory behavior. EERNN (Su et al. 2018) and KES (Liu et al. 2019) utilize RNN-based models to forecast learners’ performance. DAISim (Zhao et al. 2023) constructs learner simulations as Markov decision processes, simultaneously considering learners’ long and short-term question-answering patterns. However, the memory-based simulator is overly simplistic and cannot simulate complex interactions. Most other learner simulators simplify the student answering process and face challenges in conducting zero-shot simulations due to their reliance on data. In this paper, we employ an LLM-powered agent to simulate the student practice process, addressing these limitations.
学习者响应数据模拟学习者模拟旨在解决智能教育系统中高质量练习数据的短缺问题,并在众多先前研究中得到应用(Zhao 等人,2023;Yao 等人,2024)。基于记忆的模拟(Reddy,Levine 和 Dragan,2017)依赖于手动设计的规则来预测学习者的响应或记忆行为。EERNN(Su 等人,2018)和 KES(Liu 等人,2019)利用基于 RNN 的模型来预测学习者的表现。DAISim(Zhao 等人,2023)将学习者模拟构建为马尔可夫决策过程,同时考虑学习者的长期和短期问答模式。然而,基于记忆的模拟器过于简单,无法模拟复杂的交互。大多数其他学习者模拟器简化了学生答题过程,并且由于依赖数据,在执行零样本模拟时面临挑战。在本文中,我们采用基于 LLM 的智能体来模拟学生练习过程,以解决这些局限性。
Personalized Learning Services
Intelligent educational systems offer learners personalized learning services, including Computerized Adaptive Testing (CAT) (Chang and Ying 1996), exercise recommendation (Huang et al. 2019) and learning path suggestions (Liu et al. 2019), to help learners enhance their skills.
In this work, we select the representative and popular CAT services as our personalized learning scenarios for our study and experiments.
CAT is an advanced educational measurement method that evaluates the knowledge level of examinees in minor exercises, which has been widely used in various standardized tests (e.g., GMAT and GRE) (Zhuang et al. 2024; Bi et al. 2020; Lord 2012; Chang and Ying 1996). However, current CAT models require high-quality practice data to train a cognitive diagnosis model for evaluating learner ability or knowledge proficiency, which is often challenging to gather. Therefore, in this paper, we employ the CAT service within learning systems to assess the quality of the data generated by our agents. Additionally, we investigate the potential for enhancing CAT models using simulated data.
个性化学习服务 智能教育系统为学习者提供个性化学习服务,包括计算机化自适应测试(CAT)(Chang and Ying 1996)、练习推荐(Huang et al. 2019)和学习路径建议(Liu et al. 2019),以帮助学习者提升技能。在本工作中,我们选择具有代表性和广泛应用的 CAT 服务作为我们的个性化学习场景进行研究和实验。CAT 是一种先进的教育测量方法,通过小练习评估考生的知识水平,已广泛应用于各种标准化考试(例如 GMAT 和 GRE)(Zhuang et al. 2024;Bi et al. 2020;Lord 2012;Chang and Ying 1996)。然而,当前的 CAT 模型需要高质量的练习数据来训练认知诊断模型,以评估学习者的能力或知识熟练度,这通常难以收集。因此,在本文中,我们利用学习系统中的 CAT 服务来评估我们代理生成的数据质量。此外,我们还研究了使用模拟数据增强 CAT 模型的潜力。
LLM-based Agents
LLM-based generative agents demonstrate the remarkable capabilities to perceive their environment, make decisions, and take actions, thus, emerging a substantial amount of research (Wang et al. 2024b). The development of generative agents (Park et al. 2023), designed with profile, memory, action, and reflective capabilities, represents pioneering work in simulating human daily life. Within this general framework, agents tailored to specific tasks (Qian et al. 2023; Wu et al. 2023; Wang et al. 2023b; Huang et al. 2023; Zhang et al. 2023b, a) and simulations (Gao et al. 2023; Wang et al. 2023d; Park et al. 2023; Liu et al. 2023; Wang et al. 2023c) have been constructed. Recent research highlights bringing generative agents to educational settings (Li et al. 2024; Dan et al. 2023; Kieser et al. 2023). For example, (Qadir 2023; Rahman and Watanobe 2023) conclude the applications of ChatGPT to engineering education. (Baidoo-Anu and Ansah 2023) focus on the literature review over the published paper. SocraticLM (Liu et al. ) embodies a “Thought-Provoking” teaching paradigm, engaging students in active problem-solving, akin to a real classroom teacher. The most relevant part of our work is EduAgent (Xu, Zhang, and Qin 2024) which utilizes LLM-based agents to simulate learners studying PowerPoint presentations and videos, predicting their quiz outcomes to assess performance. However, this approach relies on expert-annotated cognitive factors to initialize agents, disregarding the understanding and analysis of exercises. In contrast, our Agent4Edu extracts cognitive factors from data using tools and captures practice styles, allowing it to simulate the detailed exercise understanding and analysis process and interact effectively with personalized learning algorithms.
基于 LLM 的智能体 基于 LLM 的生成式智能体展现出感知环境、做出决策和采取行动的卓越能力,因此涌现出大量研究(Wang 等人 2024b)。生成式智能体(Park 等人 2023)的开发,其具备档案、记忆、行动和反思能力,代表了模拟人类日常生活的开创性工作。在这个通用框架内,已经构建了针对特定任务的智能体(Qian 等人 2023;Wu 等人 2023;Wang 等人 2023b;Huang 等人 2023;Zhang 等人 2023b,a)和模拟(Gao 等人 2023;Wang 等人 2023d;Park 等人 2023;Liu 等人 2023;Wang 等人 2023c)。近期研究重点是将生成式智能体引入教育环境(Li 等人 2024;Dan 等人 2023;Kieser 等人 2023)。例如,(Qadir 2023;Rahman 和 Watanobe 2023)总结了 ChatGPT 在工程教育中的应用。(Baidoo-Anu 和 Ansah 2023)专注于已发表论文的文献综述。SocraticLM(Liu 等人)体现了一种“启发思考”的教学范式,让学生参与主动问题解决,类似于真实的课堂教师。 我们工作的最相关部分是 EduAgent(Xu, Zhang, and Qin 2024),它利用基于 LLM 的智能体来模拟学习者学习 PPT 和视频,预测他们的测验结果以评估表现。然而,这种方法依赖于专家标注的认知因素来初始化智能体,忽略了练习的理解和分析。相比之下,我们的 Agent4Edu 使用工具从数据中提取认知因素,捕捉练习风格,使其能够模拟详细的练习理解和分析过程,并有效地与个性化学习算法互动。
图 1:Agent4Edu 的整体框架。
3 Agent4Edu
Agent4Edu is a personalized learning simulator, aimed at accurately simulating learners’ response data and facilitating responsive personalized learning algorithms. It contains two key components: (1) LLM-powered generative agents that capture learners’ practice patterns and cognitive preferences to simulate their response, and (2) a personalized learning environment that interacts with agents to support accurate and interpretable evaluations and improvements of mainstream intelligent algorithms (e.g., computerized adaptive testing).
The framework of Agent4Edu is illustrated in Figure 1. All the prompts are listed in Appendix C.
Agent4Edu 是一个个性化学习模拟器,旨在准确模拟学习者的响应数据,并促进响应式个性化学习算法。它包含两个关键组件:(1)由 LLM 驱动的生成式代理,这些代理捕捉学习者的练习模式和认知偏好以模拟他们的响应,以及(2)一个个性化学习环境,该环境与代理互动,以支持主流智能算法(例如计算机化自适应测试)的准确和可解释的评估和改进。Agent4Edu 的框架如图 1 所示。所有提示都列在附录 C 中。
3.1 Task Formulation 3.1 任务公式化
Suppose there are learners, exercises in an intelligent educational system.
For a learner , his/her response data are denoted as a time-ordered set , where represents the exercise that learner practiced at step , and is ’s response to exercise , which is usually denoted as a binary value, i.e., if learner answers correctly, otherwise . denotes textual information of each exercise , e.g., textual content and corresponding knowledge concepts. We provide in a form, as the example in Figure 1.
假设在一个智能教育系统中存在 名学习者, 项练习。对于学习者 ,其响应数据表示为一个时间有序集 ,其中 代表学习者 在步骤 练习的练习, 是 对练习 的响应,这通常表示为一个二进制值,即如果学习者 正确回答 ,否则 。 表示每个练习 的文本信息,例如文本内容和相应的知识概念。我们以 的形式提供 ,如图 1 所示。
Based on the above conditions, the simulator’s overarching goal is to faithfully distill the human learners’ learning patterns and cognitive preferences, and accurately generate their future response data on unseen exercises. Please note that existing personalized learning algorithms usually assume that learners only submit each exercise once, so repeated submission is not considered in our simulation.
基于上述条件,模拟器的总体目标是忠实地提炼人类学习者的学习模式和认知偏好,并准确生成他们对未见过练习的未来响应数据。请注意,现有的个性化学习算法通常假设学习者只提交每个练习一次,因此在我们的模拟中不考虑重复提交。
3.2 LLM-powered Agent 3.2 基于大型语言模型的代理
The generative agent in Agent4Edu uses LLM as its foundational architecture, enhancing its functionality tailored for the personalized learning scenario through three specialized modules: learner profile, memory, and action modules. To mimic actual personalized practice responses akin to humans, we construct an individual agent for each learner . Each agent integrates a learner profile module aimed at reflecting personalized practice patterns and cognitive factors. Additionally, each agent is equipped with a memory module designed to store past practice records and summarize high-level ideas. To simulate learner practice behavior more cohesively, the agent is also equipped with an action module.
Agent4Edu 中的生成式代理以 LLM 为基础架构,通过三个专门模块——学习者画像、记忆和行动模块——增强其功能,使其更适合个性化学习场景。为了模拟人类般的实际个性化练习响应,我们为每个学习者构建一个个体代理 。每个代理集成了一个学习者画像模块,旨在反映个性化练习模式和认知因素。此外,每个代理还配备了一个记忆模块,用于存储过去的练习记录并总结高级观点。为了更连贯地模拟学习者练习行为,代理还配备了行动模块。
Learner Profile Module 学习者画像模块
The learner profile module represents some overall learning features of human learners, which are typically stable and derived from long-term learning experiences. We configure each agent ’s profile based on its corresponding learner ’s response data111Note that if zero-shot simulations are performed and user data is unavailable, the profile needs to be randomly generated.
请注意,如果进行零样本模拟且用户数据不可用,则需要随机生成个人资料。
学习者个人资料模块代表人类学习者的某些总体学习特征,这些特征通常稳定且源自长期学习经验。我们根据每个代理 对应的学习者 的响应数据 1 配置其个人资料。. Each agent’s initial configuration is divided into two categories: explicit practice styles and implicit cognitive factors.
每个代理的初始配置分为两类:显式练习风格和隐性认知因素。
Practice styles are statistical features explicitly derived from the available practice record of each learner , such as learning activity (Baker 2001; Gao et al. 2021), practice diversity (Bi et al. 2020), success rate, and preference. Activity indicates learners’ enthusiasm for learning and provides clues for simulating their practice behaviors. For example, learners with higher enthusiasm for learning usually perform better. Mathematically, the activity level of learner is defined as . Practice diversity reflects the knowledge coverage practiced by learners, represented as , where is the number of knowledge concepts practiced by learner . Higher diversity indicates greater curiosity in learners. Success rate correlates with the probability of learners answering questions correctly, making it another essential characteristic. The success rate for learner is mathematically represented as . Preference refers to the knowledge concepts that learners practice most frequently.
练习风格是明确从每个学习者的可用练习记录 中提取的统计特征,例如学习活动(Baker 2001;Gao 等人 2021)、练习多样性(Bi 等人 2020)、成功率以及偏好。活动表明学习者对学习的热情,并为模拟他们的练习行为提供线索。例如,对学习更有热情的学习者通常表现更好。从数学上讲,学习者 的活动水平定义为 。练习多样性反映了学习者练习的知识覆盖范围,表示为 ,其中 是学习者 练习的知识概念数量。更高的多样性表明学习者更有好奇心。成功率与学习者正确回答问题的概率相关,是另一个重要特征。学习者 的成功率从数学上表示为 。偏好是指学习者练习最频繁的知识概念。
Cognitive factors are implicit features studied in psychology (Baker 2001; Chen et al. 2024), which significantly impact learner ’s practice performance. We select problem-solving ability and knowledge proficiency (Cheng et al. 2024) for this study. Problem-solving ability is assumed to be stable during the learning process, while knowledge proficiency typically evolves with learning progress (Huang et al. 2020). Therefore, in the profile module, we only configure the ability factor , with knowledge mastery being considered in the subsequent memory module. To obtain implicit ability, we assign a psychological IRT model (Baker 2001) trained on the observed learner response records, as the tool for the agent, allowing it to infer each learner ’s ability factor from the response data . The training and use of the IRT tool are detailed in Appendix B.
认知因素是心理学中研究的隐性特征(Baker 2001;Chen 等人 2024),它们显著影响学习者 的实践表现。在本研究中,我们选择了问题解决能力和知识熟练度(Cheng 等人 2024)。问题解决能力在学习过程中被认为是稳定的,而知识熟练度通常随着学习进度而发展(Huang 等人 2020)。因此,在个人资料模块中,我们仅配置能力因素 ,而知识掌握则在后续的记忆模块中考虑。为了获得隐性能力,我们为智能体分配了一个基于观察到的学习者响应记录进行训练的心理 IRT 模型(Baker 2001),作为其工具,使其能够从响应数据 推断每个学习者 的能力因素。IRT 工具的训练和使用在附录 B 中有详细说明。
Notably, we segment the values of each of the above features into several tiers in order to better prompt the generative agent inspired by (Wang et al. 2023d).
For a detailed exposition,
refer to Appendix A.1.
Additionally, to ensure broad applicability and protect privacy, certain personal identifiers (such as name, gender, age, and occupation) are intentionally anonymized in this work (Zhang et al. 2023a; Li et al. 2023). While these attributes may help shape other types of agents, they are not primary factors affecting practice performance in education. Our approach, based on both behavioral practice styles and psychological cognitive settings, can support a comprehensive representation of real learners.
值得注意的是,我们将上述每个特征值划分为几个层级,以便更好地提示受(Wang et al. 2023d)启发的生成式代理。详细说明请参见附录 A.1。此外,为确保广泛适用性和保护隐私,本研究中有意对某些个人标识符(如姓名、性别、年龄和职业)进行了匿名化处理(Zhang et al. 2023a; Li et al. 2023)。虽然这些属性可能有助于塑造其他类型的代理,但它们并非影响教育实践表现的主要因素。我们的方法基于行为实践风格和心理认知环境,能够支持对真实学习者的全面表征。
Memory Module 记忆模块
The Memory module allows the LLM-based agent to observe and summarize its corresponding learner ’s past practice experiences step by step. This module provides insightful clues to the agent for response simulation on unseen exercises. We follow the human learning mechanism (Atkinson 1968a; Cowan 2008; Huang et al. 2020; Wang et al. 2023d) to design three types of memories for each agent: factual memory, short-term memory, and long-term memory. Each memory is initially set to empty.
记忆模块允许基于 LLM 的智能体 逐步观察和总结其对应学习者 的过往练习经验。该模块为智能体在未见过练习题上的响应模拟提供了有见地的线索。我们遵循人类学习机制(Atkinson 1968a; Cowan 2008; Huang 等人 2020; Wang 等人 2023d)为每个智能体设计三种类型的记忆:事实记忆、短期记忆和长期记忆。每种记忆最初均设置为空。
Factual Memory:
In our simulation, factual memory is defined as the true learner’s past response records (i.e., observations). When the agent obtains a new response record of learner at step , i.e., , the response record is transmitted to the factual memory for processing.
事实记忆:在我们的模拟中,事实记忆被定义为真实学习者的过往响应记录(即观察记录)。当智能体在步骤 获得学习者 的新响应记录,即 时,该响应记录将被传输到事实记忆中进行处理。
Inspired by human learning mechanisms, if an agent repeatedly practices similar questions or knowledge, their memory is strengthened (Huang et al. 2020). Therefore, we introduce an additional counter (initially set to 1) for each record in factual memory to track the number of times it has been reinforced, a simple yet effective method that has been successfully used in user preference simulation (Wang et al. 2023d). Formally, for each agent , assume it has observed factual memories is , then it is allowed to receive a new response record . We first calculate the similarity between and each existing factual memory in the current memory .
The similarity between records can be defined as a metric that can be evaluated by LLMs, cosine similarity between text vectors, and other similar measures. In this case, we use the similarity relationships between the knowledge concepts involved in the records for the calculation. Specifically, we employ the statistical tool222https://github.com/bigdata-ustc/RCD
受人类学习机制启发,如果一个智能体反复练习相似的问题或知识,其记忆会得到加强(Huang 等人 2020)。因此,我们为事实记忆中的每条记录 引入一个额外的计数器 (初始值设为 1),以跟踪其被强化的次数,这是一种简单而有效的方法,已成功应用于用户偏好模拟(Wang 等人 2023d)。形式上,对于每个智能体 ,假设它已观察到 条事实记忆,记为 ,那么它允许接收一条新的响应记录 。我们首先计算 与当前记忆 中的每条现有事实记忆 之间的相似度。记录之间的相似度可以定义为一个 LLMs 可以评估的度量,如文本向量的余弦相似度以及其他类似度量。在这种情况下,我们使用记录中涉及的知识概念之间的相似关系进行计算。具体而言,我们采用统计工具 2 released by RCD (Gao et al. 2021) to determine whether two knowledge concepts are similar. If there is a similarity between the knowledge concepts involved in two records, the two records are considered similar.
For similar records, we increment the counter for by 1 (i.e., ), indicating that it has been reinforced by , and then add to factual memory; otherwise, is directly recorded without any reinforcement.
After processing and saving a new response record, factual memory triggers updating short-term and long-term memories.
由 RCD(高等人,2021)发布的工具用于判断两个知识概念是否相似。如果两条记录中涉及的知识概念存在相似性,则认为这两条记录相似。对于相似记录,我们将 的计数器加 1(即 ),表示它已被 强化,然后向事实记忆中添加 ;否则,直接记录 ,无需任何强化。在处理并保存一条新的响应记录后,事实记忆会触发短期记忆和长期记忆的更新。
We emphasize that the agent can only save response records into factual memory but cannot directly retrieve it, thereby allowing the retention of all exercise textual information and responses without being constrained by the LLM’s context length limitations.
我们强调,该智能体只能将响应记录保存到事实记忆中,但不能直接检索它,从而允许保存所有练习文本信息和响应,而不受 LLM 的上下文长度限制。
Short-term Memory:
Human short-term memory refers to the recent and temporary information that can be retained and recalled over a relatively brief period (Atkinson 1968b). Therefore, in our simulation, short-term memory is employed to retain the details of the agent’s most recent observed records. Assuming the current factual memory of agent is , then the short-term memory storage is defined as .
短期记忆:人类的短期记忆是指在一定相对较短的时期内可以被保留和回忆的近期信息(Atkinson 1968b)。因此,在我们的模拟中,短期记忆用于保留智能体最近观察到的 记录的详细信息。假设智能体 的当前事实记忆是 ,那么短期记忆存储定义为 。
Long-term Memory:
Long-term memory is formed through the reinforcement of memories from repeated practice and self-reflection inspired by to human long-term memories (Matelsky et al. 2023). It possesses a wide receptive field, allowing it to retain information observed long ago and generate high-level insights. We design the long-term memory using three types of information:
(1) Reinforced Facts: During each update of long-term memory, the agent first goes through the current factual memory . When the count of a record exceeds a preset threshold , indicating that the memory has been reinforced times, it is converted into long-term memory.
(2) Learning Process Summary: We utilize the LLM embedded in the agent to summarize the agent’s learning status from both short-term and long-term memories by Memory Reflection. Each step of the summary replaces the previous summary.
The summary consists of linguistic descriptions of the practice process and new insights from the agent itself. It overlooks practice details to filter out noise, irrelevant content, or potentially misleading information. Furthermore, compressing memory conserves significant space and enhances operational efficiency. (3) Knowledge Proficiency: We allow the agent to use an optimized DNeuralCDM (Wang et al. 2023a) based on the observed learner response data as a tool to obtain the learner’s dynamic proficiency (segmented into several tiers) evolution of specific knowledge concepts after each step of practice. The knowledge proficiency is a kind of dynamic cognitive factor significantly reflecting human responses in education (Piech et al. 2015; Wang et al. 2024a). The training and use of DNeuralCDM are given in Appendix B.
长期记忆:长期记忆是通过重复练习和自我反思的强化形成的,灵感来源于人类的长期记忆(Matelsky 等人,2023)。它具有广阔的接受域,能够保留很久之前观察到的信息并生成高级洞察。我们使用三种类型的信息来设计长期记忆:(1) 强化事实:在每次长期记忆更新时,智能体 首先会通过当前的事实记忆 。当记录 的计数 超过预设阈值 ,表明记忆已被强化 次时,它会被转化为长期记忆。(2) 学习过程摘要:我们利用嵌入在智能体中的 LLM 通过记忆反思来总结智能体从短期和长期记忆中的学习状态。每一步的摘要都会替换之前的摘要。摘要由智能体练习过程的语言描述和智能体自身产生的新见解组成。它忽略练习细节以过滤出噪声、无关内容或潜在的误导性信息。 此外,压缩内存可节省大量空间并提高运行效率。(3) 知识熟练度:我们允许代理使用基于观察到的学习者响应数据的优化 DNeuralCDM(Wang 等人,2023a)作为工具,以获取学习者每次练习后特定知识概念的动态熟练度(分为多个等级)的演变。知识熟练度是一种显著反映教育中人类反应的动态认知因素(Piech 等人,2015;Wang 等人,2024a)。DNeuralCDM 的训练和使用方法在附录 B 中给出。
Additionally, each factual record in long-term memory may be forgotten following the human forgetting curve theory (Averell and Heathcote 2011; Huang et al. 2020) that human memory decay starts rapidly and then gradually slows over time. We define a forgetting function associated timestamp and current observed step , i.e., , to simulate human learners’ forgetting. For each factual record in the long-term memory , it is forgotten if exceeds a predetermined threshold and its reinforcement frequency in factual memory is then reset as .
Overall, the factual response records are specific, while learning memory summaries are more general. By combining them, the agent can accurately perceive the learner’s practice process. Please note that traditional simulators (Piech et al. 2015; Zhao et al. 2023) can be regarded as owning the short-term memory but no long-term memory.
To help agents interact with the personalized learning environment, we introduce three memory operations:
Memory Retrieval: This operation assists the agent in extracting related information from memory. We allow the agent to retrieve the short-term and long-term memories finding reinforced facts and conducting summary.
Memory Writing: The raw observations are firstly input into the factual memory as facts. Then the recent facts are stored in short-term memory and the reinforced facts are written into long-term memory.
Memory Reflection: This operation occurs exclusively within long-term memory containing two aspects of reflections: (1) Summary Reflection is performed to summarize high-level ideas based on short-term and long-term memories, and (2) Corrective Reflection is performed when the agent’s action is inconsistent with the real learner, which will be introduced in Action Module.
Action Module
To equip the agent with learner profiles and memory modules, enabling it to exhibit human-like problem-solving behaviors and responses based on current observations, we design a specialized action module for each agent within Agent4Edu tailored for personalized learning. This module encompasses three main categories of actions:
Cognitive-driven Actions: In our simulation, personalized learning algorithms recommend one exercise to the agent at each step. The agent read the exercise’s content and decides whether or not to practice it, based on current cognitive factors. If the exercise is too challenging relative to the agent’s assessed ability and knowledge proficiency, the agent can opt to reject the recommended exercise.
Reading and Understanding Exercises. Simulating the process of reading and understanding exercises, similar to how humans approach them, provides valuable and interpretable insights into the agents. During each practice session, the agent is first required to identify and describe a knowledge concept assessed by the current exercise. If the agent correctly matches the exercise’s knowledge concept, it demonstrates an understanding of the exercise context akin to human learners. If the agent fails to do so, a corrective reflection is triggered to guide the agent towards the correct knowledge concept. This method reduces the risk of inaccuracies and ensures the agent’s credibility in simulating learner response (Zhang et al. 2023a).
Analyzing and Solving Exercises. Analyzing and solving exercises are crucial aspects of the learning process. Unlike previous simulation methods that directly predict the learner’s response in terms of answer correctness, our simulation requires the agent to emulate the learner’s answering process, which enhances both interpretability and credibility. To simulate this complex answer process more effectively, we improve agent’s reasoning ability through a chain-of-thought approach (Wei et al. 2022). Initially, the agent combines its profile and memories to formulate an initial solution idea for the exercise. Then, it writes the final answer to the exercise based on the solution idea. Afterwards, the agent predicts whether its answer is correct (i.e., performance prediction). If the predicted response does not match the real learner’s response, a corrective reflection is triggered. Note that, if standard answers of exercises are available, a scoring program can be designed to directly assess the correctness of the agent’s answer.
3.3 Personalized Learning Scenarios
Agent4Edu simulates agent and learning environment interaction (see Appendix D for a case study). The learning environment is designed as a standalone module that incorporates a series of personalized algorithms. These algorithms can recommend exercises to agents based on their past practice data. For instance, our experiments utilize computerized adaptive testing (CAT) strategies (Bi et al. 2020) for personalized learning. The module features an open interface, allowing researchers and practitioners to integrate external personalized learning algorithms seamlessly. This adaptability ensures that Agent4Edu serves as a versatile platform for comprehensive evaluations and the future collection of valuable learner response data.
4 Experiment
Dataset
Our dataset, called EduData, is provided by iFLYTEK Co., Ltd. It comprises 18,045 time-ordered response records from 500 Chinese high school students in the subjects of mathematics and physics. Each record includes the exercise ID, correctness, and timestamp. There are 1,032 exercises and 458 knowledge concepts in total, with each exercise testing one knowledge concept. Additionally, to facilitate reasoning and reflection for LLM-based agents, the platform provider has furnished us with the textual content of the exercises. In the experiment, we translate all Chinese text of exercises into English.
Experimental Setup
We use GPT-3.5-turbo and GPT-4 through OpenAI’s API service 333The detailed GPT versions: GPT-3.5-turbo-1106 (up to Sep. 2021) and GPT-4-turbo (up to Dec. 2023). to construct the agent for experimentation. When operating under the GPT-3.5-turbo configuration, all response data is utilized for experiments. Due to cost considerations, we simulate the task records of only 100 learners under the GPT-4 setting. The temperature parameter of GPT is 0 to avoid randomness. Empirically, we set the short-term memory size to 5, the threshold for memory enhancement is 5, and the threshold for forgetting in long-term memory is 0.99. Note that, in our experiments, unless explicitly specified, the LLM used is GPT-3.5-turbo.
4.1 LLM-based Agent Simulation Evaluation
Motivation:
The LLM-based agent is the core component of Agent4Edu. Exploring whether the agent can truly simulate human learners’ practice response is crucial for enhancing intelligent educational systems. We evaluate the effectiveness of the generative agent, including response simulation, exercises’ knowledge understanding, zero-shot simulations, and ablation experiments.
Learner Simulation Evaluation
The agent aims to generate simulated learner response data that closely approximates real responses. To validate the effectiveness of the simulation, we compare it with two traditional supervised simulation methods, including DAISIM (Zhao et al. 2023) and KES (Liu et al. 2019). Additionally, to enrich our baseline for a compelling comparison, we include several Knowledge Tracing (KT) models, such as DKVMN (Zhang et al. 2017), EERNN (with Markov) (Su et al. 2018) and SAKT (Pandey and Karypis 2019), which are similar to the learner simulator in terms of response prediction.
In the experimental setup, each learner’s records are divided into a 90% training set and a 10% test set. Each baseline model which is data-driven is trained on the training data, with the last 20% records of each learner’s training data used for model validation. The agent has access to all training data to generate profiles and update its memory through reflection. During the testing phase, each trained baseline is tasked with predicting learners’ binary responses (correct or incorrect) to unseen exercises in the test data. For our generative agent, exercises from the test data are sequentially sent to it, and it performs the designed three actions to solve them. If the agent rejects an exercise due to its difficulty, we label its response as an “incorrect answer”. The evaluation metrics are selected from two perspectives. Firstly, we use accuracy (ACC) and F1-score to measure prediction accuracy. Secondly, we assess the similarity between the simulated and real data distributions using ROUGE-3, inspired by (Zhao et al. 2023). We repeatedly run each baseline model five times in the same setups and the Table 1 reports the average scores.
The experimental results indicate that Agent4Edu (GPT-3.5-turbo) demonstrates strong competitiveness compared to the supervised baselines, particularly in terms of ACC and F1-score. This suggests that the LLM-based agent has the potential to generate learner response data that closely resembles real-world datasets. Furthermore, among the baselines, EERNN performs exceptionally well by effectively modeling the exercise content as supplementary clues. Finally, an exploratory simulation conducted using Agent4Edu (GPT-3.5-turbo and GPT-4) on a subset of data with 100 learners shows that they enable the simulated distribution to closely approach the real distribution. Among them, GPT-4 performs better in terms of ACC and F1-score.
Additionally, we evaluate whether the simulated distribution of the agent’s practice success rate aligns with the actual distribution of learner data. We use the real response success rate as the ground truth and then replace the corresponding responses in the real sequence with the predicted responses from the test data to calculate the agent’s simulated success rate, as shown in Figure 2 (a). The comparison between the ground truth values and the agent’s results indicates that the simulated data effectively captures the learners’ practice patterns related to success rate.
| Model | ACC | F1-score | ROUGE-3 |
|---|---|---|---|
| KES | 50.11 | 58.32 | 25.77 |
| DKVMN | 64.39 | 76.70 | 37.24 |
| EERNN | 65.72 | 76.06 | 43.55 |
| SAKT | 65.52 | 78.33 | 31.09 |
| DAISIM | 65.63 | 78.25 | 31.72 |
| Agent4Edu (GPT-3.5-turbo) | 66.70 | 79.84 | 37.97 |
| Agent4Edu (GPT-3.5-turbo)100 | 65.40 | 78.72 | 35.14 |
| Agent4Edu (GPT-4)100 | 66.51 | 79.53 | 34.86 |
| Model | ACC |
|---|---|
| Agent4Edu (GPT-3.5-turbo) | 73.88 |
| Agent4Edu (GPT-3.5-turbo)100 | 74.57 |
| Agent4Edu (GPT-4)100 | 82.43 |
Understanding Exercise-related Knowledge
To evaluate whether the agent understands a specific exercise, the agent is tasked with generating the knowledge concept tested by the exercise. Specifically, we create a candidate list containing one actual knowledge concept related to the exercise and two random knowledge concepts unrelated to the exercise. The agent must then select the relevant knowledge concept from this list based on its understanding (detailed prompts are provided in Appendix C). We use ACC as the metric to evaluate the agent’s knowledge predictions for all exercises in the test set, treating it as a binary classification task. This section uses the same agents from the section “Learner Simulation Evaluation”. The experimental results presented in Table 2, indicate that all the agents can correctly identify the knowledge being tested in most practice exercises. This demonstrates the strong human-like ability and rich knowledge of LLMs to comprehend exercises. Furthermore, under the same conditions with 100 learners, the agent with GPT-4 is more accurate than the one with GPT-3.5-turbo, indicating that GPT-4 has a stronger semantic understanding ability compared to GPT-3.5-turbo.
Zero-shot Simulation
Zero-shot simulation presents a significant challenge in real-world applications, particularly when learners are in a cold-start situation where their response data is unavailable. This limitation restricts the applicability of previous simulation models. To validate the zero-shot simulation capability of the agent, we initialize 10 agents with randomly generated profiles and have them sequentially answer 10 randomly selected exercises with IDs . In this zero-shot scenario, we disable the corrective reflection mechanism and tools, due to the absence of learner response data. The summary reflection remains usable for the agent. Three GPT-3.5-turbo models, with a temperature parameter of 0.5, act as annotators, tasked with evaluating whether each simulated record (including exercise answers and practice summaries) is written by a real human. Records deemed to be human-written are labeled as “Agent4Edu Win”, non-human records are labeled as “Lose”, and ambiguous records are labeled as “Tie”.
The results depicted in Figure 2 (b) indicate that the agent’s performance in summarization is closely aligned with the real human responses, making differentiation between the two challenging. However, the agent exhibits certain limitations in answering exercise tasks compared to summarization tasks, primarily due to the complexity of reasoning required to solve exercises.
Ablation Study
We conduct ablation studies to evaluate the impact of key components within the GPT-3.5-turbo-powered agent. The results illustrated in Figure 2 (c) show the accuracy of the agent’s exercise-related knowledge prediction and response prediction under various conditions: without the profile module (w/o prof), without the memory module (w/o mem), without the memory enhancement (w/o enh), without the memory forgetting (w/o fgt), and without reflection (w/o ref). These findings confirm the effectiveness of each component in improving the agent’s predictive performance on learners’ response data. However, the ablation experiments indicate that the impact on knowledge prediction is not significant. This can be attributed to the fact that the original GPT-3.5-turbo model already possesses a substantial amount of knowledge, which is sufficient to support exercise comprehension.
4.2 Personalized Learning
Motivation
The primary objective of Agent4Edu is to comprehensively and accurately evaluate personalized learning algorithms and use the generated data to enhance their effectiveness. We aim to validate this objective from two perspectives: (1) through an agent-based multifaceted evaluation of personalized learning services, and (2) by assessing the potential improvements in personalized learning algorithms based on the simulated data.
| Model | satisfaction | AoD | gain |
|---|---|---|---|
| FSI | 39 | 70 | 48 |
| KLI | 39 | 43 | |
| MAAT | 42 | 68 | 45 |
Multifaceted Evaluation
Human learners have multifaceted evaluations of different personalized learning services, such as whether the recommended task difficulty is too challenging. Assuming that a generative agent can accurately simulate the behavior of real learners, its evaluation of personalized algorithms tends to align with human evaluations. We utilize the Computerized Adaptive Testing (CAT) which aims to estimate learners’ ability or knowledge proficiency with minor exercises, as the experimental environment, including FSI (Lord 2012), KLI (Chang and Ying 1996), and MAAT (Bi et al. 2020). We use 100 randomly initialized agents to generate virtual data for pretraining the cognitive diagnosis model (i.e., the IRT model (Baker 2001)) in the CAT algorithm for learner evaluation. Based on this, each adaptive algorithm iterates through 10 rounds to recommend 100 randomly initialized agents (zero-shot simulation), with one exercise recommendation per round. Upon the conclusion of personalized testing, each agent is required to evaluate each CAT algorithm. To achieve this, we design three evaluation metrics, including satisfaction, appropriateness of difficulty (AoD), and whether there was any gain. Table 3 presents a comprehensive evaluation of various strategies, where the element in the -th row and -th column represents the number of agents that consider the corresponding CAT algorithm to meet the metric . Clearly, the agent demonstrates higher satisfaction in recommending MAAT. This observation aligns with the common understanding in the research community that MAAT considers both the difficulty of exercises and the diversity of knowledge (Bi et al. 2020), making the overall service more reasonable. Additionally, FSI focuses on recommending exercises that are moderately difficult and likely to provide gain. These findings highlight the LLM-powered agent’s fine-grained evaluation level for learning algorithms.
Personalized Learning Algorithm Improvement
We investigate whether the simulated data generated by Agent4Edu can enhance personalized learning algorithms. We select CAT as our personalized learning assessment task due to their representativeness in intelligent education. If the generated data can improve the performance of CAT models, it will indicate the effectiveness of our proposed Agent4Edu.
To set up, we select 60% of the learners’ data from EduData to train the cognitive diagnosis models (i.e., the IRT model) for learner evaluation in CAT algorithms (i.e., FSI (Lord 2012), KLI (Chang and Ying 1996), and MAAT (Bi et al. 2020)). The remaining 40% of learners’ data is used to test the CAT models. Furthermore, for each learner in the test data, we simulate their responses to 20 randomly selected unseen exercises based on their profiles. Using this strategy, we generate simulated learner data, which are then merged with the training data from the original EduData to form the augmented dataset, EduData+. We train the IRT model in each CAT model using both the original EduData and EduData+, and then evaluate each CAT strategy by recommending 5 and 10 test exercises for each learner.
Table 4 lists the IRT prediction performance after retraining on the testing records via CAT, where F1-score represents scores on EduData, and F1-score+ represents scores on EduData+. The results demonstrate that CAT strategies can be effectively enhanced with the assistance of Agent4Edu. This suggests that Agent4Edu is capable of generating high-quality learner response data, even with randomly initialized agents (in zero-shot scenarios), thereby enriching the provided dataset.
| Testing length is 5 | Testing length is 10 | |||||
|---|---|---|---|---|---|---|
| Model | F1-score | F1-score+ | Imp. | F1-score | F1-score+ | Imp. |
| FSI | 80.11 | 82.39 | +2.28 | 81.10 | 82.51 | +1.41 |
| KLI | 79.45 | 81.84 | +2.39 | 80.63 | 82.82 | +2.19 |
| MAAT | 81.77 | 81.97 | +0.20 | 81.71 | 81.88 | +0.17 |
5 Conclusion
In this paper, we introduce Agent4Edu, an innovative personalized learning simulator that leverages LLM-powered generative agents to simulate learners’ response data, as well as detailed problem-solving behaviors. Our generative agents are equipped with learner Profile, Memory and Action modules specifically tailored for personalized learning scenarios. These agents exhibit human-like choosing, understanding, analyzing and answering exercises, which accurately predict their future responses. Additionally, the generative agent can interact with personalized learning environments to evaluate and enhance intelligent services. Through comprehensive and meticulous evaluation, we explore the strengths and weaknesses of Agent4Edu, emphasizing the consistency and discrepancies in practice behaviors observed between agents and learners. In the future, we plan to research multi-learner agent cooperation and multi-modal practice solutions using generative agents. We hope that our research will provide new insights into the field of intelligent education.
Acknowledge
This research was partially supported by grants from the National Natural Science Foundation of China (No.62337001, 62477044), the Key Technologies R & D Program of Anhui Province (No. 202423k09020039), and the Fundamental Research Funds for the Central Universities.
References
- Atkinson (1968a) Atkinson, R. C. 1968a. Human memory: A proposed system and its control processes. The Psychology of Learning and Motivation, 2.
- Atkinson (1968b) Atkinson, R. C. 1968b. A proposed system and its control processes. The Psychology of Learning and Motivation, 2.
- Averell and Heathcote (2011) Averell, L.; and Heathcote, A. 2011. The form of the forgetting curve and the fate of memories. Journal of mathematical psychology, 55(1): 25–35.
- Baidoo-Anu and Ansah (2023) Baidoo-Anu, D.; and Ansah, L. O. 2023. Education in the era of generative artificial intelligence (AI): Understanding the potential benefits of ChatGPT in promoting teaching and learning. Journal of AI, 7(1): 52–62.
- Baker (2001) Baker, F. B. 2001. The basics of item response theory. ERIC.
- Bi et al. (2020) Bi, H.; Ma, H.; Huang, Z.; Yin, Y.; Liu, Q.; Chen, E.; Su, Y.; and Wang, S. 2020. Quality meets diversity: A model-agnostic framework for computerized adaptive testing. In 2020 IEEE International Conference on Data Mining (ICDM), 42–51. IEEE.
- Brown et al. (2020) Brown, T.; Mann, B.; Ryder, N.; Subbiah, M.; Kaplan, J. D.; Dhariwal, P.; Neelakantan, A.; Shyam, P.; Sastry, G.; Askell, A.; et al. 2020. Language models are few-shot learners. Advances in neural information processing systems, 33: 1877–1901.
- Chang and Ying (1996) Chang, H.-H.; and Ying, Z. 1996. A global information approach to computerized adaptive testing. Applied Psychological Measurement, 20(3): 213–229.
- Chen et al. (2024) Chen, X.; Wu, L.; Liu, F.; Chen, L.; Zhang, K.; Hong, R.; and Wang, M. 2024. Disentangling Cognitive Diagnosis with Limited Exercise Labels. Advances in Neural Information Processing Systems, 36.
- Cheng et al. (2024) Cheng, K.; Peng, L.; Wang, P.; Ye, J.; Sun, L.; and Du, B. 2024. DyGKT: Dynamic Graph Learning for Knowledge Tracing. In Proceedings of the 30th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, 409–420.
- Cowan (2008) Cowan, N. 2008. What are the differences between long-term, short-term, and working memory? Progress in brain research, 169: 323–338.
- Dan et al. (2023) Dan, Y.; Lei, Z.; Gu, Y.; Li, Y.; Yin, J.; Lin, J.; Ye, L.; Tie, Z.; Zhou, Y.; Wang, Y.; et al. 2023. Educhat: A large-scale language model-based chatbot system for intelligent education. arXiv preprint arXiv:2308.02773.
- Gao et al. (2023) Gao, C.; Lan, X.; Lu, Z.; Mao, J.; Piao, J.; Wang, H.; Jin, D.; and Li, Y. 2023. S3: Social-network Simulation System with Large Language Model-Empowered Agents. arXiv preprint arXiv:2307.14984.
- Gao et al. (2021) Gao, W.; Liu, Q.; Huang, Z.; Yin, Y.; Bi, H.; Wang, M.-C.; Ma, J.; Wang, S.; and Su, Y. 2021. RCD: Relation map driven cognitive diagnosis for intelligent education systems. In Proceedings of the 44th international ACM SIGIR conference on research and development in information retrieval, 501–510.
- Huang et al. (2023) Huang, X.; Lian, J.; Lei, Y.; Yao, J.; Lian, D.; and Xie, X. 2023. Recommender ai agent: Integrating large language models for interactive recommendations. arXiv preprint arXiv:2308.16505.
- Huang et al. (2020) Huang, Z.; Liu, Q.; Chen, Y.; Wu, L.; Xiao, K.; Chen, E.; Ma, H.; and Hu, G. 2020. Learning or forgetting? a dynamic approach for tracking the knowledge proficiency of students. ACM Transactions on Information Systems (TOIS), 38(2): 1–33.
- Huang et al. (2019) Huang, Z.; Liu, Q.; Zhai, C.; Yin, Y.; Chen, E.; Gao, W.; and Hu, G. 2019. Exploring multi-objective exercise recommendations in online education systems. In Proceedings of the 28th ACM International Conference on Information and Knowledge Management, 1261–1270.
- Jin et al. (2023) Jin, D.; Mehri, S.; Hazarika, D.; Padmakumar, A.; Lee, S.; Liu, Y.; and Namazifar, M. 2023. Data-efficient alignment of large language models with human feedback through natural language. arXiv preprint arXiv:2311.14543.
- Kieser et al. (2023) Kieser, F.; Wulff, P.; Kuhn, J.; and Küchemann, S. 2023. Educational data augmentation in physics education research using ChatGPT. Physical Review Physics Education Research, 19(2): 020150.
- Li et al. (2024) Li, H.; Xu, T.; Zhang, C.; Chen, E.; Liang, J.; Fan, X.; Li, H.; Tang, J.; and Wen, Q. 2024. Bringing generative AI to adaptive learning in education. arXiv preprint arXiv:2402.14601.
- Li et al. (2023) Li, Y.; Chen, X.; Zhao, H.; Gong, J.; Zhou, G.; Rossano, F.; and Zhu, Y. 2023. Understanding Embodied Reference with Touch-Line Transformer. In ICLR.
- (22) Liu, J.; Huang, Z.; Xiao, T.; Sha, J.; Wu, J.; Liu, Q.; Wang, S.; and Chen, E. ???? SocraticLM: Exploring Socratic Personalized Teaching with Large Language Models. In The Thirty-eighth Annual Conference on Neural Information Processing Systems.
- Liu et al. (2019) Liu, Q.; Tong, S.; Liu, C.; Zhao, H.; Chen, E.; Ma, H.; and Wang, S. 2019. Exploiting cognitive structure for adaptive learning. In Proceedings of the 25th ACM SIGKDD international conference on knowledge discovery & data mining, 627–635.
- Liu et al. (2023) Liu, R.; Yang, R.; Jia, C.; Zhang, G.; Zhou, D.; Dai, A. M.; Yang, D.; and Vosoughi, S. 2023. Training socially aligned language models in simulated human society. arXiv preprint arXiv:2305.16960.
- Long et al. (2024) Long, X.; Zeng, J.; Meng, F.; Ma, Z.; Zhang, K.; Zhou, B.; and Zhou, J. 2024. Generative multi-modal knowledge retrieval with large language models. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 38, 18733–18741.
- Lord (2012) Lord, F. M. 2012. Applications of item response theory to practical testing problems. Routledge.
- Matelsky et al. (2023) Matelsky, J. K.; Parodi, F.; Liu, T.; Lange, R. D.; and Kording, K. P. 2023. A large language model-assisted education tool to provide feedback on open-ended responses. arXiv preprint arXiv:2308.02439.
- Ouyang et al. (2022) Ouyang, L.; Wu, J.; Jiang, X.; Almeida, D.; Wainwright, C.; Mishkin, P.; Zhang, C.; Agarwal, S.; Slama, K.; Ray, A.; et al. 2022. Training language models to follow instructions with human feedback. Advances in neural information processing systems, 35: 27730–27744.
- Pandey and Karypis (2019) Pandey, S.; and Karypis, G. 2019. A self-attentive model for knowledge tracing. In 12th International Conference on Educational Data Mining, EDM 2019, 384–389. International Educational Data Mining Society.
- Park et al. (2023) Park, J. S.; O’Brien, J.; Cai, C. J.; Morris, M. R.; Liang, P.; and Bernstein, M. S. 2023. Generative agents: Interactive simulacra of human behavior. In Proceedings of the 36th Annual ACM Symposium on User Interface Software and Technology, 1–22.
- Piech et al. (2015) Piech, C.; Bassen, J.; Huang, J.; Ganguli, S.; Sahami, M.; Guibas, L. J.; and Sohl-Dickstein, J. 2015. Deep knowledge tracing. Advances in neural information processing systems, 28.
- Qadir (2023) Qadir, J. 2023. Engineering education in the era of ChatGPT: Promise and pitfalls of generative AI for education. In 2023 IEEE Global Engineering Education Conference (EDUCON), 1–9. IEEE.
- Qian et al. (2023) Qian, C.; Cong, X.; Yang, C.; Chen, W.; Su, Y.; Xu, J.; Liu, Z.; and Sun, M. 2023. Communicative agents for software development. arXiv preprint arXiv:2307.07924.
- Rahman and Watanobe (2023) Rahman, M. M.; and Watanobe, Y. 2023. ChatGPT for education and research: Opportunities, threats, and strategies. Applied Sciences, 13(9): 5783.
- Reddy, Levine, and Dragan (2017) Reddy, S.; Levine, S.; and Dragan, A. 2017. Accelerating human learning with deep reinforcement learning. In NIPS workshop: teaching machines, robots, and humans.
- Su et al. (2018) Su, Y.; Liu, Q.; Liu, Q.; Huang, Z.; Yin, Y.; Chen, E.; Ding, C.; Wei, S.; and Hu, G. 2018. Exercise-enhanced sequential modeling for student performance prediction. In Proceedings of the AAAI conference on artificial intelligence, volume 32.
- Wang et al. (2024a) Wang, F.; Gao, W.; Liu, Q.; Li, J.; Zhao, G.; Zhang, Z.; Huang, Z.; Zhu, M.; Wang, S.; Tong, W.; et al. 2024a. A Survey of Models for Cognitive Diagnosis: New Developments and Future Directions. arXiv preprint arXiv:2407.05458.
- Wang et al. (2023a) Wang, F.; Huang, Z.; Liu, Q.; Chen, E.; Yin, Y.; Ma, J.; and Wang, S. 2023a. Dynamic cognitive diagnosis: An educational priors-enhanced deep knowledge tracing perspective. IEEE Transactions on Learning Technologies, 16(3): 306–323.
- Wang et al. (2023b) Wang, G.; Xie, Y.; Jiang, Y.; Mandlekar, A.; Xiao, C.; Zhu, Y.; Fan, L.; and Anandkumar, A. 2023b. Voyager: An open-ended embodied agent with large language models. arXiv preprint arXiv:2305.16291.
- Wang et al. (2024b) Wang, L.; Ma, C.; Feng, X.; Zhang, Z.; Yang, H.; Zhang, J.; Chen, Z.; Tang, J.; Chen, X.; Lin, Y.; et al. 2024b. A survey on large language model based autonomous agents. Frontiers of Computer Science, 18(6): 1–26.
- Wang et al. (2023c) Wang, L.; Zhang, J.; Chen, X.; Lin, Y.; Song, R.; Zhao, W. X.; and Wen, J.-R. 2023c. Recagent: A novel simulation paradigm for recommender systems. arXiv preprint arXiv:2306.02552.
- Wang et al. (2023d) Wang, L.; Zhang, J.; Yang, H.; Chen, Z.; Tang, J.; Zhang, Z.; Chen, X.; Lin, Y.; Song, R.; Zhao, W. X.; et al. 2023d. User behavior simulation with large language model based agents. arXiv preprint arXiv:2306.02552.
- Wei et al. (2022) Wei, J.; Wang, X.; Schuurmans, D.; Bosma, M.; Xia, F.; Chi, E.; Le, Q. V.; Zhou, D.; et al. 2022. Chain-of-thought prompting elicits reasoning in large language models. Advances in neural information processing systems, 35: 24824–24837.
- Wu et al. (2023) Wu, Q.; Bansal, G.; Zhang, J.; Wu, Y.; Zhang, S.; Zhu, E.; Li, B.; Jiang, L.; Zhang, X.; and Wang, C. 2023. Autogen: Enabling next-gen llm applications via multi-agent conversation framework. arXiv preprint arXiv:2308.08155.
- Xu, Zhang, and Qin (2024) Xu, S.; Zhang, X.; and Qin, L. 2024. EduAgent: Generative Student Agents in Learning. arXiv preprint arXiv:2404.07963.
- Yao et al. (2024) Yao, F.; Liu, Q.; Yue, L.; Gao, W.; Li, J.; Li, X.; and He, Y. 2024. Adard: An adaptive response denoising framework for robust learner modeling. In Proceedings of the 30th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, 3886–3895.
- Yue et al. (2023) Yue, L.; Liu, Q.; Du, Y.; Gao, W.; Liu, Y.; and Yao, F. 2023. Fedjudge: Federated legal large language model. arXiv preprint arXiv:2309.08173.
- Zhang et al. (2023a) Zhang, A.; Sheng, L.; Chen, Y.; Li, H.; Deng, Y.; Wang, X.; and Chua, T.-S. 2023a. On generative agents in recommendation. arXiv preprint arXiv:2310.10108.
- Zhang et al. (2023b) Zhang, J.; Hou, Y.; Xie, R.; Sun, W.; McAuley, J.; Zhao, W. X.; Lin, L.; and Wen, J.-R. 2023b. Agentcf: Collaborative learning with autonomous language agents for recommender systems. arXiv preprint arXiv:2310.09233.
- Zhang et al. (2017) Zhang, J.; Shi, X.; King, I.; and Yeung, D.-Y. 2017. Dynamic key-value memory networks for knowledge tracing. In Proceedings of the 26th international conference on World Wide Web, 765–774.
- Zhao et al. (2023) Zhao, G.; Huang, Z.; Zhuang, Y.; Liu, J.; Liu, Q.; Liu, Z.; Wu, J.; and Chen, E. 2023. Simulating Student Interactions with Two-stage Imitation Learning for Intelligent Educational Systems. In Proceedings of the 32nd ACM International Conference on Information and Knowledge Management, 3423–3432.
- Zhuang et al. (2024) Zhuang, Y.; Liu, Q.; Zhao, G.; Huang, Z.; Huang, W.; Pardos, Z.; Chen, E.; Wu, J.; and Li, X. 2024. A Bounded Ability Estimation for Computerized Adaptive Testing. Advances in Neural Information Processing Systems, 36.