1. National Key Laboratory of Oto-Electronic Information Acquisition and Protection Technology, Anhui University, Hefei 230601, China
2. School of Computer Science and Technology, Anhui University, Hefei 230601, China
xyzhanghust@gmail.com
Show less
History+
Received
Accepted
Published
2025-01-07
2025-03-03
2025-06-25
Issue Date
Revised Date
2025-03-17
2025-02-17
PDF
(4158KB)
Abstract
With the rapid development of online education, cognitive diagnosis has become a key task in intelligent education, particularly for student ability assessments and resource recommendations. However, existing cognitive diagnosis models face the diagnostic system cold-start problem, whereby there are no response logs in new domains, making accurate student diagnosis challenging. This research defines this task as zero-shot cross-domain cognitive diagnosis (ZCCD), which aims to diagnose students’ cognitive abilities in the target domain using only the response logs from the source domain without prior interaction data. To address this, a novel paradigm, large language model (LLM)-guided cognitive state transfer (LCST) is proposed, which leverages the powerful capabilities of LLMs to bridge the gap between the source and target domains. By modelling cognitive states as natural language tasks, LLMs act as intermediaries to transfer students’ cognitive states across domains. The research uses advanced LLMs to analyze the relationships between knowledge concepts and diagnose students’ mastery of the target domain. The experimental results on real-world datasets shows that the LCST significantly improves cognitive diagnostic performance, which highlights the potential of LLMs as educational experts in this context. This approach provides a promising direction for solving the ZCCD challenge and advancing the application of LLMs in intelligent education.
Haiping Ma, Changqian Wang, Siyu Song, Shangshang Yang, Limiao Zhang, Xingyi Zhang.
Large Language Models Are Zero-Shot Cross-Domain Diagnosticians in Cognitive Diagnosis.
Frontiers of Digital Education, 2025, 2(2): 17 DOI:10.1007/s44366-025-0054-y
In recent years, with the development of online education, cognitive diagnosis has become an important research task in the application of intelligent education (Liu et al., 2023; Yang et al., 2024b; Yang et al., 2024c). Cognitive diagnosis is widely used in student ability assessments and resource recommendations (Yu et al., 2024a; Yu et al., 2024b), with its goal being to diagnose students’ mastery of knowledge concepts associated with exercises through a large number of students‒exercise response logs, which facilitates personalized learning (Ma et al., 2024a; Zhang et al., 2024). The existing work is primarily based on research within well-established knowledge domains with the available response logs of students and exercises. However, for knowledge concepts in a new subject (hereinafter referred to as a domain) without response logs, existing methods struggle to provide accurate diagnoses of students. Nevertheless, research on the cold-start problem of diagnostic systems in the domain with no interaction logs is relatively scarce. As shown in Fig.1, the mathematics domain encompasses a vast number of knowledge concepts. Existing cognitive diagnosis models (CDMs) analyze exercises associated with these knowledge concepts and a large set of student responses to generate detailed diagnostic feedback for each student. For instance, the value of the diagnostic feedback indicates the student’s level of mastery of these concepts, including the system of linear equations in two variables presenting 0.4, quadratic radical equation presenting 0.6, and inverse proportionality function presenting 0.3. Mastery is represented as a value between 0 and 1, where 0 implies no understanding of the concept and 1 signifies complete mastery of the ability to solve related problems with minimal errors. However, when faced with a lack of response records for knowledge concepts in the physics domain, it is difficult to accurately diagnose students’ cognitive states in that domain. This is because although students can be asked to perform exercises from the target domain with no prior interaction records, the monotonous task of performing these exercises may lead to a subjective decline in their cognitive abilities. Moreover, when the set of knowledge concepts in the target domain is large or there are no corresponding exercises, diagnosing through the performance of exercises is not a feasible strategy. Furthermore, from a model perspective, CDMs rely on analyzing students’ response records within a specific domain to generate diagnostic feedback, which implies that in the absence of response data for a given subject, the CDM would no longer function effectively. This task is defined in this research as zero-shot cross-domain cognitive diagnosis (ZCCD), which refers to diagnosing students’ cognitive abilities in a target domain with completely blank response records, using only response logs from a source domain. ZCCD is a practical and meaningful task (Bai et al., 2025; Ma et al., 2024a). A promising approach to address this task is to build a bridge between sources and target domains and to leverage the relationships between knowledge concepts across similarity and precedence domains. This method allows students’ cognitive states in the source domain to be transferred in a zero-shot manner to the target domain, therefore enabling the diagnosis of their mastery of knowledge concepts in the target domain. To address this issue, TechCD propagates students’ cognitive states by pre-constructing knowledge concept graphs (KCGs) (Gao et al., 2023). However, the construction of KCGs depends heavily on information-rich interaction datasets and significant human efforts. Similarly, Zero-1-to-3 requires a substantial amount of response data from a subset of students in the target domain to propagate those students’ cognitive states to cold-start students (Gao et al., 2024a). Both approaches rely on extensive prior information, which constrains their applicability in real-world scenarios.
Recently, the pretrained large language models (LLMs) have demonstrated remarkable performance in general tasks across both natural language processing (NLP) and recommendation systems (Agostinelli et al., 2023; Brown et al., 2020; Hou et al., 2024; Kojima et al., 2022; Li et al., 2024; OpenAI et al., 2023; Wang et al., 2024a; Wang et al., 2024b; Yang et al., 2023a). Trained on vast and diverse textual corpora, LLMs not only exhibit exceptional language comprehension but also possess strong reasoning and generative capabilities which enable them to capture intricate semantic structures and underlie logical relationships within text. One of the intuitive advantages of LLMs lies in their ability to emulate human thought processes. The extensive knowledge acquired during pretraining allows LLMs to generate intermediate reasoning steps, and these steps resemble the analytical processes employed by educational experts, thereby providing an additional interpretative layer for diagnostic outcomes. For instance, when evaluating the relationship between force and equilibrium force or between reflection and refraction in physics, LLMs offer detailed explanations that reflect the underlying physical principles, intuitively illustrating cross-domain knowledge dependencies. A key challenge in addressing ZCCD is identifying an intermediary that bridges the source and target domains, which necessitates an understanding of both interdisciplinary relationships and the characteristics of students’ cognitive states. Leveraging their inherent world knowledge, LLMs offer a promising solution for mitigating the semantic gap between these domains. LLMs have shown exceptional summarization and reasoning capabilities in recommendation systems (Jiang et al., 2024; Ren et al., 2024). This is because after modelling user behavioural preferences as natural language descriptions, the language model understands the instructions and further completes the recommendation task. Inspired by this, this research views the cognitive modelling of students in the source domain as a language modelling task. After acquiring the student’s prior cognitive state, the diagnostic task is formatted into a natural language text, which is then treated as a language-processing task. Therefore, LLMs cannot only serve as intermediaries between the source and target domains but also replace the CDM in the target domain. To this end, a novel diagnostic paradigm for the ZCCD task is proposed to integrate the cognitive diagnosis task with LLMs for the first time. This paradigm is called LLM-guided cognitive state transfer (LCST), which reexamines students’ cognitive states from a natural language perspective. Llama 3.2, Gemma, and other advanced LLMs are chosen in this research, as educational experts for analyzing the internal relationships among knowledge concepts and diagnosing students’ cognitive states in the target domain due to their accessibility and high performance. Specifically, to obtain students’ prior abilities, the abundant response data from the source domain are leveraged. Using mainstream CDMs, this research diagnoses the students’ mastery of knowledge concepts in the source domain, which serves as the primary input for the target domain diagnostic feedback. Moreover, this research integrates students’ cognitive states from the source domain, formats their prior abilities into natural language descriptions, and designs appropriate prompt templates to enable LLMs to function as intermediaries between the source and target domains, which unlocks students’ potential in cognitive diagnosis tasks. To verify LLMs’ performance as educational experts in diagnostic tasks, the research formats the diagnostic feedback as proficiency parameters and reintegrates them into the CDM’s parameter set for diagnosing prior cognitive states. Extensive experiments are conducted on real-world datasets. The experimental results demonstrate a significant performance improvement achieved by the LCST and reveal the enormous potential of LLMs in cognitive diagnosis tasks, which offers a promising research direction.
Based on the previous research, this research propels further contributions from three aspects. First, this research identifies the real-world ZCCD task and combines cognitive diagnosis tasks with LLMs to propose the novel LCST diagnostic paradigm for the first time, which addresses the cold-start challenge in zero-shot cross-domain diagnosis scenarios. Second, given the outstanding summarization and recommendation performance of LLMs in other domains, the LCST utilizes LLMs as the cognitive diagnosis module for the target domain. Based on prior knowledge, it not only bridges the source and target domains but also enables the functionality of mainstream CDMs. Third, the research conducts extensive experiments on real-world datasets to validate the effectiveness of the proposed LCST paradigm in solving diagnostic tasks for cold-start domains.
2 Related Work
2.1 Cognitive Diagnosis
Cognitive diagnosis is a fundamental and crucial task in the field of education that aims to infer students’ mastery of knowledge concepts (Gao et al., 2024b; Ma et al., 2024a; Ma et al., 2024b; Yang et al., 2024a; Yang et al., 2023b; Yang et al., 2023c). Existing cognitive diagnostic methods are based on available student response logs and focus on modelling interaction behaviours. For instance, disentangling cognitive diagnosis improves diagnostic performance in scenarios with limited exercise labels by utilizing student response records to model students’ abilities, exercise difficulties, and label distributions (Chen et al., 2024). The oversmoothing-resistant cognitive diagnosis framework designs a novel perceptual graph convolutional network to capture key response signals in interaction behaviours, which alleviates the over-smoothing problem and enhances the diagnostic capability of the model (Qian et al., 2024). Agent4Edu introduces a personalized learning simulator that leverages LLM-driven generative agents to simulate real learner behaviour (Gao et al., 2025), which aims to bridge the gap between offline evaluation metrics and real-world online performance in personalized learning. LLM-based question generation utilizes LLMs to generate questions aligned with Bloom’s taxonomy learning objectives (Elkins et al., 2024), which ensures a more pedagogically grounded approach to question creation. While these works have made significant progress, they are all based on rich student response records. When faced with a blank domain lacking student response logs, these methods struggle to function effectively. As a solution, TechCD uses preconstructed KCGs to transfer students’ cognitive states from existing domains to the cold-start target domain (Gao et al., 2023). Zero-1-to-3 generates simulated practice logs for cold-start students by analyzing the behaviour patterns of early-bird students and refining their cognitive states using virtual data to mitigate cold-start challenges (Gao et al., 2024a). However, these approaches rely on the availability of student response records in the target domain, which are difficult to obtain and costly in real-world scenarios, to bridge the source and target domains. Therefore, this research proposes an alternative approach that leverages the advantages of LLMs to build a bridge between source and target domains and extend further to replace the cognitive diagnostic models for the target domain.
2.2 Large Language Models
Recently, LLMs have demonstrated remarkable capabilities in solving various tasks in the field of NLP (Mesnard et al., 2024; Touvron et al., 2023a; Touvron et al., 2023b). Thanks to self-supervised training on vast datasets and the extensive absorption of knowledge from various domains (Grattafiori et al., 2024), LLMs are capable of solving specific tasks through tailored prompts. As a result, the growing interest in leveraging these advanced models addresses traditional tasks. For instance, the personalized prompt for sequential recommendation effectively enhances cold-start recommendations by constructing personalized soft prompts based on user profiles (Wu et al., 2024). The spatiotemporal graph transfer learning unifies different tasks into a single template and employs a two-stage prompt pipeline with learnable prompts to achieve domain and task transfer (Hu et al., 2024), enabling the prompts to effectively capture domain knowledge and task-specific attributes. Moreover, in the medical field, the Prompt-Eng designed precise prompts that included both positive and negative aspects and highlighted that designing paired prompts helped the model generalize effectively (Ahmed et al., 2024). Despite the impressive results achieved by LLMs across many domains, no work explored the integration of LLMs with cognitive diagnosis tasks.
3 Problem Statements
The formatting for the ZCCD task is conducted as follows. The available M mature source domains are designated as S1, S2, ..., SM, and the target domain, which requires cold-start initialization, is designated as T. The overlapping set of students shared between the m-th source domain and target domain is denoted as Um. In the m-th source domain, the sets of exercises and knowledge concepts are represented as and , respectively, while in the target domain, they are represented as ET and KT. Notably, and . The student‒exercise response records in the m-th source domain Sm are recorded as = {(ui, ej, rij) | rij ∈ {0, 1}, ui ∈ Um, ej ∈ }, where rij = 1 indicates that student ui answered exercise ej correctly, otherwise rij = 0. Problem definition can be recorded RS = {R1, R2, ..., RM} given the student’s response in the source domains and the set of knowledge concepts KT in the target domain. The goal of the research is to diagnose the potential mastery of knowledge concepts in the target domain for students with blank interaction records.
4 Proposed LCST Diagnostic Paradigm
4.1 Overview
To transfer students’ cognitive states from the source domain to the blank target domain, as illustrated in Fig.2, the LCST diagnostic paradigm is proposed, which combines traditional CDMs with LLMs. The proposed LCST comprises four modules, including the pre-established cognitive state, bridge source and target domain, cognitive diagnosis of the target domain, and feedback constraint. In the pre-established cognitive state module, given rich response records in the source domain, the LCST utilizes a traditional CDM to pre-diagnose students’ mastery of knowledge concepts in the source domain. This information is then transformed into a natural language form. The zero-shot, few-shot, and chain-of-thought prompts for the bridge source and target domain module are proposed to explore the internal relationships between knowledge concepts. This module takes input from both the source and target domains, including all knowledge concepts and instruction prompts. It is used to activate LLMs’ potential to infer knowledge concept relationships. In the cognitive diagnosis of the target domain module, the information from the previous two modules is integrated, and the prompts are designed to utilize LLMs’ internal knowledge and pre-acquired prior knowledge to complete the diagnosis task in the target domain. Finally, to ensure that the feedback format is easily assessable, this research incorporates the feedback constraint module into the LCST. Notably, while the CDM in the pre-established cognitive states module requires training, LLMs do not require training or fine-tuning. They only need to be utilized with pretrained models and well-designed prompts for downstream tasks.
4.2 Pre-Established Cognitive States
The cognitive diagnosis task aims to analyze students’ response records and generate diagnostic feedback. Traditional CDMs are only applied in source domains with rich response logs. Therefore, when faced with a zero-shot target domain, the diagnostic process becomes challenging. LLMs have great potential for propagating cognitive states from the source domain to the target domain. However, due to input constraints and the sensitivity of educational contexts, it is not feasible to use all response logs from the source domain as input directly. The focus of this phase is on obtaining the cognitive states of students in the source domain, which lays the foundation for subsequent state transfer. To achieve this goal, this research leverages the available response data from the source domain and combines it with cognitive diagnostic models such as neural cognitive diagnosis (NCD) and knowledge-association based extension of NCD, to pre-diagnose students’ cognitive states in the source domain.
Specifically, for student ui ∈ Um in the m-th source domain Sm, given their response record set within that domain, the domain-specific CDM Mm will establish their cognitive state, which is expressed as follows:
where the proficiency vector represents the student’s mastery of all knowledge concepts in the source domain Sm, where indicates the student’s proficiency in the n-th knowledge concept, with values ranging from 0 to 1. represents the loss function used to pretrain the model Mm, typically, the cross-entropy loss.
Next, the parameterized proficiency values are stored in a set, which serves as a precursor for subsequent natural language formatting, as follows:
where denotes the set of proficiency values for student ui in the m-th source domain, and kn represents the name of the n-th knowledge concept in source domain Sm.
4.3 Bridge Source and Target Domain
To bridge the gap between the source and target domains, previous works have utilized preconstructed knowledge concept maps. However, this approach relies on both source and target domain response logs and incurs high labour costs. Therefore, this research proposes directly leveraging LLMs as educational experts to analyze two types of relationships between knowledge concepts, including hierarchical relationships and similarity relationships. Specifically, a specialized prompt template is designed to investigate the reasoning capabilities of LLMs regarding the internal relationships among knowledge concepts. The template consists of three components, including task description, knowledge concept injection, and format indicator. The task description is used to adapt the concept reasoning task to a natural language-processing task. The knowledge concept injection aims to provide a range of knowledge concepts from both domains to be bridged. The format indicator is employed to constrain the output format, which makes the reasoning results easier to integrate into cognitive diagnostic models based on knowledge concept relationships. This task is defined as concept reasoning, and an example of such a prompt for the smart learning partner (SLP) physics dataset is shown in Fig.3.
To further stimulate the internal knowledge of LLMs, a few-shot chain-of-thought prompt is injected into the prompt template to enhance their reasoning ability (Wei et al., 2022). By adding expert-analyzed knowledge concept relationships and reasoning justifications, the above template is adjusted as shown in Fig.4.
Compared to zero-shot prompting, injecting a small number of prompts guides the LLMs to perform the reasoning task better, thereby reducing the uncertainty of the generated text and increasing confidence in the results.
4.4 Cognitive Diagnosis of the Target Domain
After obtaining the students’ cognitive abilities in the source domain and establishing a bridge between the source and target domains, the proposed LCST is capable of utilizing this prior knowledge to complete the diagnostic process in the target domain. Given the availability of source domain response logs, the diagnostic task in the target domain is divided into two sub-tasks, including single-domain diagnosis using diagnostic feedback from a single domain as prior knowledge and multi-domain diagnosis using diagnostic feedback from multiple domains as prior knowledge.
Similar prompt templates for these two diagnostic tasks are designed to integrate the prior knowledge from the previous modules, which instructs LLMs to analyze and diagnose the students’ potential mastery of specific knowledge concepts. Each prompt includes a task description, a textual representation of the student’s prior cognitive state, a concept reasoning task, and a format indicator. The task description is used to adapt the diagnostic task to a natural-language-processing task. The format indicator is designed to assess the performance of the LLMs in generating results.
For the single-domain diagnosis task, student’s prior cognitive state is injected Cm from the m-th source domain, the set of knowledge concepts in that domain K and the knowledge concepts of the target domain KS, and the knowledge concepts of the target domain KT into the prompt template, as shown in Fig.5.
For the multi-domain diagnosis task, minor adjustments are made to the above template by replacing the prior cognitive state where represents the number of available source domains and the knowledge concept set for the source domains in the template, as shown in Fig.6.
These two tasks are adjusted based on the availability of interaction logs from the source domain. To enhance the credibility and persuasiveness of the diagnostic tasks, the explanation generation commands are injected into the prompt templates of the two tasks. These explanations are intended to clarify why a particular diagnostic conclusion is made for the student’s abilities, and they also contribute to providing personalized learning support for the students. Specifically, LLMs are requested to generate “why” explanation texts and “how” guidance texts for each diagnostic result, thereby ensuring the reasoning and transparency of the diagnostic process.
4.5 Feedback Constraints
To facilitate the evaluation and extraction of the results, format indicators are designed for each of the three tasks to constrain the output format. For the concept reasoning task, the output needs to be constrained into two parts, including pairs of knowledge concepts that exhibit similarity relationships and pairs that exhibit hierarchical relationships. The instructions were added at the end of the concept reasoning task template, as shown in Fig.7.
For the single-domain diagnosis and multi-domain diagnosis tasks, to enhance the transparency and interpretability of the diagnostic process, the format indicator in this section constrains the output into three parts, including the concept relationship analysis, the diagnostic report, and the guidance suggestion. Instructions are added to the end of the single-domain diagnosis and multi-domain diagnosis task templates, respectively, as shown in Fig.8.
Finally, the mastery score is extracted for each knowledge concept from the output, which is guided by carefully designed prompt templates. These scores are then vectorized to form a student ability vector representing the mastery levels of different knowledge concepts in the target domain. The length of this vector remains consistent with the number of knowledge concepts in the target domain.
5 Experiments
Because the key contribution of this work is to diagnose the potential mastery levels of knowledge concepts in the target domain in which interaction records are blank, comprehensive experiments on real-world datasets are conducted to address the following four research questions:
(1) Can the proposed LCST effectively handle the ZCCD task?
(2) Can the LCST effectively utilize information from diverse source domains for the ZCCD task?
(3) How does the LCST perform on the concept reasoning task?
(4) How do different LLMs perform in cross-domain diagnosis?
5.1 Dataset Description
The experiments are conducted on five real-world datasets, including mathematics, physics, chemistry, history, and geography, which are collected from the SLP online education platform (Lu et al., 2021). These datasets summarize students’ academic performance data across different subjects over three-year-study, including their response records, knowledge concept texts, and the relationships between exercises and knowledge concepts. Each subject is treated as a domain, alternating between being the cold-start target domain and other domains that serve as source domains. For each dataset, each student’s response records are divided into training and testing sets, with a 4:1 split. The test set for the target domain is used not only to evaluate the pretraining performance of the cognitive diagnostic model but also to test the performance of baselines in solving the ZCCD task. The statistics of the processed datasets are shown in Tab.1.
5.2 Experimental Setup
5.2.1 Baselines
The Llama 3.21 is the backbone of the LCST. The CDM is first pretrained using the dataset from the target domain, obtaining the model’s network parameters specific to that domain. The student ability component in these network parameters is presented with the vectorized student abilities, after which the cross-domain diagnostic performance is tested on the target domain’s test set. To demonstrate the effectiveness of the proposed LCST in solving the ZCCD task, it is applied to four widely used cognitive diagnostic models, including NCD (Wang et al., 2020), knowledge-sensed cognitive diagnosis (Ma et al., 2022), relation map-driven cognitive diagnosis (RCD) (Gao et al., 2021), and knowledge-association based extension of the NCD (Wang et al., 2022). The random oracle model as a baseline for comparison is selected, where the random oracle method represents the lower and upper bounds of the performance of CDMs in solving the ZCCD task. The random method means that predicts the initial ability of students in the target domain from a uniform (0, 1) distribution randomly, which is the most common method in existing CDMs. The oracle method uses CDM to train the target domain ability from the response records of students in the target domain directly.
5.2.2 Evaluation Metrics
Given that the true knowledge mastery of students is unknown, mainstream methods have been used in previous research to evaluate the effectiveness of CDMs by predicting students’ exercise performance based on acquired knowledge mastery vectors. Since cognitive diagnosis is a binary classification task that predicts whether a student will answer a given question correctly, three well-known metrics are selected to assess prediction performance. First, accuracy serves as an intuitive metric for evaluating overall prediction correctness. This research sets a threshold of 0.5, meaning that if the predicted probability exceeds 0.5, the student is expected to answer correctly. Otherwise, an incorrect response is predicted:
where ACC indicates accuracy, is the predicted binary response, yi is the ground truth response, N denotes the size of the dataset, and is the indicator function. However, in scenarios with imbalanced class distributions, accuracy alone may not provide a comprehensive assessment of model performance. Second, this research incorporates the area under curve (AUC) as an additional evaluation metric (Bradley, 1997). AUC offers a holistic measure of the model’s discriminative ability across different probability thresholds. Third, root mean square error (RMSE) is introduced as an evaluation metric to quantify the deviation between predicted probabilities and actual binary labels (Pei et al., 2017):
where represents the predicted probability of a correct response and N denotes the size of the dataset.
5.2.3 Parameter Settings
In the pretraining phase, this research initializes all parameters in the network using Xavier (Glorot & Bengio, 2010), and uses the Adam optimizer during training (Kingma & Ba, 2014). The batch size is fixed at 32, while the learning rate is fixed at 0.01. Then, the dimensions of the latent features for the students and exercises are set to be equal to the number of knowledge concepts. Finally, this research sets the temperature for all LLMs to 0.2 and runs each task three times, taking the average as the diagnostic feedback. All models are implemented in PyTorch, while all experiments are conducted on a Linux server equipped with a Tesla V100 GPU.
5.3 Overall Performance
To validate the effectiveness of the proposed diagnostic paradigm in solving the ZCCD task, the LCST is compared with other baselines by using single-domain and multi-domain information as prior knowledge.
5.3.1 Performance on the Single-Domain Diagnosis Task
First, this research explores the possibility of using a single domain as the source domain to solve the ZCCD task, which includes four cross-domain scenarios. This research alternates each dataset to serve as the target domain, with other datasets acting as the source domains. The experimental results shown in Tab.2 demonstrate that the proposed LCST outperforms the widely applied random method in most scenarios and even surpasses oracle method in some cases, achieving the performance upper bound. This illustrates the immense potential of LLMs, powered by their strong reasoning capabilities, in cognitive diagnosis tasks. Furthermore, four aspects of performance are shown. First, in most scenarios, the performance of the LCST is consistently close to the oracle, especially in domains such as mathematics and physics, where there is a clear hierarchical relationship between knowledge concepts. This confirms that by implementing the concept reasoning task, LLMs can enhance their diagnostic capability as educational experts effectively. Second, given that different CDMs model students and exercises, the parameterized diagnostic feedback shows performance fluctuations across various CDMs in the same scenario. Among them, NeuralCD exhibits the least fluctuation, since it does not interfere much with the student’s cognitive state parameters. In contrast, CDMs that employ more complex modelling approaches somewhat distort the parameterized diagnostic feedback but still maintain high performance. Third, at optimal performance, the proposed LCST consistently outperforms TechCD. TechCD addresses the ZCCD challenge by pre-constructing KCGs, which rely on both information-rich interaction datasets and extensive human annotations. In contrast, the LCST leverages the strengths of LLMs, which requires only a set of knowledge concepts within the domain to infer the relationships between them. By eliminating the need for complex model designs, such as those in TechCD, which may reduce interpretability, the LCST offers a more efficient and cost-effective solution while maintaining clarity and ease of implementation. Fourth, compared to mainstream CDMs pretrained on the target domain, the LCST paradigm, which integrates LLMs, is able to better capture students’ mastery of complex knowledge concepts by relying solely on prediagnosed cognitive states without directly receiving students’ response records. This highlights the promising research direction for using LLMs in cognitive diagnosis tasks.
5.3.2 Performance on the Multi-Domain Diagnosis Task
This research further explores the potential of integrating multisource domains as prior information to solve the ZCCD task. This includes three scenarios in which physics, mathematics, and chemistry are used as target domains, while the remaining subjects provided prior information as source domains. From Tab.3, three results can be extracted. First, information from multiple domains provides significantly richer prior knowledge than single-domain information. By offering higher-quality input, LLMs can provide more accurate and reliable student ability predictions. Moreover, having more domains means richer relationships between knowledge concepts, which also helps LLMs deliver more comprehensive diagnostics for students. Second, the knowledge concepts in the three domains exhibited stronger hierarchical and similarity relationships. As a result, the LCST demonstrated performance that not only far exceeded the random method in most scenarios but also slightly surpassed the oracle method. This suggests that domains with closer relationships can more effectively address the cold-start challenge in the blank target domain. Third, LLMs show excellent generalization across different scenarios, indicating that they can effectively adapt to various subjects and knowledge points. Compared to traditional CDM diagnostic paradigms, LLMs are capable of processing more diverse input features, further enhancing their performance in multitasking environments.
5.4 Performance on the Concept Reasoning Task
To explore the performance of the LCST in handling the concept reasoning task, RCD is selected, which involves knowledge concept relationship modelling, as the validation model and tests are conducted in the physics domain. The standard RCD method uses an approach to obtain knowledge concept relationships from student response logs that is referred to as statistics-based. This research utilizes the LCST-enhanced method to model the relationships among knowledge concepts. The experimental results shown in Fig.9 demonstrate that the LCST-enhanced method outperforms the statistics-based method in RCD, surpassing it on both metrics. This proves that the knowledge concept relationships inferred through the LCST are more beneficial for RCD in modelling knowledge concepts and align better with real-world scenarios.
Furthermore, as shown in Fig.10, four reasoning results of both methods regarding knowledge concept relationships can be extracted. First, the LCST-enhanced method infers a broader range of knowledge concept relationships, while the statistics-based method relies on the statistical analysis of existing response logs, leading to the loss of many knowledge concept relationships, which hinders CDM from performing fine-grained modelling. Second, the statistical results of the statistics-based method regarding the similarity relationships between knowledge concepts are biased, particularly in the upper half of the knowledge concepts shown in Fig.10(a). This is because the statistical results are constrained by the distribution of student responses and abilities in the corresponding logs. In contrast, as shown in Fig.10(b), the LCST-enhanced method infers knowledge concept relationships with a more uniform distribution and provides more comprehensive reasoning of hierarchical relationships, proving that the LCST has great potential for handling the concept reasoning task.
5.5 Performance of Different LLMs on Diagnostic Tasks
To explore the performance of different LLMs in cognitive diagnosis tasks, several mainstream open-source LLMs with varying parameter sizes are selected, including Llama3.2:3b, Gemma:7b, Qwen2.5:3b, Hermes3:3b, and Qwen2.5:7b, as the backbone of the LCST. Mathematics as the source domain and physics as the target domain are set, using NeuralCD as the pretrainer to diagnose students’ abilities in the source domain. Two results can be extracted from Tab.4. First, Gemma-7b exhibits the best diagnostic performance, albeit with the longest generation time. In contrast, Hermes3:3b shows worse diagnostic performance than the random method, which may be due to Amnesia Mode leading to unstable generated feedback (Teknium et al., 2024). Second, a larger parameter size does not necessarily indicate better performance for diagnostic tasks. Smaller LLMs may perform better than larger ones in diagnostic tasks. For example, Llama3.2:3b, despite its smaller parameter size, shows near-optimal performance and takes the least amount of time. In terms of balancing performance and efficiency, Llama3.2:3b may be the most suitable backbone for the LCST to address the ZCCD task. This is attributed to the training set and quality of the LLMs, since Qwen2.5-based LLMs, primarily pretrained on Chinese corpora, may not perform optimally under English prompt templates.
6 Discussion
In this work, the LCST method is proposed to address data scarcity in the target domain encountered in ZCCD. The experimental results indicate that the LCST not only significantly enhances diagnostic accuracy across multiple datasets but also, in certain cases, achieves performance on par with oracle methods that rely on target domain data, thereby validating the effectiveness and applicability of LLMs in cross-domain cognitive diagnosis. Nonetheless, the proposed method has several limitations, including high sensitivity to prompt template designs, inadequate coverage of domain-specific knowledge, and a lack of clarity in the model’s internal reasoning processes. Future research will focus on exploring more robust prompt designs, adaptive fine-tuning techniques, and the integration of external knowledge graphs to further improve the model’s ability to capture cross-domain knowledge and enhance diagnostic performance.
7 Conclusions
In this research, the application of LLMs in cognitive diagnosis tasks is explored to diagnose students’ potential mastery of knowledge concepts in a zero-shot cross-domain scenario. A novel diagnostic paradigm, LCST, not only establishes a bridge between the source and target domains but also facilitates the diagnostic task in the target domain. The experimental results on real-world datasets demonstrate the effectiveness of the proposed method. This research is intended to further investigate the application of LLMs tailored to the characteristics of cognitive diagnosis tasks.
Agostinelli, V., Wild, M., Raffel, M., Fuad, K.A.A., Chen, L. (2023). Simul-LLM: A framework for exploring high-quality simultaneous translation with large language models. arXiv Preprint, arXiv:2312.04691.
[2]
Ahmed, A., Hou, M., Xi, R., Zeng, X., Shah, S. A. (2024). Prompt-Eng: Health-care prompt engineering: Revolutionizing healthcare applications with precision prompts. In: Proceedings of the ACM on Web Conference 2024. New York: ACM, 1329–1337.
[3]
Bai, Y., Li, X., Liu, Z., Huang, Y., Guo, T., Hou, M., Xia, F., Luo, W. (2025). csKT: Addressing cold-start problem in knowledge tracing via kernel bias and cone attention.Expert Systems with Applications, 266(25): 125988
[4]
Bradley, A. P. (1997). The use of the area under the ROC curve in the evaluation of machine learning algorithms.Pattern Recognition, 30(7): 1145–1159
[5]
Brown, T., Mann, B., Ryder, N., Subbiah, M., Kaplan, J., Dhariwal, P., Neelakantan, A., Shyam, P., Sastry, G., Askell, A., , . (2020). Language models are few-shot learners. arXiv Preprint, arXiv:2005.14165.
[6]
Chen, X., Wu, L., Liu, F., Chen, L., Zhang, K., Hong, R., Wang, M. (2024). Disentangling cognitive diagnosis with limited exercise labels. In: Proceedings of the 37th International Conference on Neural Information Processing Systems. New York: ACM, 18028–18045.
[7]
Elkins, S., Kochmar, E., Cheung, J. C., Serban, I. (2024). How teachers can use large language models and Bloom’s taxonomy to create educational quizzes. arXiv Preprint, arXiv:2401.05914.
[8]
Gao, W., Liu, Q., Huang, Z., Yin, Y., Bi, H., Wang, M.-C., Ma, J., Wang, S., Su, Y. (2021). RCD: Relation map driven cognitive diagnosis for intelligent education systems. In: Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval. New York: ACM, 501–510.
[9]
Gao, W., Liu, Q., Wang, H., Yue, L., Bi, H., Gu, Y., Yao, F., Zhang, Z., Li, X., He, Y. (2024a). Zero-1-to-3: Domain-level zero-shot cognitive diagnosis via one batch of early-bird students towards three diagnostic objectives. arXiv Preprint, arXiv:2312.13434.
[10]
Gao, W., Liu, Q., Yue, L., Yao, F., Lv, R., Zhang, Z., Wang, H., Huang, Z. (2025). Agent4Edu: Generating learner response data by generative agents for intelligent 20 education systems. arXiv Preprint, arXiv:2501.10332.
[11]
Gao, W., Liu, Q., Yue, L., Yao, F., Wang, H., Gu, Y., Zhang, Z. (2024b). Collaborative cognitive diagnosis with disentangled representation learning for learner modeling. arXiv Preprint, arXiv:2411.02066.
[12]
Gao, W., Wang, H., Liu, Q., Wang, F., Lin, X., Yue, L., Zhang, Z., Lv, R., Wang, S. (2023). Leveraging transferable knowledge concept graph embedding for coldstart cognitive diagnosis. In: Proceedings of the 46th International ACM SIGIR Conference on Research and Development in Information Retrieval. New York: ACM, 983–992.
[13]
Glorot, X., Bengio, Y. (2010). Understanding the difficulty of training deep feedforward neural networks.Journal of Machine Learning Research, 9: 249–256
[14]
Grattafiori, A., Dubey, A., Jauhri, A., Pandey, A., Kadian, A., Al-Dahle, A., Letman, A., Mathur, A., Schelten, A., Yang, A., , . (2024). The Llama 3 herd of models. arXiv Preprint, arXiv:2407.21783.
[15]
Hou, Y., Zhang, J., Lin, Z., Lu, H., Xie, R., McAuley, J., Zhao, W.X. (2024). Large language models are zero-shot rankers for recommender systems. arXiv Preprint, arXiv:2305.08845.
[16]
Hu, J., Liu, X., Fan, Z., Yin, Y., Xiang, S., Ramasamy, S., Zimmermann, R. (2024). Prompt-based spatio-temporal graph transfer learning. In: Proceedings of the 33rd ACM International Conference on Information and Knowledge Management. New York: ACM, 890–899.
[17]
Jiang, M., Bao, K., Zhang, J., Wang, W., Yang, Z., Feng, F., He, X. (2024). Item-side fairness of large language model-based recommendation system. arXiv Preprint, arXiv:2402.15215.
[18]
Kingma, D. , Ba, J. (2014). Adam: A method for stochastic optimization. arXiv Preprint, arXiv:1412.6980.
[19]
Kojima, T., Gu, S. S., Reid, M., Matsuo, Y., Iwasawa, Y. (2022). Large language models are zero-shot reasoners. arXiv Preprint, arXiv:2205.11916.
[20]
Li, Z., Chen, Z. Z., Ross, M., Huber, P., Moon, S., Lin, Z., Dong, X. L., Sagar, A., Yan, X., Crook, P. A. (2024). Large language models as zero-shot dialogue state tracker through function calling. arXiv Preprint, arXiv:2402.10466.
[21]
Liu, S., Yu, X., Ma, H., Wang, Z., Qin, C., Zhang, X. (2023). Homogeneous cohort-aware group cognitive diagnosis: A multi-grained modeling perspective. In: Proceedings of the 32nd ACM International Conference on Information and Knowledge Management. New York: ACM, 4094–4098.
[22]
Lu, Y., Pian, Y., Shen, Z., Chen, P., Li, X. (2021). SLP: A multi-dimensional and consecutive dataset from K-12 education. In: Proceedings of the 29th International Conference on Computers in Education. 261–266.
[23]
Ma, H., Li, M., Wu, L., Zhang, H., Cao, Y., Zhang, X., Zhao, X. (2022). Knowledge-sensed cognitive diagnosis for intelligent education platforms. In: Proceedings of the 31st ACM International Conference on Information & Knowledge Management. New York: ACM, 1451–1460.
[24]
Ma, H., Wang, C., Zhu, H., Yang, S., Zhang, X., Zhang, X. (2024a). Enhancing cognitive diagnosis using un-interacted exercises: A collaboration-aware mixed sampling approach.Proceedings of the AAAI Conference on Artificial Intelligence, 38(8): 8877–8885
[25]
Ma, H., Xia, A., Wang, C., Wang, H., Zhang, X. (2024b). Diffusion-inspired cold start with sufficient prior in computerized adaptive testing. arXiv Preprint, arXiv:2411.12182.
[26]
Mesnard, T., Hardin, C., Dadashi, R., Bhupatiraju, S., Pathak, S., Sifre, L., Rivière, M., Kale, M. S., Love, J., Tafti, P., , . (2024). Gemma: Open models based on Gemini research and technology. arXiv Preprint, arXiv:2403.08295.
[27]
Open AI, Achiam, J., Adler, S., Agarwal, S., Ahmad, L., Akkaya, I., Aleman, F. L., Almeida, D., Altenschmidt, J., Altman, S., , . (2023). GPT-4 technical report. arXiv Preprint, arXiv:2303.08774.
[28]
Pei, H., Yang, B., Liu, J., Dong, L. (2017). Group sparse bayesian learning for active surveillance on epidemic dynamics. In: Proceedings of the 32nd AAAI Conference on AI and 13th Innovative Applications of AI Conference and 8th AAAI Symposium on Educational Advances in AI. New York: ACM, 800–807.
[29]
Qian, H., Liu, S., Li, M., Li, B., Liu, Z., Zhou, A. (2024). ORCDF: An oversmoothing-resistant cognitive diagnosis framework for student learning in online education systems. In: Proceedings of the 30th ACM SIGKDD Conference on Knowledge Discovery and Data Mining. New York: ACM, 2455–2466.
[30]
Ren, X., Wei, W., Xia, L., Su, L., Cheng, S., Wang, J., Yin, D., Huang, C. (2024). Representation learning with large language models for recommendation. In: Proceedings of the ACM on Web Conference 2024. New York: ACM, 3464–3475.
[31]
Teknium, R., Quesnelle, J., Guang, C. (2024). Hermes 3 technical report. arXiv Preprint, arXiv:2408.11857.
[32]
Touvron, H., Lavril, T., Izacard, G., Martinet, X., Lachaux, M.-A., Lacroix, T., Rozière, B., Goyal, N., Hambro, E., Azhar, F., , . (2023a). Llama: Open and efficient foundation language models. arXiv Preprint, arXiv:2302.13971.
[33]
Touvron, H., Martin, L., Stone, K., Albert, P., Almahairi, A., Babaei, Y., Bashlykov, N., Batra, S., Bhargava, P., Bhosale, S., , . (2023b). Llama 2: Open foundation and fine-tuned chat models. arXiv Preprint, arXiv:2307.09288.
[34]
Wang, F., Liu, Q., Chen, E., Huang, Z., Chen, Y., Yin, Y., Huang, Z., Wang, S. (2020). Neural cognitive diagnosis for intelligent education systems.Proceedings of the AAAI Conference on Artificial Intelligence, 34(4): 6153–6161
[35]
Wang, F., Liu, Q., Chen, E., Huang, Z., Yin, Y., Wang, S., Su, Y. (2022). NeuralCD: A general framework for cognitive diagnosis.IEEE Transactions on Knowledge and Data Engineering, 35(8): 8312–8327
[36]
Wang, W., Fang, T., Li, C., Shi, H., Ding, W., Xu, B., Wang, Z., Bai, J., Liu, X., Cheng, J., , . (2024a). CANDLE: Iterative conceptualization and instantiation distillation from large language models for commonsense reasoning. arXiv Preprint, arXiv:2401.07286.
[37]
Wang, W., Jiao, W., Huang, J., Dai, R., Huang, J.-t., Tu, Z., Lyu, M. R. (2024b). Not all countries celebrate Thanksgiving: On the cultural dominance in large language models. In: Proceedings of the 62ndAnnual Meeting of the Association for Computation Linguistics. Bangkok: ACL, 6349–6384.
[38]
Wei, J., Wang, X., Schuurmans, D., Bosma, M., Xia, F., Chi, E., Le, Q. V., Zhou, D. (2022). Chain-of-thought prompting elicits reasoning in large language models. arXiv Preprint, arXiv:2201.11903.
[39]
Wu, Y., Xie, R., Zhu, Y., Zhuang, F., Zhang, X., Lin, L., He, Q. (2024). Personalized prompt for sequential recommendation.IEEE Transactions on Knowledge and Data Engineering, 36(7): 3376–3389
[40]
Yang, F., Chen, Z., Jiang, Z., Cho, E., Huang, X., Lu, Y. (2023a). PALR: Personalization aware LLMs for recommendation. arXiv Preprint, arXiv:2305.07622.
[41]
Yang, S., Chen, M., Wang, Z., Yu, X., Zhang, P., Ma, H., Zhang, X. (2024a). DisenGCD: A meta multigraph-assisted disentangled graph learning framework for cognitive diagnosis. arXiv Preprint, arXiv:2410.17564.
[42]
Yang, S., Qin, L., Yu, X. (2024b). Endowing interpretability for neural cognitive diagnosis by efficient Kolmogorov–Arnold networks. arXiv Preprint, arXiv:2405.14399.
[43]
Yang, S., Wei, H., Ma, H., Tian, Y., Zhang, X., Cao, Y., Jin, Y. (2023b). Cognitive diagnosis-based personalized exercise group assembly via a multi-objective evolutionary algorithm.IEEE Transactions on Emerging Topics in Computational Intelligence, 7(3): 829–844
Yang, S., Zhen, C., Tian, Y., Ma, H., Liu, Y., Zhang, P., Zhang, X. (2023c). Evolutionary multi-objective neural architecture search for generalized cognitive diagnosis models. In: Proceedings of 2023 the 5th International Conference on Data-Driven Optimization of Complex Systems. New York: IEEE, 1–10.
[46]
Yu, X., Qin, C., Shen, D., Ma, H., Zhang, L., Zhang, X., Zhu, H., Xiong, H. (2024a). RDGT: Enhancing group cognitive diagnosis with relation-guided dual-side graph transformer.IEEE Transactions on Knowledge and Data Engineering, 36(7): 3429–3442
[47]
Yu, X., Qin, C., Zhang, Q., Zhu, C., Ma, H., Zhang, X., Zhu, H. (2024b). DISCO: A hierarchical disentangled cognitive diagnosis framework for interpretable job recommendation. arXiv Preprint, arXiv:2410.07671.
[48]
Zhang, D., Zhang, K., Wu, L., Tian, M., Hong, R., Wang, M. (2024). Path-specific causal reasoning for fairness-aware cognitive diagnosis. In: Proceedings of the 30th ACM SIGKDD Conference on Knowledge Discovery and Data Mining. New York: ACM, 4143–4154.
RIGHTS & PERMISSIONS
Higher Education Press
AI Summary 中Eng×
Note: Please be aware that the following content is generated by artificial intelligence. This website is not responsible for any consequences arising from the use of this content.