Skip to main content

The role of confidence in the gaze bias effect among economics trainee teachers — results from a digital assessment of economic content knowledge

Abstract

In the present study, we recorded the eye movements of 20 criterion-based selected trainee teachers in economics while they responded to 25 single choice (SC) items in an economic content knowledge (CK) test and rated their confidence for each response in a digital assessment. By using a multilevel modeling approach with crossed random effects, we confirmed prior findings from eye-tracking research on SC tests, which showed longer dwell time on the correct response options (attractor) and shorter dwell time on the distractors are positively linked to correct options. Furthermore, we identified an additional effect on dwell time on the attractor in a moderator model with participants who highly rated their confidence for correct response options. Thus, we identified a specific role of students’ confidence in their CK on the gaze bias effect. We interpret these results in terms of students’ actual understanding of test contents from assessments of their professional knowledge and draw implications for further research and teacher education.

Introduction and research questions

Computer-based digital teaching-learning and assessment tools are increasingly being used in teacher education for vocational training to assess and promote professional teacher knowledge and its central facets such as content knowledge (CK) and pedagogical content knowledge (PCK) (Nilsson and Karlsson 2019; Zlatkin-Troitschanskaia et al. 2019a). Competent teachers should not only have sufficient professional knowledge (CK, PCK, PK or pedagogical knowledge), but also should be able to assess and apply it appropriately in a self-reflective manner (Schön 1987). The teachers’ self-reflection on their role in the classroom, their personality, their situational teaching behavior as well as on their own learning processes and thus their professional knowledge has a high impact on the learning processes of the students (Korthagen 2004; Rodríguez et al. 2014). The development of the ability to self-reflect is not only relevant to correcting one’s own misconceptions but also to establishing a strategic learning process (Shulman 1986). Self-reflection is shown to be particularly important in relation to a teacher’s professional knowledge and its application (Feucht et al. 2017; Schön 1987).

In connection with teachers’ learning and self-reflection abilities, the more specific question arises to what extent they are able to monitor and correctly assess the learning processes and their professional knowledge. While (teacher) students recognize the importance of professional knowledge, they often do not feel confident in their knowledge or, in contrast, overestimate it. One explanation in the literature refers to deficits in students’ ability to evaluate their knowledge and recognize deficiencies (Kruger and Dunning 1999; Eva et al. 2004). In SC tests, which are widely established to measure professional knowledge in teacher studies, confidence in one’s knowledge is not usually assessed. Therefore, teacher student participants may answer correctly for instance by simply guessing (Walstad et al. 2018). When responding to a content knowledge SC task and deciding one response option, this (type of) reasoning can be based on prior knowledge or subjective “good feelings”, e.g., described by Kahneman (2011) as “System 2,” i.e., a “subjective feeling of confidence”. For instance, in a study that analyzed economics students while responding to economics knowledge test tasks, it was identified that naïve students tend to respond to these tasks, albeit with self-reported high uncertainty, with a bipolar superficial approach that reflects their good feelings more adequately than the elaborateness of their knowledge (Leiser and Aroch 2009). Some further studies explored to what extent students have confidence in their content knowledge or, in contrast, to what extent they respond to test questions based on a “strategically selected option” (e.g., guessing, Sanders et al. 2016).

Few studies consider not only the correctness of a response but also the students’ confidence in their answers when investigating knowledge development (Khan et al. 2001; Gardner-Medwin 1995; Cordova et al. 2014). This research illustrates that the awareness of confidence in relation to professional knowledge has a major influence on the development of knowledge (Stankov 2013). Research on knowledge assessment and confidence testing (e.g., Bruno 1993; Davies 2002) indicates that confidence in the correctness of one’s response to a task in a knowledge test can be considered as an appropriate indicator of the extent to which a student’s response is based on knowledge vs. (strategic) guessing (Kolbitsch et al. 2008).

To confront teachers with the relation between their confidence and knowledge, confidence ratings have also been used in teacher education research (e.g., Dassa and Nichols 2019; Kim and Klassen 2018). The frequently observed tension between teachers’ actual knowledge and their confidence rating has been researched intensively (e.g., Podgoršek and Lipovec 2017; Brückner and Zlatkin-Troitschanskaia 2018) and raises the question whether teachers are self-aware about the discrepancy in their perception of knowledge and their demonstrated understanding. Although this topic is relevant to a variety of domains in teacher education, initial eye tracking studies between the domains of physics and economics have shown that trainee teachers for vocational education in economics are less reliable in correctly assessing their content knowledge than trainee teachers in physics (Klein et al. 2019). Thus, a need has been identified to more comprehensively investigate the accuracy of self-assessments among teachers of vocational education in economics.

The CK of (trainee) teachers is usually assessed using (SC) items that are especially helpful in the context of the increasing digitalization of teaching-learning processes and are easy to implement and whose data are more readily fed into and scored in a Learning Management System (Parkes and Zimmaro 2016). SC tests are regularly used in courses as a part of audience or classroom response systems, as they are able to provide learners with immediate feedback on their actual performance and learning progress (Greving et al. 2020).Footnote 1 There are a number of tests available for assessing economic knowledge in vocational education which can be used in a valid way, especially for trainee teachers in economics (Walstad et al. 2007, 2013; Zlatkin-Troitschanskaia et al. 2019b).

To investigate actual comprehension and performance using computer-based digital CK and PCK tests, which are currently gaining increasing popularity in teacher education, it is necessary to analyze the processing of digital learning or testing materials by (trainee) teachers using eye-tracking in addition to the analysis of learning performance based on the CK/PCK test scores. An analysis of the response processes occurring while answering test items is important to gain insight into students’ cognitive processing (Ercikan and Pellegrino 2017; Zumbo and Hubley 2017). To find out how attention is spatiotemporally directed to different item areas, eye-tracking studies have been carried out, which facilitate dedicated analyses of respondents’ gaze behavior (Holmqvist et al. 2011). Recent literature reviews highlight the increasing importance of eye-tracking in in empirical research in (vocational) education (Mayer et al. 2023).

One particular focus of the current study is the interaction between information content processing and self-reflective abilities, which are measured as confidence in response correctness. Such self-reflective abilities are considered in research to be necessary components of economic teachers’ professional economic knowledge in vocational education but have hardly been studied to date. Only the analysis of students’ processing of content can reveal important information about the extent to which differences exist between trainee teachers that possess knowledge and rate their confidence in this knowledge differently. Therefore, in this study, the intraindividual spatiotemporal processing of trainee teachers of economics in their bachelor’s program is investigated based on their response processes when answering a professional CK test in economics.

Based on the preliminary work of Brückner et al. (2020); Zlatkin-Troitschanskaia et al. (2019b), this paper presents an eye-tracking study examining how economics trainee teachers perform on an economics knowledge test administered digitally and focuses on two research questions (RQ). With the first RQ, we seek to confirm findings on SC items in teacher education using the gaze bias effect (for a definition, see Sect. 2.1) identified in prior research (Lindner et al. 2014).

RQ1

To what extent can the economics trainee teachers’ dwell time on whole SC items and individual response-relevant or response-irrelevant parts of the items of a CK test in economics predict the correct or incorrect response of these SC items?

In addition, the confidence with which trainee teachers respond to the item is considered an “essential skill for efficient study and work practice” (Gardner-Medwin 1995, p. 81). Eye-tracking studies conducted by Brückner et al. (2020) and Klein et al. (2020) revealed that confidence also affects gaze behavior. This leads to the second RQ:

RQ2

To what extent is item-related confidence related to dwell times on single elements of a correctly responded item?

Based on prior eye-tracking research (Brückner et al. 2020; Lindner et al. 2014; Klein et al. 2020), we propose working hypotheses and explain the research design, including the CK test used and the sample of students. After presenting the results from multilevel models that take the interactions between students and items into account, we discuss the limitations of the study and implications for future research in education of teachers in economics for vocational education.

Theoretical background

Eye-tracking research and gaze bias effect in standardized educational assessments

Eye tracking has been used in cognitive and educational research for many years (Holmqvist et al. 2011; Mayer et al. 2023). It is increasingly applied in the analysis of well-structured learning environments and standardized educational assessments in various disciplines (Han et al. 2017; Klein et al. 2019; Lindner et al. 2014; Saß et al. 2017; Tsai et al. 2012). Here, the focus has often been placed on the validation of the construct like graph comprehension and knowledge by elaborately investigating gaze behavior during task processing (Zumbo and Hubley 2017).

Eye-tracking research with respect to the knowledge of students assumes that there are associations between visual perception, interpretation, and understanding between learners that possess more or less knowledge within that domain. The cognitive theory of visual expertise (Gegenfurtner et al. 2023) is an example of a possible underlying theory. Classical approaches like the “immediacy assumption” establish the link between cognitive activity, the order of its processing, and the sequence of visual perception, i.e., cognitions that occur during an action, e.g., solving an economics task (Just and Carpenter 1980; Holmqvist and Andersson 2017). The “eye-mind assumption” associates the moment of visual perception with the moment of attention and information processing (Holmqvist et al. 2011), however does not adequately reflect the complex relationships between knowledge and visual processing.

For instance, novice learners who have a lower level of domain knowledge will foveally perceive, understand, and mentally process typical challenges of the domain differently than learners with a higher level of knowledge (Larkin et al. 1980). In addition, learners with a higher level of knowledge are considered to be more efficient at selecting relevant and ignoring irrelevant information than novice learners (Haider and Frensch 1999). They perceive information from the environment by foveal and parafoveal vision and keep them in a visual register for a short time, they bundle several pieces of visually perceived information into so-called image chunks, which, in addition to the classical assumptions, enable holistic mental representations that are kept as retrieval cues in working memory (Gegenfurtner et al. 2023) and thus allow faster information processing than perceiving information individually and sequentially. In this way, advanced learners are better able to connect their mental capacity and knowledge from long-term memory with their representations of e.g., economics concepts presented in the tasks, like economics principles and rules, and to attach meaning to them and to process them in a resource-efficient way connected with the suitable domain knowledge (Gegenfurtner et al. 2023).

Through this association of perception, cognitive processing, and memory, it can be assumed that both the relative frequency and duration of perception in certain relevant and irrelevant areas of interest (AoI) can serve as an indicator for naive or advanced learners, also among economics trainee teachers.Footnote 2 With respect to SC item responding, various cognitions are shown to play a role in determining the selection of a particular response option from multiple response options, e.g., in the initial reception of information and its interpretation to the prediction of a preference for a specific response option and its final selection and evaluation (Parkes and Zimmaro 2016). The correct options (attractor) and the incorrect ones (distractors) represent the central response-relevant and response-irrelevant features of the item, respectively, and indicate the intensity with which the students deal with certain item content.

A phenomenon often observed in the investigation of gaze behavior during SC tests is the so-termed ‘gaze bias effect’ or ‘gaze cascade effect,’ which plays a major role in visual decision-making (e.g., in marketing research or face recognition, Shimojo et al. 2003; Glaholt and Reingold 2009; Saito et al. 2017). Lindner et al. (2014, p. 738) describe the gaze bias effect as a positive correlation between the preference for an object and the duration with which this object is viewed. For example, when buying a car, the car that is purchased is more likely to be viewed and analyzed by the buyer for a longer period of time than cars that are ultimately not purchased. They transferred this effect to SC tests for the first time and showed that a gaze bias effect can also be detected in decision-making between given response options. When people are asked to choose one of several response options, they usually spend more time, i.e., have fixations of longer durations, looking at the response option they will ultimately choose than at the other options, e.g., students responding correctly to the item should focus on the attractor longer (Gegenfurtner et al. 2023; Lindner et al. 2014).

In further studies from physics education research, the eye-tracking studies using a kinematic graph comprehension test (Klein et al. 2020), incorrect responses were also associated with longer dwell times on attractive distractors and lower dwell times on attractors and vice versa for correct responses. In Tsai et al. (2012), students processing a meteorological task spent more time on the response options they chose. Moreover, incorrect respondents had more difficulties understanding the question and extracting relevant information.

These partly different findings might be due to different tests and analysis foci, e.g., analyzing eye movements in terms of dwell time to describe task-response behavior. Based on prior research on the gaze bias effect, we intend to replicate the findings from Lindner et al. (2014) and Klein et al. (2020) in a first step, showing that economics trainee teachers correctly responding to the item can be expected to have a longer dwell time on the attractor than those with incorrect responses.

Since an economics knowledge test has not yet been subjected to an eye-tracking analysis, the abovementioned findings are used as the theoretical foundation for the research hypotheses. Based on the gaze bias effect for SC tests (Lindner et al. 2014), we suggest:

H1

The longer the average duration of fixation on the attractor, the higher the probability of a correct response.

H2

The shorter the average duration of fixation on the distractors, the higher the probability of a correct response.

Effect of economics trainee teachers confidence on gaze behavior and test scores

The relationship of knowledge to confidence, as an indicator of accuracy of teacher reflective abilities, is of great importance to teaching competence (Dassa and Nichols 2019; Podgoršek and Lipovec 2017). Confidence in one’s own expertise in a knowledge test is critical in achieving learning success and applying acquired knowledge in learning environments (Gardner-Medwin 1995). As part of teacher competence, confidence influences the learner’s actions and provides an insight regarding the likelihood with which a learner’s task response might be correct (Stankov and Lee 2008).

The relationship between knowledge and confidence is intensively studied, e.g., in the heuristics-bias approach to explain why individuals overestimate or underestimate their performance and the ways in which this disparity manifests itself in practice (Stankov and Lee 2008). Confidence ratings have already been used in several educational assessments in various disciplines, e.g., to obtain an indication of whether guessing or competent learning behavior is used in responding to an item via the discrepancy between confidence and test score (Brückner and Zlatkin-Troitschanskaia 2018). Studies generally assume that higher knowledge is also associated with higher confidence (Gardner-Medwin 1995). In studies of graph comprehension with bachelor students and trainee teachers in economics and physics, trainee teachers in economics were found to estimate their knowledge of graphs in the economics domain less accurately than physics students in their own domain. While there was a positive correlation, there was still a domain difference that necessitates more specific investigation of the correlations in the domain of economics teaching in vocational education (Brückner et al. 2020; Klein et al. 2019). Therefore, we expect:

H3

The share of correct responses should be higher for responses with high confidence than for responses with medium or low confidence.

Connections with confidence were also explored using eye tracking (Brückner et al. 2020; Klein et al. 2020). These studies have demonstrated that, in general, higher test scores on a graph test in economics were correlated with higher confidence, indicating that high, medium, or low confidence can be reflected in gaze patterns. Building on prior studies on the gaze bias effect (Lindner et al. 2014) and the assumed positive correlation between confidence and test scores, we assume:

H4

The higher the confidence of economics trainee teachers, the higher the probability of a correct response due to the extended (shortened) dwell time on the attractor (distractors).

Design and sample

The descriptions in this chapter take into account the twelve reporting standards for eye tracking studies as recommended in Dunn et al. (2023), which we complemented by several aspects.

Test and areas of interest (AoIs)

In this study, we used the economics knowledge test, which comprises 25 SC items (for details, see Zlatkin-Troitschanskaia et al. 2019b). Each item consists of one question, one attractor and three distractors (for an example, see Fig. 1). The test covers basic economic content that is generally required in economics teacher education worldwide (Holtsch et al. 2019). Each correct response is coded as 1 and incorrect responses are coded as 0. A maximum of 25 points can be achieved by each participant.

Fig. 1
figure 1

Sample item from the WiWiKom test with the five labeled AoIs (colored rectangles) (translated version on the left, taken from Walstad et al. 2007)

In this eye-tracking study, the five components of the items (four distractors and one attractor) were defined as AoIs for the analyses. Gaze data were collected specifically for these areas, which were spatially defined with a high degree of separation and without overlaps (Fig. 1). They reach beyond the text area, as deviations in the measured gazes were taken into account due to the precision values. Moreover, marginal areas were defined for each AoI that were at least 1° of the visual angle. A distance of 2° was defined between AoIs to avoid any confounding in the data (Holmqvist et al. 2011). An additional ‘global’ indicator was also created that showed students’ overall processing of a task at the millisecond level.

After processing the test and selecting a response option for each item, economics trainee teachers were given a six-point Likert scale to assess their confidence in their response (1 = not confident, …, 6 = very confident). This scale was aggregated into three categories of 0 = low confidence, 1 = medium confidence, 2 = high confidence (Gardner-Medwin 1995; Klein et al. 2020) to increase the robustness of statistical analyses in the cross-random effects model.

Apparatus

The items were implemented using the software UniparkFootnote 3. The assessment was then implemented in the web-stimulus element of the eye-tracking software Tobii Pro Lab³ with version 1.152.30002 (x64) and presented to the test participant on a desktop computer with a 22-inch monitor with a resolution of 1920 × 1080 pixels and a refresh rate of 60 Hz. The total system latency was 11 ms. Below the monitor, an infrared-based stationary eye-tracker Tobii Pro X3-120Footnote 4 using a pupil-corneal reflection method with a sampling frequency of 120 Hz was mounted, which allowed the trainee teachers to move their heads freely without a chinrest or similar objects and to assess both eye positions accurately. The laboratory in which the study took place was darkened and indirectly lit to prevent interference from other infrared sources. Precision was 0.24° of visual angle.

To calculate dwell times, the fixation duration metric was specified. Based on the manufacturer’s specifications, fixations were measured on a millisecond basis using the identification by velocity threshold (I-VT) filter with a threshold of 30°/s of visual angle and a minimum fixation duration of 60 ms. Each fixation was always preceded and followed by a saccade, i.e., a metric to measure the eye movement between two fixations. Saccade classification was defined as when the acceleration of the eyes exceeded 8500° s − 2 and velocity exceeded 30°s − 1.

Procedure

The participants’ distance to the eye-tracker was between 60 and 70 cm. A timed screen-based eye-tracking calibration was conducted with 9 black bullet points on a white background, 4 of which served as validation indicators to ensure the accurate and precise measurement of the trainee teachers’ gazes. Measurement deviations of less than 0.8° of the visual angle were tolerated. The items were presented to the test participants in randomized order. By pressing a button, a response option (A, B, C, or D) could be selected. During the processing, the test administrator monitored the assessment on a second screen and gave correction instructions whenever the precision of the measurement was no longer sufficient or the participants were outside the measurement range.

The participants were allowed to work on the items freely, there was no time limit. The average processing time for the 25 items was 15.13 (Minimum = 7.58 min; Maximum = 25.42 min) minutes in total, which results in an average processing time of 36.32 s (Median = 32.74 s) per item.

To ensure high test motivation, each participant received course-specific credit points for participating in this assessment, which trainee teachers need to accumulate a certain amount of to complete their course of studies.

Sample

For this study, we used the data of 20 economics trainee teachers from the degree course of economics education for vocational training from one German university. There were originally 22 participants in the study (15 female, 5 male, 2 missing), but the data from two of them were lost because the recording of gaze data was incomplete due to technical problems. The sampling approach used here was similar to that of the nation-wide representative main study (for details, see Zlatkin-Troitschanskaia et al. 2019b), i.e., the sample group was selected intentionally based on different descriptive criteria.

The average grade of the university entrance qualification in the sample was 2.25 (SD = 0.500). The average age was 23.1 years (SD = 4.518). The participants had, on average, completed 3.5 semesters (SD = 3.488). Half of the participants had completed a commercial vocational training before starting their university studies.

Analysis

Since each participant completed 25 items, a total of N = 500 response processes were available, which were clustered within students and items. Due to the nested data structure, variance splitting was necessary to account for the relationships between dwell times and test score, and confidence and dwell times. Therefore, multilevel models with crossed random effects (Rabe-Hesketh and Skrondal 2012; Snijders and Bosker 2012) were used, which are recommended for analyses of response process data (Strobel et al. 2018). Especially in unbalanced designs, random effects models have proven to be efficient and allow for the inclusion of dwell times and confidence as predictors of test scores (Rabe-Hesketh and Skrondal 2012).

Since the dependent variable item response is binary (correct vs. incorrect), a logit link function was used as a generalized linear mixed effects model. Moreover, it is assumed that confidence as a moderating variable can affect the dwell time on certain AoIs during the response process. Therefore, interaction terms, which accounts for the interaction between confidence and dwell time, were integrated into the multilevel model in addition to the main effects (moderator model).

To improve the interpretability of the results, some modifications were made: The dwell time was presented in seconds instead of the measured milliseconds. To compare the dwell time on the attractor with those on all three distractors combined, an additional variable was calculated showing the average dwell time on the three distractors.

Since a comparison between dwell times on attractor and distractor in terms of initiated cognitive processes is only possible if the stimuli are comparable, the average number of words of the two types of response options was also compared (Lindner et al. 2014). The average word count per attractor (M(SD) = 7.76(5.53) did not differ significantly from the average word count per distractor (7.31(3.63) (t = 0.697, dpooled=0.013, p = .492). Each response option was phrased in approx. seven words (Fig. 1).

There was neither an item- nor sample-specific accumulation in the occurrence of the confidence and correct or incorrect responses. Therefore, to analyze the differences in the dwell times, responses in which the high confidence and low scores occurred were extracted (N = 41) and compared with responses that were also incorrect but had a medium or low confidence estimate or a correct response with high confidence. There were no values for 13 responses.

Results

Across all items, there were 302 correct responses and 198 incorrect responses. On average, 60% of the responses were correct. To obtain indications regarding the generalizability of the findings, the distribution of item difficulties calculated in the main study based on the performance of approximately 5,000 students (Zlatkin-Troitschanskaia et al. 2019b) were compared to the distribution of item difficulties in this eye-tracking study using a two-sample Kolmogorov-Smirnov test. The results (Z = 0.990, p = .281) indicated that neither distribution significantly differs from the other.

For this analysis, we specifically focus on task processing at the individual level rather than on item comparisons. To investigate to what extent the dwell time differs depending on the selected response option, i.e., responding to the item either correctly or incorrectly, an exploratory repeated-measures analysis of variance was conducted (response option = within-factor; score = between-factor). The mean dwell time on the attractor (MW = 3.87) was significantly higher than the mean dwell time of the distractors, with a small effect size (MW = 3.41) (F(1, 519) = 9.064, p < .01; η²p=.017). This finding, however, is score-independent, since in each response process the students obviously paid more attention to the attractors than to the distractors, and it does not take into account compensatory effects. Although the overall dwell time on the response options is longer for incorrect responses (MW = 3.417) than for correct responses (MW = 3.87), it is not evident how the overall dwell time is distributed between the individual response options, and whether this distribution differs for correct and incorrect responses (Fig. 2). The results of the non-parametric Friedman test for dependent samples confirm this finding (χ²(df) = = 8.23(1), p < .01). The average dwell time spent on the AoI question was about 11.39 s for both correct and incorrect responses. A response-specific examination of dwell times revealed that the mean dwell time for correct responses (MW = 12.74, SD = 12.063) differed significantly (MW = 10.46, SD = 10.35) from that for incorrect responses (t = 2.310, p < .05, dpooled=0.206) for the AoI question. The results of the non-parametric U-test for dependent samples confirm this finding (Z=-2.626, p < .05).

Fig. 2
figure 2

Average dwell time (in seconds) of incorrect and correct respondents on the AoI attractor (left), mean of the three distractor AoIs (middle), and the AoI question (right). The error bars represent 1 standard error of the mean (SEM)

Considering the nested data structure, we first computed a variance component model, i.e., baseline model without covariates (Model 1 in Table 1). To assess the significance of the dwell times on each individual AoI for correct responses, we controlled for the effects of the dwell times on the other AoIs. In the multilevel model with crossed random effects, the log odds had different values (Model 2 in Table 1). When controlling for the dwell times on the attractor and distractors, no significant correlation was found between the dwell time on the AoI question and the test score (estimate=-0.009, z=-0.87, p = .382). However, the time spent on the distractor showed a highly significant negative correlation with the response (estimate=-0.356, z=-6.09, p < .001). Thus, if the dwell time on any distractor increases by one unit, the probability of a correct response decreases by 30%. Conversely, a longer dwell time on the attractor increases the probability of a correct response by 16% (estimate = 0.151, z = 3.93, p < .001). When comparing the two predictors, dwell time on the distractors proves to be more indicative of a correct response (Table 1). Therefore, H1 and H2 can be confirmed based on model 2, which shows that, as a predictor, the AoI question is no longer significant; this was also suggested by the ANOVA.

Table 1 Random intercept model with a binary logistic regression function and fixed effects on score

Relevant to H3, Table 2 illustrates that the proportion of correct responses is larger the higher the confidence rating is which corresponds to the general assumption. In line with previous research (Brückner and Zlatkin-Troitschanskaia 2018; Klein et al. 2020), the likelihood of a correct response is linked to the participant’s confidence in their response ω (Cohen 1988) (χ²(df) = 43.5874(2), p < .001, ω = 0.299) (Table 2). H3 can thus be confirmed.

Table 2 Cross table score × confidence

To test H4, first, a random intercept model that only includes confidence as a fixed effect was calculated. Taking into account the nested and unbalanced data structure and compared with responses made with low confidence, the likelihood that an item was responded to correctly was four times higher when students’ confidence ratings were high (odds ratio = 4.312, z = 5.26, p < .001). There was no significant effect when students were medium confident (odds ratio = 1.097, z = 0.36, p = .772). In Model 3, confidence was included as a covariate in addition to the dwell times on the AoIs (Model 3: odds ratio = 3.391, z = 4.21, p < .001), indicating that the assessment of confidence is a significant predictor for the likelihood of responding to the items correctly. Taking into account the dwell times on AoIs, it can be seen — in addition to Table 2 — that in the group comparison, the group with high confidence in particular shows a large correlation with a correct response (H3).

To implement the moderator model (Model 4 in Table 1), the significant main effects of the log odds of the average distractor and the attractor were each extended by an interaction effect with confidence. No significant interaction effect was found for the distractor, but a significant interaction effect beyond the significant mean effect was evident for the attractor when students were highly confident (odds ratio = 1.573, z = 2.73, p < .01). Thus, for participants with a high confidence, the probability of a correct response increases significantly with a dwell time on the attractor longer than 2.5 s (see Fig. 3). Hypothesis 4 can therefore only be partially confirmed, since a significant correlation between high confidence and longer dwell time on the attractor was found but not between a change in dwell time on the distractors and correct response.

Fig. 3
figure 3

Interaction between dwell time on attractor and students with low, medium, and high confidence and its predictive power regarding the average test score for all participants (with 95% confidence interval, dashed lines) (seconds < = 12)

These findings were also confirmed when dwell times on attractors and distractors for cases with correct solutions at high confidence are compared to those with correct solutions at low confidence. Part of the group with high confidence achieved incorrect responses( in 41 cases) (Table 2). To compare the dwell times, t-tests with independent samples were calculated. The different total dwell times, depending on the response process (Table 3) indicate that, for a comparison of the dwell times per AoI, the relative dwell times have to be used.

Table 3 t-tests with the total dwell time

The analyses (Table 4) illustrate that students with high confidence and a correct solution were more able to identify the correct solution shown by a higher preference for the attractor. Since students with a high degree of confidence and a correct solution are more likely to have the domain knowledge required to answer the task, the longer focus on the attractor also reflects their preference for this answer option. They were able to identify the correct answer from a variety of incorrect answer options.

Table 4 t-tests with the dwell time on attractor

Conversely, the relative dwell times on the distractors show that, in cases with high confidence and a correct solution, the dwell time on the distractors is similar to that with high confidence and an incorrect solution (Table 5).  However, students with greater uncertainty and incorrect solutions spend longer on the preferred incorrect solutions. This suggests that the relative dwell times may also reflect different task solving strategies, as solutions with greater uncertainty are worse at distinguishing between the correct and incorrect answer options, while cases with high confidence and correct solutions use the attractor purposefully.

Table 5 t-tests with the dwell time on distractor

Discussion

Given the tension between self-reflective skills and knowledge of prospective economics teachers for vocational education, this study examined the extent to which a change in the length of time spent on different AoIs influenced the test score depending on confidence.

Regarding RQ1: The assumption that correct responses are associated with shorter total dwell times was confirmed at the response process level (person × item), with a small effect size. The findings related to the comparative analysis of dwell times on distractors and attractors between trainee teachers who responded to the items correctly or incorrectly are consistent with prior research (Klein et al. 2020; Lindner et al. 2014). This indicates that participants who responded to the items correctly tend to dwell longer on attractors (H1) than on distractors and vice versa (H2). The reversed effect found for the AoI ‘question’ may be due to the fact that the eye-tracking metric ‘dwell time‘ can refer to fixations on the question or the response options and can therefore serve as an indicator for different cognitive functions/processes (gaze bias effect) (Lindner et al. 2014). With regard to the response options, economics trainee teachers tend to spend more time focusing on certain areas of an item or a particular response option if they are inclined to choose that response option (Thomas et al. 2019).

Using multilevel models with crossed random effects for each AoI, as expected, there was a positive correlation between a correct response and a longer dwell time on the attractors (H1) and a negative correlation between an incorrect response and the dwell time on the distractors. A shorter dwell time on the AoI question indicates a tendency to respond to the item correctly, however, this correlation was not significant. Since the economics knowledge test items focus primarily on the activation of (mental) schemes and less on the activation of complex mental activities like in problem-solving processes, one explanation might be that the performance of participants who possess the required knowledge can faster infer the meaning of the questions. This is in contrast to the findings of Klein et al. (2020) but replicates the findings of Lindner et al. (2014), who also found that students who responded to the item correctly tend to have shorter total dwell times than students who did not. This indicates, for economics knowledge test items, a longer dwell time on the question is associated with comprehension difficulties or more elaborate information processing by test participants (Tsai et al. 2012). Once all AoIs had been integrated into one model, the dwell time on the question did not appear to have any significant negative correlation with the response. However, the dwell times on the individual response options were highly significant, confirming previous findings on SC tests from other domains (Lindner et al. 2014).

Regarding RQ2: Confirming Klein et al. (2020) and another study that assumed a positive relation between economics knowledge and self-reported confidence (e.g., Leiser and Aroch 2009), economics trainee teachers’ confidence was positively correlated with the overall test score (H3). Thus, economics trainee teachers with higher economics knowledge tend to be able to self-reflect adequately. A high or low level of confidence was also reflected in the dwell times on the individual AoIs, which in turn were predictive of whether the item was responded to correctly. Longer dwell time at low confidence can be explained by actions characterized by higher doubt and hesitant deliberation (Stankov and Lee 2008). A high level of confidence was linked to faster response processes in the economics knowledge test, as the individual aspects of the item content were more quickly evaluated by the economics trainee teachers in terms of their relevance. However, the interaction model (Model 4) no longer shows a general confidence effect, indicating that the dwell times on the distractors and the attractor essentially determine the probability of responding to the item correctly. The moderator effect becomes evident in the interaction between dwell time on the attractor and high confidence. When confident responses were accompanied by a longer dwell time on the attractor, the probability of a correct response increased (H4).

Since the distractors address typical misconceptions of economics, they may also provide more in-depth insights into low confident economics trainee teachers’ misunderstandings. For example, the sample item (Fig. 1) describes that the income of the population in Germany is increasing overall, which apparently also leads to a general increase in consumption. If the economics trainee teachers chose one of the first two response options (distractors), it can be assumed that they do not understand the significance of a general increase in income and its effect on consumption.

In addition to findings from previous studies (Lindner et al. 2014; Klein et al. 2020), this study shows a correlation between confidence assessment dwell times on specific task parts and economics knowledge among trainee teachers in economics. At the same time, however, different effects can occur. Thus, it is necessary to capture the self-assessed confidence of trainee teachers in economics from a metacognitive perspective. It seems obvious that the dwell time on the different AoIs may be an expression of different task solving strategies that are used when confidence is high or low. At the same time, however, it seems necessary to diagnose the different facets of teachers’ professional knowledge even more precisely to find out what the explanations might be for different levels of confidence in answering tasks. Since knowledge tasks also depict different topics and concepts, it is obvious that teacher knowledge also varies and that alternative task solving strategies are used in cases of self-assessed uncertainty because the correct answer is not directly recognized. This has been emphasized before, e.g., Leiser and Aroch (2009, p. 381) conclude from their study: “On the one hand, they declare on average not to understand the concepts very well. On the other, they are quite willing to judge how changes in one economic variable would affect another. Our interpretation is that what enables the economically untrained to answer is their superficial approach to the issues.” Thus, comprehensive assessment of economics teacher knowledge also requires the measurement of teachers’ self-reflection to find out about their strategies for answering the content knowledge tasks. However, to diagnose how this is reflected in performing specific, professional tasks in their teaching job, such as responding to CK items, more extensive analyses using authentic tasks and log data analysis beyond eye tracking are required).

Limitations and future research

While the presented results are mostly in line with previous studies, further areas of research emerge for a more in-depth analysis of the significance of self-reflection as part of the response process. CK represents only one facet of teaching competence. The extent to which the phenomenon of correct and incorrect solutions with different levels of confidence and its effect on dwell times might also be evident in other knowledge dimensions, e.g., PCK, has not yet been explored. Likewise, the relationship of this phenomenon and the associated eye movements to teachers’ actual classroom performance is only vaguely suggestible. In particular, the effects of high and low confidence with high or low economic knowledge on classroom behavior, e.g., instruction or economics teachers’ detailed attention to student errors, remain to be investigated. Further studies using the corresponding assessments are still required.

In the present study, the eye movements of economics trainee teachers were investigated for the first time in the context of self-assessed confidence and economics knowledge for vocational education. This is a domain-specific finding. Moreover, the question arises whether these findings can be generally assumed for other teaching domains. First interdisciplinary studies on the domain comparison of graph comprehension suggest that there might be domain-specific differences (Klein et al. 2019; Brückner et al. 2020). However, empirical evidence has yet to be provided.

In this study, only SC tests in a traditional task format with one correct and several incorrect options were applied, as they were also commonly used in other studies (Klein et al. 2020; Tsai et al. 2012; Han et al. 2017). However, comparisons are not always possible, as these studies refer to other disciplines and do not exclusively focus on teacher education research. Including other constructs e.g., PCK entails also including tasks with other format representations, e.g., graphical rather than textual, which might be the addition of representation on a whiteboard in a classroom to the tasks. Since the response process can be affected by the type of representation (textual vs. graphical) as well as the specified cognitive demands (simply recalling content from memory vs. problem analysis) of an SC item or by specific content and teacher knowledge demands, different expectations should be formulated for different types of SC items (Saß et al. 2017).

When responding to SC items, participants have to choose one of several response options. Here, too, comprehension plays a role, but the focus mainly lies on the ‘attractiveness’ of the response options, one of which must be selected by the economics trainee teachers. As studies from other disciplines indicate (Lindner et al. 2014; Klein et al. 2020), the time the trainee teachers spend looking at the response options, i.e., dwell time, tends to be indicative of which response option they prefer and will eventually choose. In further studies, the individual distractors should be taken into consideration in a more differentiated manner, e.g., by analyzing them based on their ‘attractiveness’ and by matching eye-movement data with the item difficulty and discrimination parameters (derived from more comprehensive field studies) or other classroom specific parameters. For instance, in assessing PCK, the distractors and attractors might include different economic student or teacher statements that need to be evaluated.

Another (general) methodological limitation lies in the definition of AoIs (Bojko 2013; Holmqvist et al. 2011), which include textual content. The size of the AoIs significantly determines the dwell times and fixation frequencies to be assessed and was standardized across all items for this study.

Moreover, the question arises whether similar findings would have been obtained with mobile eye-trackers and paper-based SC tests. For instance, due to the particular setup of the experimental situations with a participant-to-administrator ratio of 1:1 (which differs from field surveys), the survey situations were highly controlled in terms of time, place, and person, and the participants always made an effort to work intensively on the items, which is less common for low-stakes surveys (with large samples). In future studies, a variation of audience-response systems or clickers should be implemented to find out how the feedback affects the (visual) perception of items. In addition, mobile eye trackers are often used to analyze classroom events (Goldberg et al. 2021).

How difference in content knowledge and teachers self-reflection could therefore also be investigated in the context of specific actions in the classroom and, together with teacher educators, an objectified evaluation of the actions could be compared with the trainee teachers’ self-reflections of these situations. For example, the controversially discussed Dunning-Kruger effect (Kruger and Dunning 1999), could also be a significant factor that needs to be investigated in more detail to obtain indications of different task strategies. The effect describes the phenomenon of deficits in one’s (here: content) knowledge with a concurrent high self-assessment of this knowledge, and is therefore to be seen critically, especially with regard to the necessary self-reflection in teaching (Dassa and Nichols 2019). To date, it is largely unclear how this manifests itself in the visual perception and selection of SC response options and it is discussed whether it is just a statistical artefact or not (Gignac et al. 2020).

In cross-linked mixed-effects models, the effects of dwell times on scores have been investigated by taking into account the cross-classification of dwell times in relation to both items and participants simultaneously (Strobel et al. 2018). Further predictors can be used at different levels, and future studies should also analyze gaze behavior in relation to item difficulty. For this purpose, adjustments and estimates of random effects are necessary, which require a larger sample — a greater number of items and participants.

When expanding the sample, different levels of expertise should be systematically taken into account, e.g., advanced students and first-year students, to analyze developments over the course of the study (Brückner et al. 2020). Furthermore, the present study did not aim to analyze how dwell times changed during the response process; thus, no analysis of the chronological sequence of dwell times on AoIs in specific time intervals has been conducted so far. In particular, multilevel models with autoregressive covariance structures and crossed-random effects might provide some valuable insights into time-dependent analyses. However, Lindner et al. (2014) showed in a gaze-likelihood analysis (across the response process of SC items) that the fixation times of participants with higher and lower performance levels on different task intervals was overall comparable in terms of their attention distribution over time, which was not the matter of this study.

Conclusion

Competent economics teachers should not only have sufficient professional knowledge (CK, PCK, PK), but also assess and apply it appropriately in a self-reflective manner. Self-reflection such as self-confidence is shown to be particularly important in relation to teachers’ professional knowledge and its application. In previous studies, economics teachers’ self-reflective competencies were theoretically modelled and empirically assessed (Brückner and Zlatkin-Troitschanskaia 2018). These studies empirically identified significant correlations between these self-reflective competencies and teaching skills. They suggest that such self-reflective competencies, in addition to professional knowledge, are a necessary foundation for professional action in the classroom (Feucht et al. 2017; Schön 1987).

Our study is based on prior research, in which a diagnosis and analysis of the differences between students’ confidence and their knowledge has already been used to explain differences in economics knowledge test performance based on isolated eye movements that provide insight into participants’ analytic information processing. To date, little research has been conducted to analyze the relationship between confidence, knowledge and eye movements as it pertains to (prospective) teachers, and no study was available for the domain of economics for vocational education. Therefore, based on research from the other domains (physics, biology), this eye-tracking study contributes towards bridging this research gap. The findings indicate that trainee teachers who exhibit differences between confidence and knowledge also differ in their gaze behavior from students who correctly assess their CK in economics. The results of this study thus not only indicate deficits in self-reflective skills in line with previous studies on teachers’ self-reflective competencies, but also point to the significant role these skills play in the acquisition and application of correct CK.

Further research is needed to investigate this phenomenon in other teacher professional knowledge areas such as PCK and PK. To this end, we are currently conducting an analogue eye-tracking study using a validated PCK test among economics students for vocational education (Kuhn et al. 2016). Here, it is of particular interest whether differences between confidence, eye movements, and knowledge that became evident in this study using a CK test can also be found in economics trainee teachers while responding to a PCK test.

In terms of practical implications, it can be concluded that such self-reflective skills need to be more explicitly addressed in economics teacher education. This is especially true in the context of increased digital learning and the use of freely available online information in economics teacher education, to prevent the acquisition of erroneous knowledge and misconceptions.

Data availability

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Notes

  1. The use of SC items in online assessments in domains with large student populations, like medicine or economics, has become more widespread (Calhoun and Mateer 2012).

  2. Eye tracking complements the introspective method of thinking aloud since unconscious cognitions can be tapped into by analyzing students’ gaze. This method is non-reactive and creates no additional mental load that may influence test behavior compared to the think-aloud method in which verbalized thoughts can confound cognitions (Neuert and Lenzner 2019).

  3. https://ww3.unipark.de/www/front.php.

  4. Tobii Pro Lab User Manual at https://www.tobiipro.com/siteassets/tobii-pro/user-manuals/Tobii-Pro-Lab-User-Manual/?v=1.152.1.

Abbreviations

CK:

Content Knowledge

PCK:

Pedagogical Content Knowledge

SC:

Single-Choice

AoI:

Areas of Interest

I-VT:

Identification by Velocity Threshold

References

Download references

Acknowledgements

We would like to thank all economics trainee teachers who participated in this study. Furthermore, we would also like to thank the reviewers for their differentiated feedback and their contribution to the improvement of this manuscript. Detailed comments, efforts, and expertise of the reviewers were outstanding and certainly contributed significantly to the quality of this manuscript.

Funding

This study was funded by the Center of School, Education and Higher Education Research at the Johannes Gutenberg University Mainz and the PLATO project funded by the Ministry for Science, Continued Education and Cultural Affairs Rhineland-Palatinate, Germany.

Author information

Authors and Affiliations

Authors

Contributions

Both authors contributed substantially to this work. Conceptualization, SB and OZT; methodology, SB; validation, SB and OZT; investigation, SB; resources, writing—original draft preparation, SB and OZT; project administration, OZT and SB; all authors have read and agreed to the published version of the manuscript.

Corresponding author

Correspondence to Sebastian Brückner.

Ethics declarations

Competing interests

The authors declare no conflict of interest.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Brückner, S., Zlatkin-Troitschanskaia, O. The role of confidence in the gaze bias effect among economics trainee teachers — results from a digital assessment of economic content knowledge. Empirical Res Voc Ed Train 16, 2 (2024). https://doi.org/10.1186/s40461-024-00156-2

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s40461-024-00156-2

Keywords