Katrin
Schuessler
*a,
Michael
Giese
b and
Maik
Walpuski
a
aUniversity of Duisburg-Essen – Chemistry Education, Schützenbahn 70, Essen 45127, Germany. E-mail: katrin.schuessler@uni-due.de
bUniversität Duisburg Essen – Institut für Organische Chemie, Universitätsstr. 7, Essen 45141, Germany
First published on 15th October 2025
When visual representations of molecules (e.g., skeletal formula) must be decoded to process a task (e.g., determine the absolute configuration of a molecule) and the corresponding schemas are not yet sufficiently automated, paper–pencil format notes may help select relevant information, organize it appropriately, and integrate knowledge without exceeding the working memory capacity (encoding and external storage). This article examines the extent to which task difficulty and invested mental effort differ for digital and paper–pencil-based tasks on the topic of chirality (RQ1) and the extent to which note-taking impacts students’ working memory load when working on paper–pencil-based chirality tasks (RQ2). The dataset is based on the responses of 80 students from Germany who completed 19 chirality task tandems (each consisting of one digital and one paper–pencil-based task) and rated their invested mental effort for each task. Item response theory analyses, group comparisons, and moderation analyses were conducted. Paper–pencil-based chirality tasks were found to be significantly easier than digital chirality tasks, and students invested significantly less mental effort in completing the paper–pencil-based chirality tasks (RQ1). Students who took notes in the paper–pencil format were found to be more capable of solving chirality tasks in both formats. Both groups invested a comparable amount of mental effort. A moderation analysis revealed that when note-taking was low, the relationship between invested mental effort and the probability of solving the task was strongest. For the note-takers, the relationship between the invested mental effort and the probability of solving the task decreased as the number of notes increased (RQ2). The results indicate that notes as external storage are relevant for processing tasks requiring handling representation. As the digital format does not offer comparable options for taking notes, notes represent a subject-specific format difference.
Perceptual fluency (for an overview, see Kellman and Garrigan, 2009) enables experts to extract relevant information from representations unconsciously and without cognitive effort. Accordingly, experts can derive all the necessary information from visual representations of molecules and convert it into other visual (Rau, 2018) or verbal (Kozma and Russell, 1997) representations. Kozma and Russell (1997) defined the ability to recognize “different surface features as all representing the same principle, concept, or chemical situation” and transform them into another form of representation as representational competence. Expert-novice comparisons show that experts are significantly more proficient in recognizing commonalities regardless of representation and are more proficient in translating representations into text than novices (Kozma and Russell, 1997). Novices struggle to translate between verbal and visual representations (Kozma and Russell, 1997; Cooper et al., 2010; Bodé et al., 2016; Rau, 2018), as they struggle to comprehend visual representations, assign meaning to them (Anzovino and Lowery Bretz, 2016; Graulich and Bhattacharyya, 2017; Asmussen et al., 2023; Dood and Watts, 2023), and derive information from them (Asmussen et al., 2023). Students unable to assign meaning to representations struggle with organic chemistry (Anzovino and Lowery Bretz, 2016; Graulich and Bhattacharyya, 2017), usually rely only on memorization (Grove and Lowery Bretz, 2012), and fail with this strategy (Anderson and Bodner, 2008).
The close relationship between content knowledge and representational competence is problematic for learning: students must work with representations to learn concepts but require concepts to interpret representations (representational dilemma; Rau, 2017). Accordingly, learners with higher content knowledge have higher representational competencies (Sim and Daniel, 2014; Rho et al., 2022) and representational competence mediates between prior and content knowledge in organic chemistry (Dickmann et al., 2019).
For organic chemistry, the International Union of Pure and Applied Chemistry (IUPAC) requires the use of the skeletal formula (Fig. 2) as the standard form of representation (Brecher, 2008). Thus, organic chemistry textbooks (e.g., Clayden et al., 2001; Grossman, 2020) and molecular editors (e.g., ChemDraw) use the skeletal formula. Skeletal formula represents molecules in a more reduced form (carbon atoms and the hydrogen atoms attached to them are not represented explicitly but only by C–C bonds). For students, the degree of abstraction of the skeletal formula is a challenge to be mastered as part of their professionalization (Dood and Watts, 2023).
No explicit introduction of the skeletal formula at the university level is currently planned in Germany (Society of German Chemists, Gesellschaft Deutscher Chemiker, GDCh, 2021), and organic chemistry textbooks offer insufficient opportunities for practice (Gurung et al., 2022). However, working with representations of individual molecules is a crucial basis for more complex tasks in further learning (Stowe and Esselman, 2023). Without explicit promotion of representational competence, the duration and course level of chemistry lessons at school remain significant predictors of representational competence after several semesters of study (Taskin et al., 2017).
In summary, successful handling of representations such as the skeletal formula is crucial for processing tasks in organic chemistry. Students must learn to read skeletal formulas to consider implicit information, identify relevant structures, and relate them to scientific concepts. Provided these processes are insufficiently automated, they represent a challenge for working memory load.
Study results on the cognitive load when working with molecules of different sizes show a correlation between the level of cognitive load and the degree of expertise in drawing comparatively small molecules (Tiettmeyer et al., 2017). The probability of solving tasks decreases with increasing element interactivity: while 80% of molecules with six or fewer atoms are correctly translated from the half-structure formula into a Lewis structure; this applies to only 30% of molecules with seven or more atoms (Cooper et al., 2010). This indicates a cognitive overload due to a lack of sufficient schema automation (Sweller, 1988).
In summary, for novices, working with relatively small molecules is already associated with a high cognitive load because many elements and their relationships with each other (element interactivity) must be processed.
When schema automation is insufficient and element interactivity is high, markings and notes can help reduce the load on working memory by noting partial solutions (rather than holding them in working memory). Notes thus help in selecting, organizing, and integrating relevant information (Mayer, 2012, 2014). A note-taking review describes note-taking when reading texts, listening to lectures, or watching videos as a five-step process (Jansen et al., 2017): (1) the material presented must be understood, (2) key points or aspects must be identified, (3) the material must be related to prior knowledge and previous notes, (4) new information must be paraphrased or summarized, and (5) the result must be transformed into a written form. Therefore, taking notes is more cognitively demanding than just merely listening; however, if sufficient cognitive capacity is available, it can lead to better learning outcomes because information is processed more deeply (encoding effect, Di Vesta and Gray, 1972). A second function of note-taking is known as the external storage effect (Rickards and Friedman, 1978; Kiewra et al., 1991). Notes support learning by presenting selected relevant information that can be studied. If visual representations of molecules must be decoded to process a task and the corresponding schemas are not yet sufficiently automated, notes may help in encoding and providing external storage to select relevant information, organize it appropriately, and integrate knowledge without exceeding the working memory capacity.
While notes in paper format can be used easily and flexibly, digital molecular editors offer significantly fewer and less flexible options for working with notes (Schuessler et al., 2024a,b). Therefore, the format (paper–pencil-based or digital) is a relevant factor when investigating the role of notes in processing tasks in chemistry.
![]() | ||
| Fig. 3 Step-by-step determination of the absolute configuration of a molecule based on notes; (a–g) show the individual solution steps. | ||
Hence, if the absolute configuration of a molecule represented by the skeletal formula must be determined, this representation must be used to (1) extract contained information (e.g., non-depicted hydrogen atoms, Fig. 3b), (2) identify relevant structures (e.g., carbon atoms with four different substituents, Fig. 3c), (3) use prior knowledge (e.g., CIP rules to determine the priority of substituents, Fig. 3d), and (4) integrate information to generate partial solutions (e.g., the absolute configuration of a chirality centre, Fig. 3e). These partial solutions can (5) be noted to save them temporarily (e.g., note “S” after determining the absolute configuration of the first chirality centre, Fig. 3f, to free up working memory capacity for determining the configuration of the second chirality centre). In the first four steps, notes function more in the sense of encoding; in the fifth step, the external storage function predominates.
Traditionally, research on note-taking occurs when learning from lectures, texts, or videos (Rickards and Friedman, 1978; Kiewra et al., 1991; Jansen et al., 2017). The role of note-taking in working with chemical representations appears to have been minimally studied. Recent studies in chemistry investigating the handling of molecule representations have instead focused on the advantages and disadvantages of two- and three-dimensional representations (Habig, 2020; Keller et al., 2021). Therefore, as molecular editors do not offer note-taking options comparable to the paper format, digital tasks must be solved without notes or an analogue scribble sheet must be used (which can be inconvenient because the molecules must first be drawn). Thus, the format (digital or paper–pencil-based) could be relevant when examining the role of notes in working with chemical representations.
To what extent do task difficulty and invested mental effort differ for digital and paper–pencil-based tasks on the topic of chirality? (RQ1)
To what extent does note-taking impact students’ working memory load when working on paper–pencil-based chirality tasks? (RQ2)
Based on the literature, it is expected that digital tasks are solved correctly less often than paper–pencil-based tasks (H1.1). The mental effort invested may be higher for digital tasks because more information must be held in working memory (external storage, H1.2a) or lower because less deep processing occurs without taking notes (encoding, H1.2b). Two hypotheses can also be formulated regarding RQ2: students who take notes report a lower cognitive load because they relieve their working memory (external storage, H2.1a). Alternatively, students who take notes may also report a higher cognitive load because they process information more intensively (encoding, H2.1b).
Three additional students from the winter semester of 2023/24 participated in the data collection. As they only completed part of the survey, they were excluded from the data analysis.
The data collection occurred as part of the organic chemistry exercise the students attended. Students in all three degree programmes attended an introduction to organic chemistry in which they learn about the structure of simple organic compounds, functional groups, and the associated basic reaction types (radical substitution, nucleophilic substitution, elimination, and electrophilic addition). The exercise was supervised by two lecturers (Authors 1 and 2). Author 1 is an expert in chemistry education, and Author 2 is an expert in organic chemistry. The exercise accompanied an organic chemistry lecture. As part of the exercise, students worked on digital exercises at home every week. A molecular editor (Kekule.js) was implemented in the e-learning and e-assessment tool JACK to enable the digital implementation of organic chemistry tasks (Schuessler et al., 2024b), such as the chirality tasks, enabling molecules to be drawn using the skeletal formula. The student drawing was translated by the system into an InChI code (Heller et al., 2015) that uniquely described the molecule stereospecifically. InChI codes generated from student input are compared with the stored sample solution and can thus be evaluated automatically. The system also has the option to evaluate multiple-choice tasks.
In class, exercise sessions were used to work together on further tasks and discuss questions and problems from the lecture or the digital exercises. The students completing the exercises of this study were given feedback on their current performance level (regardless of whether they had agreed to the data analysis for scientific purposes). In the summer semester of 2024, students could collect bonus points to improve their grades on passing the exam by actively participating in the exercise in which this study was conducted on a specific date (consent to use the data for scientific purposes was independent of the awarding of bonus points). The introduction of bonus points had an impact on the number of students who participated in the exercise.
It was not (individually) checked whether and how often the students had practiced with the digital tasks as part of the exercise. The only condition for completing the digital chirality exercises was that an introductory set of digital exercises from the first exercise session had been completed.
Occasionally, students asked whether they could use a scribble sheet during digital data collection. This was permitted. The scribbles were not collected and were therefore not considered for the data analysis.
For the paper–pencil-based task processing, the first author also coded whether markings or notes were used during task processing. All processing traces in the paper-based test booklets that could be clearly assigned to a task were considered. If a task had traces of processing, it was coded as 1. Tasks without recognizable traces of processing were coded with 0. Fig. 5 provides an overview of the processing traces. Fig. 5a shows markings and notes also used by the lecturers supplying the exercise. Fig. 5b shows students’ markings and notes not used by the lecturers supplying the exercise.
![]() | ||
| Fig. 5 Example traces of processing. Red highlighting subsequently added by the authors. (a) Markings and notes used in the exercise; (b) alternative markings and notes. | ||
tasks = 38, Ninvested
mental
effort
ratings = 38) were satisfactory.
The task-based dataset examined comprised 19 tasks for which the following variables were available: relative task difficulty in paper format, relative task difficulty in digital format, invested mental effort in paper format, and invested mental effort in digital format. Easier chirality tasks (solved correctly by many students) are associated with a lower Rasch value, while more difficult tasks are associated with a higher Rasch value. Tasks in which many students have invested much mental effort are associated with a lower Rasch value. Tasks in which many students have only invested a small amount of mental effort are associated with a higher Rasch value.
RQ1 was answered by using t-tests to compare the relative task difficulty and the invested mental effort between the formats (paper vs. digital). As the students worked on both task types and tandem tasks were designed to address similar content, t-tests for paired values were used and a d for repeated measures (dRM) was reported. Additionally, correlation analyses were conducted. The t-tests and correlation analyses were calculated using SPSS (version 29). The d value for repeated measurements was calculated using the Psychometrica website (Lenhard and Lenhard, 2017).
RQ2 was answered by labelling students with at least one processing trace coded for at least one task as note-takers. Processing traces represented all types of notes or markings that appeared to have been made in connection with the task processing (examples are shown in Fig. 5). Winsteps was used to examine the extent to which the person abilities of students who left traces of processing differed from the person abilities of other students (differential item function [DIF] analyses; an effect size d was reported for groups of different sizes). Winsteps was also used to calculate the solution probabilities (observational averages) for the digital and paper–pencil-based chirality tasks as well as for the invested mental effort in both formats separately for both groups (note-takers vs. students whose test booklets had no processing traces). This resulted in eight new variables for each task: observational average for the paper–pencil-based chirality tasks, observational average for the digital chirality tasks, observational average for the paper–pencil-based invested mental effort ratings, and observational average for the digital invested mental effort ratings – once each for the note-takers and once each for the students whose test booklets showed no traces of processing (Table 1).
| Note-takers | Students whose test booklets show no traces of processing | ||||||||
|---|---|---|---|---|---|---|---|---|---|
| Observational average for the paper-based chirality tasks | Observational average for the digital chirality tasks | Observational average for the paper-based invested mental effort ratings | Observational average for the digital invested mental effort ratings | Observational average for notes (paper format) | Observational average f or the paper-based chirality tasks | Observational average for the digital chirality tasks | Observational average for the paper-based invested mental effort ratings | Observational average for the digital invested mental effort ratings | |
| Tasks | |||||||||
Additionally, the observational average for notes was calculated for each paper format task (based on the number of notes taken by the note-takers) and added as an additional variable for each task. Observational averages were imported to SPSS, which was used for further analyses. Paired t-tests were used to compare the solution probabilities for the chirality tasks and the invested mental effort for the note-takers and non-note-takers within the formats (digital and paper). The solution probability for the invested mental effort in the paper format for the note-takers was centered using the group mean for further analyses. Moderation analysis was used to examine the extent to which an interaction term comprising note-taking and invested mental effort explains the probability of solving the paper–pencil-based drawing tasks better than either variable alone.
The SPSS PROCESS tool (version 4.2) was used to analyse the influence of note-taking on the relationship between invested mental effort and solution probability (PROCESS Model #1). The results section reports R2 and changes in R2 for the model. Additionally, it reports t-ratios, regression coefficients (B), and standard errors (SEs) for the model parameters as well as the results based on Johnson–Neyman output. For all analyses, 95% bias-corrected and accelerated confidence intervals, based on 1000 bootstrap samples (95% CI), are reported in parentheses.
Table 2 provides an overview of the descriptive statistics before and after the Rasch analysis.
| Item | Paper–pencil format | Digital format | ||||||||
|---|---|---|---|---|---|---|---|---|---|---|
| Average score (solution) | Rasch value (solution) | Average score (invested mental effort) | Rasch value (invested mental effort) | Number of people who have taken notes | Average score (solution) | Rasch value (solution) | Average score (invested mental effort) | Rasch value (invested mental effort) | ||
| Chirality centre | C01 | 0.91 | −2.29 | 3.05 | 1.04 | 5 | 0.89 | −1.99 | 3.54 | 0.77 |
| C02 | 0.59 | −0.02 | 3.33 | 0.87 | 15 | 0.71 | −0.68 | 3.63 | 0.72 | |
| C03 | 0.60 | −0.08 | 3.84 | 0.60 | 11 | 0.66 | −0.40 | 4.31 | 0.34 | |
| C04 | 0.70 | −0.61 | 3.71 | 0.67 | 10 | 0.69 | −0.54 | 4.56 | 0.21 | |
| C06 | 0.78 | −1.13 | 3.94 | 0.55 | 9 | 0.75 | −0.90 | 4.63 | 0.13 | |
| Identical or enantiomers | C07 | 0.63 | −0.20 | 4.73 | 0.09 | 9 | 0.70 | −0.61 | 5.89 | −0.57 |
| C08 | 0.69 | −0.54 | 5.06 | −0.10 | 11 | 0.63 | −0.20 | 4.93 | 0.00 | |
| C10 | 0.68 | −0.47 | 5.24 | −0.20 | 9 | 0.40 | 0.90 | 5.60 | −0.43 | |
| C11 | 0.54 | 0.23 | 4.61 | 0.16 | 6 | 0.48 | 0.53 | 5.20 | −0.19 | |
| C12 | 0.73 | −0.75 | 4.74 | 0.09 | 3 | 0.66 | −0.40 | 5.30 | −0.25 | |
| Draw | C13 | 0.60 | −0.08 | 4.44 | 0.24 | 13 | 0.50 | 0.39 | 4.84 | −0.08 |
| C14 | 0.60 | −0.10 | 4.40 | 0.14 | 11 | 0.36 | 1.06 | 5.08 | −0.33 | |
| C16 | 0.49 | 0.47 | 5.29 | −0.32 | 11 | 0.26 | 1.59 | 5.80 | −1.13 | |
| C18 | 0.35 | 1.15 | 5.85 | −0.71 | 10 | 0.28 | 1.41 | 5.43 | −1.01 | |
| Absolute configuration | C19 | 0.55 | 0.17 | 4.39 | 0.27 | 21 | 0.50 | 0.01 | 3.80 | 0.19 |
| C20 | 0.60 | −0.08 | 4.69 | 0.07 | 27 | 0.49 | 0.01 | 3.95 | −0.04 | |
| C21 | 0.43 | 0.78 | 4.68 | 0.00 | 31 | 0.31 | 0.98 | 4.53 | −0.46 | |
| C22 | 0.63 | −0.20 | 5.13 | −0.21 | 22 | 0.15 | 2.15 | 4.43 | −0.33 | |
| C23 | 0.55 | 0.16 | 5.20 | −0.30 | 31 | 0.43 | 0.27 | 4.48 | −0.50 | |
![]() | ||
| Fig. 6 Boxplot diagram for relative task difficulty for the paper–pencil and digital formats. The circle labelled six describes an outlier. This is the task shown above as an example (Fig. 4). | ||
For the invested mental effort ratings, the relative task difficulty for the total sample of students, N = 80, differed between paper–pencil-based tasks, M = 0.16, SE = 0.10, and digital tasks, M = −0.16, SE = 0.11. The difference, ΔM = 0.31, 95% CI [0.21, 0.41], was significant, t(18) = 6.42, p ≤ 0.001, dRM = 1.68 (Fig. 7). Thus, students invested significantly less mental effort in completing the paper–pencil-based chirality tasks than in completing the digital chirality tasks.
In both formats, a significant correlation was found between relative task difficulty and invested mental effort. The correlation between invested mental effort and relative task difficulty was slightly weaker for the paper format, r = −0.65, p = 0.002, 95% CI [−0.898, 0.024], than for the digital format, r = −0.75, p ≤ 0.001, 95% CI [−0.913, −0.506]. Thus, tasks with higher task difficulty involved higher investment of mental effort.
Students who took notes, M = 0.72 and SE = 0.13, showed higher person abilities in the chirality tasks than students whose paper-based test booklets showed no processing traces (DIF analyses), M = −0.03, SE = 0.16, ΔM = 0.75, SE = 0.20, F(1,78) = 15.82, p ≤ 0.001, d = 1.73. Hence, students who took notes were more able to solve chirality tasks than those whose paper-based test booklets showed no processing traces.
Students who took notes, M = −0.16 and SE = 0.07, showed equivalent invested mental effort to students whose paper-based test booklets showed no processing traces (DIF analyses), M = −0.14, SE = 0.10, ΔM = 0.01, SE = 0.13, F(1,78) = 0.01, and p = 0.06. Hence, students who took notes invested a comparable amount of mental effort to students whose paper-based test booklet showed no evidence of processing.
Since the two groups of students (note-takers and students whose paper-based test booklets showed no traces of processing) differed in their ability to solve the chirality tasks (DIF analyses), the observational averages (solution probabilities) calculated separately by Winsteps for both groups for the paper–pencil-based and digital chirality tasks and the mental effort invested in both formats were used for further analyses. Additionally, the number of notes was used to calculate the probability that the note-takers would take notes for a task. It was then determined to what extent the probability of solving the chirality tasks and the invested mental effort differed for both groups of students (note-takers and students whose paper test booklets showed no traces of processing) in both formats (paper vs. digital).
For the paper–pencil-based chirality tasks, the solution probability differed for the note-takers, M = 0.67, SE = 0.03, and the non-note-takers, M = 0.53, SE = 0.04. The difference, ΔM = −0.15, 95% CI [−0.23, −0.07], was significant, t(18) = 3.63, p = 0.003, dRM = 0.71. For the digital chirality tasks, the solution probability also differed for the note-takers, M = 0.62, SE = 0.04, and the non-note-takers, M = 0.44, SE = 0.05. The difference, ΔM = −0.18, 95% CI [−0.256, −0.09], was again significant, t(18) = 4.31, p = 0.002, and dRM = 0.9805. Therefore, the higher personal ability of the note-takers was not solely due to the paper format. Instead, note-takers performed significantly better with both formats.
The (solution probability for the) invested mental effort for note-takers and non-note-takers in paper–pencil-based (note-takers: M = 3.63, SE = 0.23, non-note-takers: M = 3.72, SE = 0.12) and in digital format (note-takers: M = 4.23, SE = 0.24, non-note-takers: M = 4.11, SE = 0.14) did not differ significantly, paper–pencil-based: t(18) = 0.71, p = 0.470, digital: t(18) = 0.89, and p = 0.362. Hence, the effort invested by both groups did not differ based on the respective format.
For the note-takers, a moderation analysis was conducted to better explain the higher probability of solving the paper–pencil-based chirality tasks with comparable mental effort. For this purpose, the invested mental effort for the group of note-takers for the paper–pencil-based tasks was first centered using the mean value. Fig. 8 illustrates the relationship between the probability of solving the task and the invested mental effort for two extreme groups of items: items with few notes versus items with many notes. For items for which many notes were made, the probability of solution was relatively constant, regardless of the amount of mental effort invested. For items for which few notes were made, the probability of solution varied more widely. Items with a high probability of being solved tended to have below-average levels of mental effort invested, while items with a low probability of being solved tended to have above-average levels of mental effort invested.
The results indicate that invested mental effort, B = −0.18, SE = 0.03, t = 5.63, p < 0.001, 95% CI [−0.25, −0.11], and note-taking, B = −0.37, SE = 0.50, t = 2.74, p = 0.015, 95% CI [−2.44, −0.31], were significant predictors for the probability of solving the task. The interaction term (invested mental effort × note-taking) was another significant predictor, B = 0.33, SE = 0.12, t = 2.72, p = 0.016, 95% CI [0.07, 0.58]. Overall, the model accounted for 80% of the variance of the probability of solving the task, with the interaction term explaining 10% variance beyond the main-effects terms.
The interaction indicated that when note-taking was low (1 SD beneath the mean), a significant negative relationship existed between the invested mental effort and the probability of solving the task (B = −0.14, 95% CI [−0.19, −0.10], SE = 0.02, t = 6.75, p < 0.001). When note-taking was at the mean, the relationship between the invested mental effort and the probability of solving the task was weaker (B = −0.09, 95% CI [−0.12, −0.05], SE = 0.02, t = 4.90, p < 0.001). When note-taking was high (1 SD above the mean), the relationship between the invested mental effort and the probability of solving the task was no longer significant (B = −0.03, 95% CI [−0.10, 0.04], SE = 0.03, t = 0.81, p = 0.431). Hence, the correlation between the invested mental effort and the probability of solving the task decreased as the number of notes increased.
The second finding is consistent with hypothesis H1.2a (the mental effort invested is higher for digital tasks because more information must be held in working memory, external storage). Consequently, hypothesis H1.2b (the mental effort invested is lower for digital tasks because less deep processing occurs without taking notes, encoding) must be rejected. Students invest mental effort to solve digital chirality tasks, but without external storage, they must invest more effort and are less successful. This interpretation is supported by the findings of correlation analyses: for both formats, a significant negative correlation was found between relative task difficulty and the Rasch parameter for invested mental effort. Tasks with higher task difficulty involved a higher investment of mental effort. Therefore, the solution of digital chirality tasks appears impaired. From an instructional efficiency perspective (Paas and van Merriënboer, 1993; van Gog and Paas, 2008), processing paper–pencil-based chirality tasks appears more favourable than processing digital chirality tasks. When working on digital chirality tasks, an additional load appears to exist on the working memory, which is reflected in the greater mental effort invested. The inability to relieve the working memory by taking notes is a possible cause. The significantly higher mental effort invested in the digital chirality tasks supports this assumption and indicates a possible cognitive overload in the absence of the opportunity to relieve working memory by taking notes (external storage, Rickards and Friedman, 1978; Kiewra et al., 1991).
In view of the results, neither the external storage nor the encoding hypothesis is supported. Students who take notes do not indicate a lower cognitive load (external storage, H2.1a) but also do not report a higher cognitive load (encoding, H2.1b). From an instructional efficiency perspective (Paas and van Merriënboer, 1993; van Gog and Paas, 2008), note-taking appears more efficient, as these students gain better results with the same amount of effort. However, since these students also perform better in the digital format, where they cannot take notes, they may also have a higher level of expertise and therefore perform better with the same amount of mental effort. One possible explanation is the relativity of cognitive load (Brünken et al., 2012): since no objective criterion exists for the amount of mental effort invested, two people can associate different things with, for example, a medium amount of mental effort (Krieglstein et al., 2025).
A moderation was calculated to predict the observational average for the chirality tasks for note-takers using invested mental effort and note-taking as predictors. The moderation analysis indicated that invested mental effort, note-taking, and the interaction term (invested mental effort × note-taking) are significant predictors of the probability of solving the task. Overall, the model accounted for 80% of the variance of the probability of solving the task, with the interaction term explaining 10% variance beyond the main-effects terms. When note-taking was low, the relationship between invested mental effort and the probability of solving the task was strongest. When note-taking was high, the relationship between invested mental effort and the probability of solving the task was no longer significant. Hence, the correlation between the invested mental effort and the probability of solving the task decreases as the number of notes increases. However, it is unclear why note-takers also perform better in digital format. In principle, it could be expected that the note-takers using the digital format (where they cannot take notes) would achieve similar results to the students whose paper-based test booklets had no processing traces (or that their mental effort increased disproportionately).
A further limitation in connection with the analysis is the scribbles that individual students used in digital format, as these were not included in the study. Our analysis of the notes relates exclusively to notes that were made in paper format. We did not consider the individual notes that were made in digital format for two reasons. Firstly, we do not consider these notes to be of equal value, as they do not primarily support helpful processing procedures, but also require a high degree of (actually unnecessary) drawing of the molecule from the screen. It is therefore very likely that they are associated with a high level of split attention and extraneous cognitive load. Secondly, it was the exception that students used this option, so that a quantitative evaluation would not have been possible, also because a clear assignment to the associated tasks is hardly possible with these scribbles.
The decreasing correlation between the mental effort invested and the probability of solving the task as the number of notes increases indicates that taking notes is a means to manage a high working memory load, reduce it, and thus also solve difficult tasks (external storage). However, future studies should aim to replicate the results with clearer group comparisons (students who must take notes for each task and students who are not allowed to take notes), perhaps combined with previous note-taking training for the note-taking group. Additionally, the level of task complexity (e.g., determining the absolute configuration of molecules with one, two, or three chirality centres) at which students depend on notes when solving tasks should be investigated.
Studies on the importance of note-taking are predominantly from learning situations, whereas the data described here are from a test situation. We assume that our test subjects were still novices at the time of testing, whose corresponding schemas are not yet sufficiently automated. In this sense, the test situation could also be seen as a further learning opportunity. The relationship between the impact of notes and developing expertise (full schema automation) should be investigated further in the future.
Additional data on student motivation, extraneous cognitive load in the digital format, and working memory resources could be helpful in investigating whether more effort is invested at some point due to a lack of motivation, the extent to which extraneous cognitive load increases the mental effort required in the digital format, and the extent to which tasks may not be solved due to a lack of working memory resources. Furthermore, the study was conducted in Germany, where there are no specific requirements for the introduction of the skeletal formula. We observe that it tends not to be used at school, whereas it is standard (and often not explicitly introduced) in organic chemistry at the university level. To gain better insights, it would be useful to collect more systematic information on how teachers at school and at university introduce and use the skeletal formula. Finally, we examined students from three degree programs at one university. We did not investigate the extent to which there are differences between students on the different degree programs. A replication of the results with a different sample (university or degree program) would therefore appear desirable.
The results of a moderation analysis showed that the correlation between the invested mental effort and the probability of solving the task decreased as the number of notes increased. This indicates that notes as external storage are relevant for processing tasks requiring handling representation. As the digital format does not offer comparable options for taking notes, notes represent a subject-specific format difference.
| This journal is © The Royal Society of Chemistry 2026 |