Structuring productive group work with Marzano's taxonomy: a study of cognitive scaffolding and group learning

Andrew Kreps , Kodi Dailey and Renee Cole *
Department of Chemistry, University of Iowa, Iowa City, IA 52242, USA. E-mail: akreps@tacomacc.edu

Received 21st June 2025 , Accepted 13th July 2025

First published on 14th July 2025


Abstract

This study explores the impact of cognitive scaffolding on group learning to establish the effective use of Marzano's taxonomy as both a diagnostic and developmental tool. Using the cognitive levels of processing as defined by Marzano's taxonomy, activities in an introductory chemistry course were revised to embed support in the structure of questions. Data was collected from two cohorts: Cohort 1, which engaged with original activities, and Cohort 2, which used our cognitively scaffolded activities. The ICAP framework provided the theoretical framework for analyzing social processing, knowledge dynamics, and modes of reasoning within the groups. Findings suggest that cognitive scaffolding helped sustain higher cognitive engagement while promoting the use of reasoning during higher-order questions. While social dynamics did not show as pronounced changes, these findings demonstrate the potential of cognitive scaffolding to support student learning. The study additionally highlights the nuanced nature of collaborative learning, with factors such as group composition, course climate, and facilitation methods influencing engagement outcomes.


Introduction

An increasing body of research highlights the advantages of active learning, which Theobald et al. (2020) define as “…any approach that engages students in the learning process through in-class activities, with an emphasis on higher-order thinking and group work.” Group work, in particular, provides students with opportunities to develop social skills and critical thinking while enhancing their engagement with course content (DeChurch and Mesmer-Magnus, 2010; Janssen et al., 2010; Chi and Wylie, 2014). The effectiveness of group work can be influenced by several factors, including the nature of the activity, assigned group roles, and group dynamics (Leupen et al., 2020; Premo et al., 2022).

Studies suggest that group work is most effective for tasks that are cognitively challenging and allow for student autonomy; in contrast, retention-focused questions tend to favor individual work (Kirschner et al., 2009; Scager et al., 2016; Wilson et al., 2018). While challenging tasks can enhance learning, they may also lead to cognitive overload and uneven participation, particularly in newly formed groups (Janssen et al., 2010). As Reid et al. (2022a,b) and Nennig et al. (2023) observe, group work is often implemented without adequate structure or feedback, and students may lack awareness of effective collaborative strategies. They advocate for the inclusion of explicit prompts that scaffold students toward productive reflection. These findings underscore the critical role of activity design in shaping the quality of student engagement in active learning environments and form the foundation for the present work.

Scaffolding

Scaffolding refers to the instructional practice of offering support to help students engage with content or skills that are currently beyond their independent ability (Davis and Miyake, 2004; Shvarts and Bakker, 2019). Unlike general course structure, which often focuses on organizing the overall learning environment, scaffolding emphasizes in-the-moment guidance during the learning process itself. As originally described by Wood et al. (1976), scaffolding involves six core functions of support: recruiting students' interest, reducing degrees of freedom, maintaining direction, highlighting critical features, managing frustration, and providing demonstrations (Wood et al., 1976). These supports can take various forms across a course and are typically adjusted or removed as students gain competence.

While the benefits of scaffolding, more specifically the structure of questions, are well established, there is limited research exploring how students in general chemistry respond to specific scaffolds, with even fewer studies focusing on the impact scaffolding has on group learning. Vo et al. (2025) discussed the prevalence of studies in organic chemistry courses, emphasizing the need for further investigations into scaffolding within general chemistry settings, particularly those that address the cognitive, behavioural, affective, and social dimensions (Fredricks et al., 2004; Vo et al., 2025). Although prior studies highlight the importance of balancing structure and autonomy to support reasoning, much of this work relies on lengthy, iterative revisions or interviews to refine scaffolded prompts, underscoring the time- and labour-intensive nature of developing effective materials (Noyes and Cooper, 2019; Noyes et al., 2022). While these approaches offer valuable models, they may be difficult for instructors to adopt without significant support.

In contrast, we propose the use of Marzano's taxonomy as a structured and more accessible framework for evaluating and developing scaffolded activities based on the levels of cognitive processing the taxonomy describes. Work by Toledo and Dubas (2016), along with others, demonstrated the successful application of Marzano's taxonomy to redesign instruction and assessment in a general chemistry course (Toledo and Dubas, 2016; James and LaDue, 2021). Our study builds on this by presenting a method for revising in-class group activities that embeds cognitive scaffolding.

This work addresses key gaps in the literature by first establishing the viability of using Marzano's Taxonomy to design cognitively scaffolded in-class activities, analyzing their impact on students' reasoning, knowledge dynamics, and social interactions, compared to groups working with non-cognitively scaffolded activities. We analyzed student discussions for evidence of knowledge construction, social processing, and reasoning strategies as they engaged with cognitively scaffolded in-class activities. The following sections introduce the coding scheme used in this study, provide relevant background, and explain how they relate to our central research question:

How does the cognitive scaffolding of in-class group activities developed with Marzano's taxonomy impact students’ cognitive and social engagement with chemical ideas?

Knowledge dynamic, social processing, and reasoning

Previous research examining scaffolded activities has employed a range of approaches, including open coding and thematic analysis (Heisterkamp and Talanquer, 2015; Yuriev et al., 2017; Kararo et al., 2019; Noyes and Cooper, 2019; Sheppard and Bauer, 2023), Toulmin's Argument Pattern (Moon et al., 2016, 2017a; Stanford et al., 2016), complexity-based frameworks (Broman et al., 2018) and analyses focused on correctness or error identification (Parker Siburt et al., 2011; Gencer, 2023). These studies often concentrate on an individual's engagement with the scaffolded tasks, which typically are focused on a specific outcome like developing problem-solving skills or reasoning. In contrast, the current study seeks to examine the broader impact of cognitive scaffolding on group interaction and engagement. To achieve this, we selected codebooks grounded in collaborative learning and discourse analysis.

Nokes-Malach et al. (2015) broadly define collaboration as “active engagement and interaction among group members to achieve a common goal,” emphasizing that collaborative learning is most effective when both cognitive and social processes are supported. Prior research has demonstrated that specific forms of discourse, such as complex reasoning, justification of ideas, and building upon peers’ knowledge, can improve learning outcomes and deepen conceptual understanding in chemistry (Towns, 1998; Michaels et al., 2008; Criswell, 2012; Moon et al., 2017b; Nennig et al., 2023). Therefore, to analyse the effects of cognitive scaffolding on group dynamics, we adopted a coding scheme that captures the social and epistemic dimensions of group interaction as well as the complexity of students’ reasoning.

Knowledge dynamic. Knowledge within a group setting can be described as a dynamic process stemming from interactions regarding how information is used and shared (Nonaka et al., 2006; Ford, 2008). To categorize a group's knowledge dynamic, it is essential to consider what is done with information once distributed among the group. Information or ideas that are shared without being questioned, interpreted, or built upon would represent a dynamic of knowledge sharing (van Aalst, 2009). When a group focuses on applying and relating information to a question or problem, the group dynamic is categorized as knowledge application. Information or ideas evaluated, situated, reflected, and built upon by the group represent a dynamic of knowledge construction. Knowledge construction is associated with deeper learning and promotes similar benefits found broadly in active learning, where it can lead to the restructuring of knowledge and new ways of thinking (Biggs, 1987; van Aalst, 2009).
Social processing. Social cohesion is generally understood to be an important part of group learning and can play a role in both the group's knowledge dynamics and participation (Forman, 1989; Johnson et al., 2007). Work from Kumpulainen and Kaartinen (2003) devised categories for the modes of social activity: collaborative, tutoring, domination, and confusion. Collaborative is where students equally participate and co-construct ideas. Tutoring is where a student is asked a question or questions by another member or members of the group. Domination is where one student primarily constructs the response while ignoring or rejecting input from other students in the group. Confusion is when the group is uncertain about how to proceed or lacks group understanding. For this work, we used an expanded list of social processing codes from the work of Reid et al. (2022a,b) and States et al. (2023) that includes individualistic, where the students work independently, and leader, where primarily one student constructs a response due to a lack of contribution from others. The additional codes were introduced due to observations during previous studies and are meant to add additional clarity to the types of social dynamics (Reid et al., 2022a; States et al., 2023).

Reasoning is integral to productive discourse as it promotes and supports critical thinking, developing deeper understanding, constructing arguments, and providing justification (Rojas-Drummond and Mercer, 2003; Mercer et al., 2004; Michaels et al., 2008). Yet, chemistry students tend to rely on a relational type of reasoning, where the connection between the features of a system and its behavior is made, leading to the construction of simple causal explanations and the reduction of variables considered (Sevian and Talanquer, 2014; Moreira et al., 2019; Deng and Flynn, 2021). Sevian and Talanquer (2014) defined four modes of reasoning that describe different levels of complexity: descriptive, relational, linear causal, and multicomponent. The descriptive and relational modes of reasoning represent surface-level reasoning, focusing on explicit features of systems and relying on relationships to drive reasoning. Linear causal and multicomponent are then where students construct a cause-and-effect type of argument; however, for the multicomponent modes, students consider and explore multiple variables, whereas, in linear causal, there may be a reference to various variables that are then reduced to a primary agent.

Theoretical framework

The framework of Marzano and Kendall's (2006) taxonomy of educational objectives was used as the lens for developing the cognitive scaffolding, and the ICAP framework, which describes four modes of cognitive engagement (Interactive, Constructive, Active, and Passive), was used for analyzing and interpreting student dynamics.

Marzano's taxonomy

While many types of scaffolding exist—from hard scaffolds (embedded, static supports) to soft scaffolds (dynamic and adaptable), and from content-dependent to content-independent supports—students’ responses can be just as varied (Vo et al., 2025). This variability is further complicated by the intended goal of the scaffolding, which may range from eliciting specific types of reasoning to modelling problem-solving strategies. Given this complexity, there is a clear need for a more approachable and systematic method for evaluating and designing scaffolding. To this end, we propose the use of Marzano's taxonomy, primarily the first four cognitive levels of processing: Retrieval, Comprehension, Analysis, and Knowledge Utilization (see Appendix A for the definitions of each level) (Marzano and Kendall, 2006). These levels describe how information is processed across three memory functions: sensory, permanent, and working memory. Because the levels are hierarchical, they offer a useful framework for implementing and evaluating scaffolding based on the cognitive processing required by a task, rather than its instructional goal or type.

This work is therefore informed by the taxonomy through its application in restructuring in-class group activities to incorporate cognitive scaffolding. In cognitive literature, this is referred to as cognitive structuring, which Tharp and Gallimore (1988) define as “… the provision of a structure for thinking and acting.” Previous work regarding cognitive structuring demonstrates that by carefully considering the changes in the cognitive load placed on the learner, you can reduce the chance of cognitive overload (Sweller, 1994; van de Pol et al., 2015; Zulu et al., 2018). This approach offers benefits similar to established methods of scaffolding, but without requiring the labour-intensive iterative development often needed to refine prompts. Moreover, it is flexible enough to be applied across an entire course, from overall design to in-class activities.

As an example of how the taxonomy was used, Fig. 1 demonstrates sample questions from one of the activities used in this study before and after cognitive restructuring. The first three questions in both versions of the activity are identical and relatively straightforward, asking students to calculate the change in entropy and Gibbs free energy. According to Marzano's taxonomy, these questions reflect Retrieval-level processing because they require students to recall information from permanent memory or apply formulas already provided. The final questions in each version (Questions 11 and 9e) represent Analysis-level processing, where students must synthesize prior knowledge to generate new insights—for instance, by evaluating the effect temperature can have on the direction of a reaction. In the original activity, students are asked to move directly from a Retrieval-level task to an Analysis-level one. This abrupt jump in cognitive processing increases task complexity and may raise the risk of cognitive overload (Sweller, 2010; Chen et al., 2023).


image file: d5rp00225g-f1.tif
Fig. 1 Representative example of the cognitive scaffolding implemented in the revised activities. The question set on the left (questions 8–11) is representative of questions written without cognitive structuring, while the questions on the right (questions 9a–9e) represent the revised set of questions that employ cognitive structuring.

To address this, the revised activity introduces an intermediate question (9d), which asks whether the reaction is favorable and prompts them to explain their reasoning. This question engages Comprehension-level processing by drawing attention to key features of the Gibbs free energy equation, such as the variables it contains. This question scaffolds their reasoning and supports a smoother cognitive progression by prompting students to reflect on these critical features before making interpretive judgments or providing reasoning. Highlighting critical features before asking students to generate new insights is one example of the cognitive scaffolding made possible through restructuring guided by Marzano's taxonomy. For further discussion about the taxonomy and its application in evaluating cognitive scaffolding, see our previous work (Kreps et al., 2024).

ICAP

When discussing active learning, it can become difficult to define when and how students engage with the content. To bring some clarity and guide the analysis for this project, we chose the ICAP framework (Chi and Wylie, 2014). The ICAP framework allowed us to understand the outcomes of active learning through the four levels of cognitive engagement: Interactive, Constructive, Active, and Passive. The ICAP framework defines engagement as “… the overt behavior the student exhibits while undertaking an activity.” Passive is defined as receiving information without doing anything else. Active is defined quite literally to be a student using some physical movement, action, or manipulation of materials, but the main feature remains that the student's attention is focused. Constructive is the process of building on top of what is provided to the students and must contain new ideas that are being generated. Lastly, interactive is defined by two criteria: the discourse must be constructive, and there must be sufficient back and forth between the students (Chi, 2009; Chi and Wylie, 2014; Chi et al., 2018). The constructive and active levels are distinct because they consider not only what an individual student is doing but also how they interact. We know that a group's social interactions are a factor in productive group learning, and the ICAP framework allows us to explore this aspect using the social processing codes mentioned above. In this work, we categorize the students' knowledge dynamics, social processing, and modes of reasoning, and the ICAP framework guided this study by providing a lens through which we evaluate these outputs to determine the impact our cognitive scaffolding had on the group learning setting.

Methods

Setting and data collection

The study was conducted at a large research-intensive university in the United States in two iterations of a second-semester introductory chemistry course. These courses each spanned fifteen weeks, featuring a lecture component three times a week for fifty minutes, a weekly three-hour laboratory or ninety-minute case study, and a weekly fifty-minute discussion session. The setting for this research focused on the discussion sections, where students worked through guided inquiry activities in small groups facilitated by graduate teaching assistants (GTAs). Each discussion section had approximately twenty-eight students, who worked in groups of three to four, with each group seated at its own table (see Fig. 2 for the layout of the discussion room).
image file: d5rp00225g-f2.tif
Fig. 2 The layout of the room where all discussions took place, with a representative example of the placement of the video and audio recorder.

The primary goal of the discussion was to provide an opportunity for students to work together on concepts they were learning, and it was expected that they would answer as many questions as possible in the allotted time. Nevertheless, answering every question in the given amount of time was not always possible, so grading was done by the GTA based only on the completion of questions answered. Each activity contained an overview of the learning objectives, applicable skills for the activity, questions, and a reflection. The reflections usually occurred in the middle of the activity, focusing on a specific process skill characteristic, and used portions of the Enhancing Learning by Improving Process Skills in STEM (ELIPSS) rubrics to aid students' reflection (Cole et al., 2019; Reynders et al., 2020; Czajka et al., 2021). The activities also included the roles of Manager, Recorder, Spokesperson, and Reflector as described by the POGIL (Process Oriented Guided Inquiry Learning) Project (Hoffman and Richardson, 2019). Students were meant to cycle through the roles each week to promote team learning, foster collaborative skills, and ensure no one student would always have the role of recorder, who was responsible for recording the group's answers and submitting the completed activity to the course learning management system.

Data collection occurred over two iterations of the second-semester introductory chemistry course, hereafter referred to as Cohort 1 and Cohort 2. Between these cohorts, a change was made to the facilitation methods and the activities the students worked on. For Cohort 1, each discussion began with a brief iClicker quiz before groups started the activity. For Cohort 2, the iClicker quiz was removed, allowing groups to begin the activity immediately.

In the first two weeks of each semester for both cohorts, researchers obtained consent from course instructors and GTAs. In week two, video recordings of the discussion sections were collected to help identify the groups to observe for the rest of the semester. Purposeful sampling, based on the days of the week and video recordings (Patton, 2014), ensured that selected groups worked collaboratively and did not complete activities prior to the discussion section. Two groups were selected from each semester: one meeting on Monday and the other on Thursday. This selection ensured that the groups sampled reflected any content knowledge gained during the lectures they attended throughout the week. In week three, researchers collected consent from the students in these groups, and from that point forward, the data collection process remained consistent across both cohorts. From weeks three to fifteen, each group was recorded using a video camera in the room, an audio recorder at the group's table, and an iPad with the week's activity uploaded and screen recording enabled. Students used the iPads to complete and submit their activities for GTA grading. Table 1 contains the reference names of each group along with the pseudonyms assigned to the students.

Table 1 The pseudonyms for each student group that participated in the study, including the day of the week their discussion took place
Monday Cohort 1 Cohort 2
Gold group Blue group
Hodgkin Sheeana
Daly Alice
Molina Dorothy
Olga

Thursday Cohort 1 Cohort 2
Red group Purple group
Ball Miles
Arrhenius Henry
Bohr Helen
Bragg Linus


The audio–video data from each discussion was synced and uploaded to MAXQDA coding software (VERBI Software, 2021). Once uploaded, we used the software to “timestamp” sections of the video where groups worked on each question. These timestamps marked specific units of discourse to be analyzed during the coding process, starting from when students began a question and ending when they moved on to the next. This approach was intended to capture the evolving dynamics within each group as they progressed through the activities rather than coding individual tasks in isolation. Screen recordings from the iPads supplemented the audio/video files, providing additional context if group conversations were unclear or offering further insight into the group's thinking and work while answering questions.

Ethical considerations

All data collection was approved by the Institutional Review Board (IRB #201309825). All participants provided written consent and had the option to decline or withdraw at any time. Researchers involved in the project had no connection to the participants. To ensure privacy and confidentiality, data was de-identified, and pseudonyms were assigned.

Revising discussion activities

Revisions to all fifteen activities primarily targeted the cognitive structure of the questions, with some additional adjustments to the reflection sections. The cognitive structure was assessed using Marzano's cognitive levels of processing. Each question in the original Cohort 1 activities was coded, after which two researchers reviewed them to identify necessary rewrites or rearrangements to enhance cognitive scaffolding. The revisions pursued two main goals: (1) to reduce the number of Retrieval-type tasks while increasing the emphasis on higher-order cognitive processing, and (2) to ensure cognitive support was provided in key areas of the activities. This process of assessing cognitive structure and revising to implement scaffolding demonstrates the effective use of the taxonomy as a diagnostic and developmental tool. Activity 12 serves as a representative example of our approach to implementing cognitive scaffolding, with both the original and revised versions included in the ESI.Fig. 3 illustrates the changes in cognitive levels of processing before and after the revisions.
image file: d5rp00225g-f3.tif
Fig. 3 Cognitive levels of processing for each question in Activity 12 for Cohorts 1 and 2. The black diagonal-patterned boxes in the Cohort 2 plot represent Pre-class questions (PcQ 1–3). Colored boxes highlight identical questions asked in both cohorts, with the two questions highlighted in different shades of blue indicating similar questions. Activity 12 is a representative example of the type of cognitive scaffolding revisions made to each of the activities for the semester-long course.

In the original Cohort 1 activities, students were instructed to watch a linked video related to the activity's content before attending discussion, where they would then summarize what they learned as a group. Most of these videos were replaced with “Pre-class Questions” (PcQs), which students completed individually before briefly comparing answers or definitions with group members. For Activity 12, these PcQs asked students to: list two strong and two weak acids (PcQ 1), describe what a resonance structure is (PcQ 2), and identify the features they use to determine acid strength and draw the conjugate base (PcQ 3). These PcQs established a shared foundation of knowledge, helping to minimize cognitive overload during more complex questions (van Merrienboer et al., 2003). They also supported the development of a shared language, consistent with other studies on scaffolding, and helped reduce confusion during peer or GTA interactions (Parker Siburt et al., 2011; Vo et al., 2022).

For the remainder of the activity, two of the authors revised the question order, added new prompts, or rewrote existing ones based on cognitive processing levels. In the original version, Question 4a–d asked students to write the reactions of a strong and weak acid in water, determine which proceeds to a greater extent, and justify their reasoning. Based on Marzano's taxonomy, Questions 4a and 4b are Retrieval-level, 4c is a Comprehension-level task that emphasizes key conceptual features, and 4d is an Analysis-level question requiring students to draw a connection between concepts. While this sequence shows some cognitive structuring, our goal was to reduce low-level tasks, which typically do little to promote collaborative learning (Kirschner et al., 2009; van de Pol et al., 2015). Additionally, by analyzing the cognitive levels of the questions, we recognized a disconnect between our intended focus for the activity and the structure of the questions, particularly in Question 4d, which asked students to connect reaction extent with acid strength. While this question reflects higher-order thinking, it promotes conceptual integration that students may not yet be ready for, especially as they are still developing foundational strategies for evaluating acid–base strength.

In the revised Cohort 2 activity, we redirected higher-order thinking toward understanding acid strength by adding Questions 2–5. These questions adopted a problematizing approach, prompting students to reassess their understanding and identify knowledge gaps (Phillips et al., 2017). Questions 2 and 3a are Comprehension-level tasks where students write their own method for evaluating acid strength and draw the conjugate base of given molecules. These tasks emphasize structural and electronic features in Lewis structures, such as formal charge, electronegativity, and resonance, encouraging representational reasoning grounded in visual interpretation. This prepares students for the Analysis-level Question 3b, which asks them to identify the stronger acid and justify their answer. Question 4 presents a new set of molecules for ranking in terms of acidity, while Question 5 prompts students to re-evaluate their original method for determining acid strength. These revisions ensure that higher-order tasks are now more intentionally scaffolded, supporting deeper understanding and promoting the transfer of knowledge to long-term memory.

The final questions in the revised (Cohort 2) activity shift focus on ranking base strength. While no direct cognitive scaffolding is provided for these questions, this aligns with the literature on fading support, which suggests that scaffolding should gradually decrease to foster independent problem-solving (Belland, 2014). Although the prior questions focused on identifying and reasoning about acid strength, students must now apply their learning to bases without direct guidance. Whether this is an ideal example of fading remains an open question, but it highlights the taxonomy's broader utility, especially for addressing gaps in the scaffolding literature related to investigating the impact of fading scaffolding (Vo et al., 2025). Future research could build on this by designing activities with varying types of cognitive scaffolding to systematically explore the effects of fading (see the work of Kreps et al. (2024) for further discussion around different types of cognitive structuring).

Each activity also included a reflection section, completed collaboratively before proceeding to the next set of content questions. These reflections incorporated categories from the ELIPSS rubrics (Cole et al., 2017; Czajka et al., 2021), with each reflection focused on a specific process skill. To guide discussion, the reflections included observable indicators and suggestions for improvement. In Cohort 1, reflections were presented as pauses within the activity. For Cohort 2, transitions were restructured to encourage more intentional and content-connected reflection. This change was motivated by research on constructive alignment and self-regulated learning, which emphasizes that reflection promotes active engagement when closely tied to the task (Biggs, 1996; Zimmerman, 2002). In the revised activities (Cohort 2), reflections began with a brief description of a process skill the group had just practiced, followed by a prompt to consider its importance before completing the rubric.

Analysis

This work used multiple codebooks to characterize the activities and each group's cognitive and social dynamics. For the activities, we employed the cognitive levels of processing defined in Marzano's taxonomy and the revised Anchoring Concepts Content Map (ACCM) for general chemistry (Marzano and Kendall, 2006; Holme et al., 2015). Knowledge dynamic, social processing, and modes of reasoning were used to categorize each group's interactions while completing the activities.
Activities. Marzano's taxonomy was used to establish the cognitive structure of the discussion activities. To ensure coding reliability, three researchers independently coded each question in each activity and then met to reach a consensus on each item. After coding, the first author revised each activity to enhance cognitive scaffolding. Subsequently, two researchers applied the cognitive levels of processing to the revised activities and again met to reach a consensus. The ACCM for general chemistry was similarly applied to the activities used for Cohorts 1 and 2. Two researchers independently applied the first level of ACCM to each question, learning objective, and skills in the activities, then met to discuss and come to a negotiated agreement.
Student groups. To characterize the groups' knowledge dynamic, social processing, and reasoning, codebooks from the work of Reid et al. (2022a,b) and Sevian and Talanquer (2014) were applied to the discourse from the group during each question using MAXQDA (VERBI Software, 2021). A full description of each code can be found in Appendix A.

There are four knowledge dynamic codes: not observable, knowledge sharing, knowledge application, and knowledge construction (Reid et al., 2022b). Knowledge sharing refers to groups exchanging information without it being added to or challenged by other members. A common example of this dynamic is the discourse surrounding Pre-class questions, where a student shares a definition of entropy that they have already written down, and the group merely agrees or says, “Sounds good to me.” Knowledge application refers to the use of information to solve a problem, with a clear understanding of its relevance. For instance, in a question asking which conjugate base is more stable, students first discuss stability before applying that knowledge to the specific molecules. The key distinction between knowledge sharing and knowledge application is that students actively channel the information to address the situation and arrive at an answer. Knowledge construction involves critiquing and building upon information. While knowledge application focuses on using information, knowledge construction emphasizes evaluating, critiquing, and testing the information. Not observable is just that; when no observable knowledge dynamic was seen, it would be coded as such.

The social processing codes are non-interactive, collaborative, tutoring, leader, domination, individualistic, confusion, multicomponent, and other (Reid et al., 2022b). Non-interactive was used when no social interactions were taking place. Collaborative represents groups that work together to find a solution to a question. Tutoring was used when a group's interaction primarily consisted of a student answering questions from other group members. Leader is representative of a single student constructing the response due to a lack of engagement from the other group members. Domination was used when a single student would construct a response while ignoring or rejecting other students’ input. Individualistic is representative of a group that does not interact with one another while constructing a response. Where individualistic differs from non-interactive is that individualistic includes an interaction where students may check-in on an answer that was developed independently. Confusion was used when the group spent the majority of time too confused to construct a response. Multicomponent was used for longer timestamps when the group displayed more than one dominant social interaction, and two other social processing codes would be selected as sub-codes for the interaction. The most common example of this was when students were confused about how to answer a question and spent an extended period trying to figure it out before asking for help from the GTA, after which they worked collaboratively on finishing the question and, therefore, would have been coded as multicomponent: confusion/collaborative. Other was used when the group's social interactions did not fall into another category and will be accompanied by a description when applicable. Inter-rater reliability (IRR) was determined using the Kappa statistical analysis for knowledge dynamic and social processing. Twenty percent of the interactions were randomly selected, and a final value of 0.80 and 0.83 was calculated for Cohorts 1 and 2, respectively, indicating strong agreement (McHugh, 2012).

The Modes of Reasoning codes are descriptive, relational, linear causal, and multi-component (Sevian and Talanquer, 2014). Descriptive reasoning refers to instances where a group's focus is on explicit features of a system or question without making connections to cause. For instance, a question that prompts students to rank acid strengths might elicit a response like, “This molecule has a lower pKa, so it must be stronger,” where the reasoning is based on a straightforward observation of a property. Relational was representative of when a group connected two features/properties of a system without explanation and contained an additional level: multi-relational, where multiple relations between features/properties were used (Weinrich and Talanquer, 2016). For example, students might expand on the previous reasoning with a statement like, “This molecule is more stable after losing hydrogen; therefore, it is stronger,” indicating a basic cause-and-effect relationship. Linear causal adds complexity by introducing a sequential chain of events, incorporating both cause and effect along with a justification or explanation. Here, students not only recognize relationships but also elaborate on them, explaining why increased stability would contribute to the strength of a molecule's acidity. Multi-component represents when a group identifies relevant interactions, and the effects of several features/properties are considered and weighted. Multi-component contains two subcodes, isolated where the effects of several variables are recognized and considered separately, and integrated where the effects of several variables are interconnected and systematically explained. For any case where no reasoning was present during discourse, an N/A code was applied. Two researchers independently coded the students' discourse for Cohort 1 and then discussed each code to come to a negotiated agreement. For Cohort 2, inter-rater reliability (IRR) was determined using the Kappa statistical analysis. Twenty percent of the interactions were randomly selected, and a final value of 0.73 was calculated, indicating substantial agreement (McHugh, 2012).

Results and discussion

In the results section, we compare the group dynamics to broadly examine the impact of our cognitive scaffolding and provide context for the subsequent analysis. These sections will be organized by each cohort's collective social processing, knowledge dynamics, and mode of reasoning. Pie charts containing the categorical data for each codebook and group can be found in Appendix B. The analysis will conclude with an examination of selected questions from Activity 12.

As shown in Fig. 4, the main result of the revisions was the redistribution of questions across different levels of processing. Although the number of Retrieval questions increased, most of these were Pre-class questions. The overall number of Comprehension questions decreased, while the Analysis level of processing saw the most pronounced increase. This reduction in Comprehension questions is important because, although useful for identifying critical features, these questions are less conducive to group learning compared to those requiring analysis (King, 2008). In the following sections of each analysis, Pre-class questions are excluded from the counts, as they were designed to be answered before the discussion and do not reflect the collaborative group learning environment. Additionally, only questions answered by both groups in each cohort are included in the analysis.


image file: d5rp00225g-f4.tif
Fig. 4 Counts for the number of questions at each cognitive level of processing across all fifteen activities.

The results from revising each activity's reflection were less impactful, as observations from Cohort 1 revealed that groups rarely completed reflections collaboratively. Similarly, the recorder for groups in Cohort 2 often completed the reflection independently, either before the rest of the group arrived or while they were engaged in off-topic discussions. Although literature suggests that scaffolding can promote collaborative reflection, a critical component is providing proper instruction and demonstration on how to engage effectively in these practices (Desautel, 2009; Harvey et al., 2016; Türkmen, 2024).

Social processing

In Fig. 5, we can see that students in Cohorts 1 and 2 were mostly collaborative while answering questions. Cohort 2 demonstrated more leader and tutoring dynamics for Analysis questions and a slight decrease in domination, confusion, and individualistic interactions. The ICAP framework suggests that higher engagement requires constructive dialogue, and an increase in interactions where students ask each other questions for clarification (tutoring) could indicate higher engagement. Another notable difference between the cohorts was the presence of confusion across all levels of cognitive processing. In Cohort 1, 14% (n = 30) of social processing codes contained confusion, whereas Cohort 2 demonstrated confusion in only 6% (n = 23) of coded discussions. This result could suggest that while our cognitive scaffolding may not have greatly impacted the social dynamics of Cohort 2, it may have promoted increased cognitive support as seen by the reduction of confusion. While these small differences in social dynamics make it difficult to attribute any results directly to cognitive scaffolding, they highlight another factor: the assigned group roles, which were intended to rotate weekly.
image file: d5rp00225g-f5.tif
Fig. 5 Collective distribution of social processing by cognitive levels of processing for Cohorts 1 and 2. N values represent the number of discourse segments coded at the corresponding cognitive level of processing.

Each activity prompted students to assign the roles of Manager, Recorder, Spokesperson, and Reflector, with the Recorder responsible for documenting the group's answers and submitting the completed activity. In Cohort 1, the red and gold groups consistently rotated roles each week, whereas in Cohort 2, the blue and purple groups did not, meaning that the same students served as Recorders every week. Notably, these Recorders were often the ones who constructed responses when the other students remained silent or answered questions from the other group members (leader and tutoring interactions). This finding underscores the importance of facilitation within a group learning environment and will be discussed further in a later section.

Knowledge dynamic

As seen in Fig. 6, both cohorts primarily engaged in knowledge sharing for Retrieval-type tasks. Cohort 1 completed a total of 61 Retrieval questions, while in Cohort 2, most of these questions were assigned as Pre-class questions to be completed prior to discussion. As such, they are not included in Fig. 6. If the Pre-class questions were included, the percentage of not observable and knowledge sharing dynamics for Cohort 2 would be 26% and 59% compared to 9% and 43%, respectively, without the Pre-class questions. This finding is noteworthy, as it lends credence to our decision to reposition the majority of the Retrieval questions to be completed before the discussion. It also aligns with existing literature on group learning, which indicates that Retrieval-type questions are generally less beneficial for collaborative learning compared to more cognitively complex questions (Kirschner et al., 2009; Scager et al., 2016).
image file: d5rp00225g-f6.tif
Fig. 6 Collective distribution of knowledge dynamics by cognitive levels of processing for Cohorts 1 and 2. N values represent the number of discourse segments coded at the corresponding cognitive level of processing.

In addition, Cohort 2 demonstrated greater engagement in knowledge dynamics broadly for the Analysis questions compared to Cohort 1, where a knowledge dynamic was not observed in 6% of interactions. This finding is notable when considering that Cohort 2 had triple the number of discursive interactions between students while working on higher-order questions (Analysis) than Cohort 1. Although the overall trends in knowledge dynamics reinforce the connection to question level and did not differ drastically with the addition of cognitive scaffolding, our use of Marzano's cognitive levels of processing in its development led to the inclusion of more higher-order questions. As a result, students in Cohort 2 spent more time at a higher level of cognitive engagement. This suggests that our approach to developing cognitive scaffolding supported sustained engagement across a broader range of questions by reducing cognitive load.

Modes of reasoning

The analysis of the students' modes of reasoning revealed findings similar to those observed in the knowledge dynamics, with the most notable difference being in the Analysis questions, as shown in Fig. 7. The modes of reasoning presented are not hierarchical or meant to imply one mode is more advantageous for students to use. Instead, they simply represent the complexity of students' reasoning (Weinrich and Talanquer, 2016). The prevalence of no reasoning (N/A) during Retrieval questions across both cohorts aligns with expectations based on cognitive processing levels and the ICAP framework, given that Retrieval questions primarily require basic recall rather than complex reasoning (Marzano and Kendall, 2006; Chi and Wylie, 2014; Chi et al., 2018). While the prevalence of descriptive and relational reasoning at the Analysis level of processing aligns with previous findings, it may also suggest that cognitive scaffolding was not entirely effective in encouraging students to engage in more complex reasoning. This indicates that additional support may be needed to foster the use of more complex reasoning (Moon et al., 2016; Deng and Flynn, 2021; Nennig et al., 2023).
image file: d5rp00225g-f7.tif
Fig. 7 Collective distribution of the modes of reasoning by cognitive levels of processing for Cohorts 1 and 2. N values represent the number of discourse segments coded at the corresponding cognitive level of processing.

Our future work will investigate this further, but as other studies have discussed, the reasoning used by students can vary depending on a myriad of factors. One factor we believe warrants closer examination is the impact of facilitator interactions on student reasoning. For example, in upcoming analyses, we will examine how teaching assistants interacted with groups in Cohort 2—particularly during a question where both cohorts expressed confusion. While this currently falls outside the scope of the present study, such analysis could help address a gap in the literature related to improving the facilitation of scaffolding through informed pedagogical training (Vo et al., 2025).

While completing Analysis questions, a greater proportion of Cohort 1 answered without providing any evidence of reasoning (N/A = 22%) compared to Cohort 2 (N/A = 12%), suggesting that Cohort 2 engaged in reasoning more consistently during these questions. The use of more complex reasoning, such as linear causal and multicomponent, remained consistent between the two cohorts. This result aligns with the findings on the group's knowledge dynamics, with the primary difference being an increased use of descriptive reasoning in Cohort 2. This suggests that our cognitive scaffolding may have both supported the sustained use of reasoning across more questions and encouraged greater use of low-level reasoning during higher-order questions, even if it did not necessarily lead to greater use of complex reasoning. This could indicate that to promote more complex reasoning, explicit instruction or guidance may be needed to deepen students' reasoning skills and fully engage them in more complex reasoning processes.

Literature on reasoning and argumentation highlights that scaffolding argument structures and prompting students to provide reasoning are effective strategies for fostering reasoning. However, without clear demonstrations or guidance on how to construct arguments or offer reasoning, students may view reasoning as just the answer to a question, rather than as an integral part of the science learning process (Ford, 2008; Russ et al., 2009; Leupen et al., 2020). Considering this, it seems that our cognitive scaffolding may have effectively encouraged greater student engagement through reasoning without necessarily increasing the complexity of that reasoning. To further iterate the point, the multicomponent reasoning in both cohorts was from the same groups (Cohort 1/gold group n = 1 and Cohort 2/purple group n = 3), implying that the group composition or dynamics may play a critical role in a group's use of more complex reasoning (Kumpulainen and Kaartinen, 2003; Jensen and Lawson, 2011; Nennig et al., 2023).

Activity 12

To better understand the role cognitive scaffolding in the revised activities played in group learning, we will discuss the differences between Cohorts 1 and 2 while working on Activity 12 (see the ESI for the entire activity). In Table 2, the knowledge dynamics used during Questions 2a–2e and 1a–1e for Cohorts 1 and 2, respectively, can be seen. The questions that remained the same or similar before and after the revision are colored accordingly.
Table 2 The assigned knowledge dynamic for each group in Cohort 1 and 2. Colored boxes highlight identical questions asked in both cohorts, with the two questions highlighted in different shades of blue indicating similar questions
image file: d5rp00225g-u1.tif


The primary difference between the activities is that Question 2d from Cohort 1 was removed, and Question 1c from Cohort 2 was added. Question 2d (Cohort 1) asked the students which hydrogen on acetic acid would be associated with a higher pKa value, while Question 1c (Cohort 2) asked the students to explain why electronegativity or resonance increased the molecule's stability. The reasoning behind the change was to reduce the change in concepts as defined by ACCM and more clearly promote identifying the critical features that can help determine the strength of an acid. Reducing the change in concepts helps to minimize cognitive load, helping students to integrate new information with their existing knowledge more effectively (Sweller, 2010, 2011). In this way, the change in the question represents context-specific scaffolding by adjusting the task to better align with the student's current level of understanding and focusing their attention on the most relevant concepts (McNeill et al., 2006; Ford, 2012; Leupen et al., 2020). By shifting the focus to the stability factors of the molecule, the revised question provides a more targeted opportunity for students to engage with the critical elements of acid strength, fostering deeper comprehension.

For the blue questions (2c and 1d), the students were asked to circle the most acidic proton from the molecule provided; the reason for the difference in the cognitive level of processing stems from the addition of asking students to justify their choice based on their answer to the previous question (1c). Thus, the students needed to connect the concepts of electronegativity, resonance, and stability to an acid's strength and were asked to engage in an Analysis level of processing. This shifted the knowledge dynamics for this question from knowledge sharing in Cohort 1 to knowledge application and construction in Cohort 2.

Table 3 contains the social processing codes applied to each group in Cohorts 1 and 2. While some studies indicate that scaffolding can positively influence groups' social dynamics, these studies have focused on online project-based learning and teacher-led scaffolding (Rojas-Drummond and Mercer, 2003; Kraatz et al., 2020; Cortázar et al., 2022). In contrast, our work focuses on the impact of using Marzano's taxonomy to develop activities that can be used as hard scaffolds, which is support that is static and pre-planned. Our results indicate that cognitive scaffolding did not have a large impact on social dynamics within the groups. For Cohort 1, both groups were collaborative, which, according to the ICAP framework, is a behavior that indicates higher engagement. However, a closer examination of the group's conversations reveals that, while they collaborated to answer questions, they may not have been as collaborative in explaining or reasoning through their answers. For example, while the gold group was working on Question 2e, Hodgkin was not sure how they could provide an explanation for the entropic argument Hodgkin: I have no idea how to go about explaining the other side,” and Daly responded by explaining Daly: The other side would be entropically…because you lose that proton and have that negative ion.” While it initially appears that the two students are working collaboratively to build their understanding, confusion about which atoms on the molecule they are referring to sets in, ultimately leading them to give up on constructing the explanation: Daly: I feel like I can kind of explain it both ways but… I don’t really know what it is [exactly], I would still [say] more energetically affected.”

Table 3 The assigned social processing for each group in Cohort 1 and 2. Colored boxes highlight identical questions asked in both cohorts, with the two questions highlighted in different shades of blue indicating similar questions. The Confusion/Leader code for the purple group, question 1c, is an example of the multicomponent code. Where the group was first too confused to progress the question before an intervention, at which point their social dynamic was Leader
image file: d5rp00225g-u2.tif


For the blue group, who mainly worked collaboratively, the students explicitly mentioned early on that “[they are] trying to get this done so [they] can go watch the [eclipse].” The desire to get the activity done quickly led them to start answering questions before Alice arrived. As the group starts working on Question 1d, Alice arrives, and the rest of the group puts no effort into getting them caught up, eventually leading Alice to try and contribute without being fully aware of what the group is working on.

Alice: [Reading off their personal iPad] As electronegativity goes up, size goes down. Is that what you were asking?
Sheeana: No, we were talking about stability.
Alice: Oh, sorry
Sheeana: No, you are good.

Leupen et al. (2020) states that “…teams should always be working on the same problem; otherwise, teams will be less motivated to be interested when others report,” and the above dialogue is a good example of this, especially when we consider that Alice does not participate for the remaining questions. In addition, both the red and purple groups would begin reading the next question while the recorder was still writing out the answer to their current question. Given that these more complex social interactions are present across both cohorts, this finding suggests that cognitive scaffolding alone may not be sufficient to positively impact a group's social dynamics. The social processing codes for the purple group illustrate how social dynamics can shift quickly, as one student consistently took on the role of recorder for almost every week and was the source of the domination during their discussion. Literature identifies factors such as this among relevant prior knowledge/experiences, holding students accountable for their answers, and specifically, facilitation methods as crucial for promoting high-quality social dynamics (Dallimore et al., 2006; Knight et al., 2016; Pecore et al., 2017; Leupen et al., 2020; Newton et al., 2020).

Table 4 contains the progression in each group's reasoning while working through the questions. For this section, we will focus on Questions 2d and 1c for Cohorts 1 and 2, respectively. For the gold group, there was initial confusion about what the pKa means Hodgkin: Well a lower pKameans more stable, right? Or is it the opposite, that it is a strong acid?” before they were ultimately able to land on a relational type of reasoning Daly: Just because comparatively it is not as electronegative, therefore [carbon] can’t hold that negative charge as well as an oxygen would.” The red group relied more on their general understanding of pKa to dictate their reasoning, leading them to engage in descriptive reasoning: Bohr: Since it is weaker, it would have a higher pKa…and if it was stronger, it would have a lower pKa.”

Table 4 The assigned mode of reasoning for each group in Cohort 1 and 2. Colored boxes highlight identical questions asked in both cohorts, with the two questions highlighted in different shades of blue indicating similar questions
image file: d5rp00225g-u3.tif


For Cohort 2, the question was modified to better align with the concept of scaffolding and cognitive structuring, which aims to guide students through complex topics by prompting them to consider the ‘why’ behind a concept before applying it to higher-order questions. By adjusting the question to better correspond with the concepts explored in other questions and emphasizing the reasoning behind electronegativity and stability, we expected to promote interactive or constructive engagement and more complex reasoning in their work (Chi and Wylie, 2014). The blue group begins this question with confusion before they settle on descriptive reasoning for their answer.

Dorothy: I don’t know why, all I know is that it is that way.
Sheeana: So, if it is more electronegative, it holds the electrons tighter.
Dorothy: Okay, that makes sense.
Olga: It makes them less available to form bonds.

This type of reasoning is similar to the red group from Cohort 1, which focuses on restating the salient entities without any justification or explanation (Weinrich and Talanquer, 2016). According to the ICAP framework, reasoning that merely reasserts explicit properties (descriptive reasoning) and applies shared knowledge (knowledge application) reflects active engagement, which does not align with the expected interactive or constructive engagement for this type of question. However, it is important to consider that the group had previously expressed a desire to finish the activity quickly to observe the eclipse. By communicating this intention, the group had already committed to limiting their engagement in favor of completing the task rapidly (Draskovic et al., 2004). In contrast, the purple group showed a markedly different outcome. For instance, during this question, we observe a key aspect of engagement and reasoning: critique (Michaels et al., 2008; Osborne, 2013). Below is a section of the transcribed group discussion during Question 1c.

Helen: Why do we want it to have a higher electronegativity? What is the actual reason? Because I know it has something to do with the negative charge but don’t we… [Linus starts talking]
Linus: I think it has to do with size and how easy it gives up its proton. Or…
Helen: That's size.
Linus: Oh yeah, that is size and not electronegativity.
Miles: They are more stable when they have a higher electronegativity because you don’t want it to change, you don’t want it to just be giving them up, or accepting them. Right? That is what makes it more stable
Helen: Hmm… My thing is why. Because more electronegative means that they take more electrons, wouldn’t that concentrate it more? I want to know the reason
Miles: Oh yeah, that's a good question.

In this discussion, both Linus and Miles proposed potential explanations for the effect of electronegativity on molecular stability, while Helen challenged their ideas by critiquing the reasoning. Although the question was coded as relational reasoning, this classification is primarily limited by the student's inability to continue the dialogue due to their confusion, which eventually led them to seek guidance from the GTA. Through the lens of the ICAP framework, this scenario exemplifies interactive or constructive engagement: the students not only articulate their reasoning but also critically engage with and evaluate each other's ideas. This collaborative process results in a co-construction of knowledge.

Conclusions

In this study, we aimed to demonstrate how Marzano's taxonomy can be used to develop cognitively scaffolded group activities. Our results indicate that this scaffolding was effective in reducing cognitive load while promoting sustained cognitive engagement and reasoning, while also increasing the time students spent engaging with higher-order questions. These results align with cognitive load theory, which describes how cognitively structured activities can help students process complex material without becoming overwhelmed, fostering a more manageable and focused learning experience (de Jong, 2010; Chang and Yang, 2023; Hao et al., 2023). The results also align with current literature on scaffolding, suggesting that Marzano's taxonomy can be used by facilitators and researchers as a diagnostic and developmental tool for cognitive scaffolding.

Determining the effectiveness of scaffolding is no easy task, as it requires careful consideration of a myriad of factors (Palincsar, 1998; van de Pol et al., 2010). To this end, we employed coding schemes that aided in our analysis of the group's social and knowledge dynamics and reasoning. While the cognitive scaffolding led to increased cognitive engagement and more frequent use of lower-level reasoning, it did not appear to support the development of more complex reasoning or consistently benefit the social dynamics within groups. Moreover, while the scaffolding encouraged basic reasoning processes, the limited presence of complex reasoning may reflect broader curricular constraints—suggesting that instructional approaches must go beyond activity design to more explicitly and consistently promote advanced forms of reasoning throughout the course. These findings indicate that cognitive scaffolding with Marzano's taxonomy alone may not be sufficient to support effective collaboration or deeper reasoning. Our future work will explore how facilitation strategies influence group interactions and whether informed pedagogical approaches can enhance the impact of scaffolding during group work.

Limitations

The small number of student groups observed limits the generalizability of our findings. Understanding the impact of cognitive scaffolding on group learning dynamics is a complex topic requiring multiple codebooks. Therefore, purposeful sampling was used to select groups that mainly worked together on the questions to ensure the analysis could be performed. Consequently, the sample size may not be representative of all settings and populations. Additionally, all the activities were presented in a second-semester introductory chemistry course, with energy/thermodynamics, kinetics, and equilibrium as the primary concepts; therefore, the findings may not represent all chemistry curricula. Our findings were developed from audio and video recordings with discussion sections taught by different GTA's. Although we attempted to limit the number of variables by selecting discussion sections that occurred on the same days of the week between Cohorts 1 and 2, many variables are challenging to control and may have impacted findings.

Implications

We believe the primary implications of this work lie in demonstrating how Marzano's taxonomy can be used to both develop and assess cognitive scaffolding. Rather than being constrained by the specific content or type of scaffolding, this approach is grounded in established cognitive processes of learning. As such, it can be applied broadly—from the design of individual in-class activities to the restructuring of entire courses.

Although the design and impact of scaffolding remains an active area of research, we believe Marzano's taxonomy provides a much-needed foundation for a shared language and approach to future studies. For instance, while our work did not address metacognitive scaffolding directly, metacognition is explicitly included in the taxonomy. This presents a clear opportunity for future research to explore how metacognitive supports might be incorporated into cognitive scaffolding without requiring an entirely separate framework. Additionally, there has been a call for further investigation into the fading of scaffolds, and our previous work provides examples of what that may look like using Marzano's taxonomy and how it can aid in further research on the topic (Kreps et al., 2024).

While the impact of our cognitive scaffolding on social interactions was less evident, these outcomes were likely shaped by other variables, such as group dynamics, instructional facilitation, and the overall classroom environment. Future research should examine how these contextual factors influence the effectiveness of scaffolding, particularly in collaborative learning settings. Doing so would not only address gaps in the literature but also contribute to the development of more robust facilitation practices and pedagogical training. By adopting a shared, cognitively grounded framework, facilitators across disciplines can more easily communicate instructional strategies and goals, ultimately making pedagogical practices more transferable and comprehensible. This benefit can also extend to students. When instructors are able to clearly communicate the purpose and benefits of cognitive scaffolding, it fosters greater transparency and can help build student trust in the learning process (Vo et al., 2022).

Although more research is needed to refine and expand this approach to cognitive scaffolding, our findings already point to several practical applications for instructors and facilitators. While we focused on revising in-class discussion activities, the taxonomy can also guide the design of higher-order thinking questions and help identify points where students may need additional support. For example, recognizing that a task requires Analysis-level processing might prompt the instructor to include an intermediate Comprehension-level prompt, such as a guided line of inquiry, to scaffold student reasoning. In this way, Marzano's taxonomy serves as both a developmental and diagnostic tool, supporting more intentional and effective instructional design.

Conflicts of interest

There are no conflicts to declare.

Data availability

Data collected from human participants, as described in the paper, are not available for confidentiality reasons.

Appendices

A. Codebook definitions

Tables 5–8.
Table 5 The four cognitive levels of processing as defined by Marzano and Kendall (2006)
Level of cognitive processing Definition
Level 1 – Retrieval Recalling information from a provided source or permanent memory.
Level 2 – Comprehension Identification of critical features needed to transfer knowledge from working memory to permanent.
Level 3 – Analysis Creating new insights from previous knowledge or using previously obtained knowledge in novel situations (focus is on the concept/skill).
Level 4 – Knowledge utilization Using knowledge for novel and specific situations (focus is on the situation).


Table 6 The four knowledge dynamic codes and definitions from Reid et al. (2022a,b)
Category Definition
Knowledge sharing The focus of the group interactions is based on sharing information to answer the question without questioning of the utterances presented
Knowledge application The focus of the group interactions is based on applying a formula/method/concept and relating that to a clear understanding of how it relates to the explanation or process of solving the problem
Knowledge construction The focus of group interactions is based on sharing information and building upon the ideas of others by questioning or critiquing the ideas presented
Not observable No knowledge dynamic is seen due to a lack of student interaction


Table 7 Social processing codes and definitions from Reid et al. (2022a,b)
Category Definition
Non-interactive Students are not having a conversation, and there is no proof of individualistic work
Collaborative Students are co-constructing ideas and generating products together
Tutoring One or more students ask questions that another student responds to by guiding the students through the problem asking for the tutees ideas or just by explaining their reasoning without asking input from the student who asked the question
Leader One student primarily constructs the response due to a lack of contribution from others
Domination One student constructs the response for the group while not considering, ignoring, or rejecting input given by others
Individualistic Students are working independently and are not having conversations about the question products
Confusion Students are too confused to really generate the expected product or make confident progress for a question
Multicomponent Students engage in more than one social processing
Other Describes a way that students are interacting with one another that is not described by another code


Table 8 Modes of reasoning codes and definitions from Sevian and Talanquer (2014) and Weinrich and Talanquer (2016)
Category Definition
Descriptive Salient entities in a system are identified or recognized. Explicit properties are described, verbalized. Functions or properties of entities are seen as sufficient explanation for their behavior. A phenomenon is seen as an instantiation of reality; it may be re-described by merely asserting that things are as they are without referring to causes. Reasoning mostly based on experiences and knowledge from daily life. Strong influence of surface similarity and recognition on judgment and decision making.
Relational Salient entities in a system are identified or recognized. Explicit and implicit differentiating properties are highlighted. Spatial or temporal relations between entities are noticed. Correlations between properties and behaviors are established but not explained or justified. A phenomenon is seen as an effect of a single entity, the natural outcome of a single property or the linear combination of several properties; no mechanisms are proposed. Reduction of variables and overgeneralization constrain reasoning.
Multi-relational Explanation based on multiple relations
Linear causal Salient entities in a system are identified or recognized. Explicit and implicit differentiating properties are highlighted. Spatial or temporal organization of and connections between entities and noticed. Relevant direct interactions between entities invoked. Although the influence of many factors may be recognized, phenomena tend to be seen as (reduced to) the result of the actions of a single agent on other entities; proposed mechanisms involve linear cause-effect relations and sequential chains of events. Reduction of variables and overgeneralization frequently constrain reasoning.
Multicomponent Salient entities in a system are identified or recognized. Explicit and implicit differentiating properties are highlighted. Spatial or temporal organization of and connections between entities are noticed. Relevant interactions between entities are invoked. Effects of several variables are considered and weighed.
Isolated Complex phenomena are seen as the result of the static and dynamic interplay of more than one factor and the direct interactions of several components. Effects of several variables are considered and weighed separately.
Integrated Complex phenomena are seen as the result of the dynamic interplay of more than one factor and the direct and indirect interactions of several components. Explanations as interconnected stories of how variables affect the entities involved. The effects of different variables are more thoroughly and systematically explained for multiple entities involved.
NA No mode of reasoning is present


B. Group data

Fig. 8–10.
image file: d5rp00225g-f8.tif
Fig. 8 Pie charts representing the percentage of each knowledge dynamic code for Cohort 1 (Gold and Red Group) and Cohort 2 (Blue and Purple Group).

image file: d5rp00225g-f9.tif
Fig. 9 Pie charts representing the percentage of each social processing code for Cohort 1 (Gold and Red Group) and Cohort 2 (Blue and Purple Group).

image file: d5rp00225g-f10.tif
Fig. 10 Pie charts representing the percentage of each mode of reasoning for Cohort 1 (Gold and Red Group) and Cohort 2 (Blue and Purple Group).

Acknowledgements

We would like to acknowledge the National Science Foundation (DUE Grant #1915047) for funding this work. The authors would like to acknowledge the participants who were willing to share their discussion materials and group work with the team for analysis. We would also like to acknowledge the work and insight provided by Hannah Nenning during the coding process.

References

  1. Belland B. R., (2014), Scaffolding: Definition, Current Debates, and Future Directions, in Handbook of Research on Educational Communications and Technology, Spector J. M., Merrill M. D., Elen J. and Bishop M. J. (ed.), Springer, pp. 505–518 DOI:10.1007/978-1-4614-3185-5_39.
  2. Biggs J. B., (1987), Student Approaches to Learning and Studying. Research Monograph, Australian Council for Educational Research Ltd.
  3. Biggs J., (1996), Enhancing teaching through constructive alignment, High. Educ., 32(3), 347–364 DOI:10.1007/BF00138871.
  4. Broman K., Bernholt S. and Parchmann I., (2018), Using model-based scaffolds to support students solving context-based chemistry problems, Int. J. Sci. Educ., 40(10), 1176–1197 DOI:10.1080/09500693.2018.1470350.
  5. Chang C.-C. and Yang S.-T., (2023), Interactive effects of scaffolding digital game-based learning and cognitive style on adult learners’ emotion, cognitive load and learning performance, Int. J. Educ. Technol. High. Educ., 20(1) DOI:10.1186/s41239-023-00385-7.
  6. Chen O., Paas F. and Sweller J., (2023), A Cognitive Load Theory Approach to Defining and Measuring Task Complexity Through Element Interactivity, Educ. Psychol. Rev., 35(2), 63 DOI:10.1007/s10648-023-09782-w.
  7. Chi M. T. H., (2009), Active-Constructive-Interactive: A Conceptual Framework for Differentiating Learning Activities, Top. Cogn. Sci., 1(1), 73–105 DOI:10.1111/j.1756-8765.2008.01005.x.
  8. Chi M. T. H., Adams J., Bogusch E. B., Bruchok C., Kang S. and Lancaster M. et al., (2018), Translating the ICAP Theory of Cognitive Engagement Into Practice, Cogn. Sci., 42(6), 1777–1832 DOI:10.1111/cogs.12626.
  9. Chi M. T. H. and Wylie R., (2014), The ICAP Framework: Linking Cognitive Engagement to Active Learning Outcomes, Educ. Psychol., 49(4), 219–243 DOI:10.1080/00461520.2014.965823.
  10. Cole R., Lantz J., and Ruder S., (2017), ELIPSS: Enhancing Learning by Improving Process Skills in STEM, Available at https://elipss.com/(Accessed: 20 June 2025).
  11. Cole R., Reynders G., Ruder S., Stanford C. and Lantz J., (2019), Constructive Alignment Beyond Content: Assessing Professional Skills in Student Group Interactions and Written Work, in Research and Practice in Chemistry Education: Advances from the 25th IUPAC International Conference on Chemistry Education 2018, Schultz M., Schmid S. and Lawrie G. A. (ed.), Springer, pp. 203–222 DOI:10.1007/978-981-13-6998-8_13.
  12. Cortázar C., Nussbaum M., Alario-Hoyos C., Goñi J. and Alvares D., (2022), The impacts of scaffolding socially shared regulation on teamwork in an online project-based course, Internet High. Educ., 55, 100877 DOI:10.1016/j.iheduc.2022.100877.
  13. Criswell B. A., (2012), Reducing the degrees of freedom in chemistry classroom conversations, Chem. Educ. Res. Pract., 13(1), 17–29 10.1039/C2RP00002D.
  14. Czajka D., Reynders G., Stanford C., Cole R., Lantz J. and Ruder S., (2021), A Novel Rubric Format for Providing Feedback on Process Skills to STEM Undergraduate Students, J. College Sci. Teach., 50(6), 48–56 DOI:10.1080/0047231X.2021.12290533.
  15. Dallimore E. J., Hertenstein J. H. and Platt M. B., (2006), Nonvoluntary Class Participation in Graduate Discussion Courses: Effects of Grading and Cold Calling. J. Manage. Educ., 30(2), 354–377 DOI:10.1177/1052562905277031.
  16. Davis E. A. and Miyake N., (2004), Guest Editors’ Introduction: Explorations of Scaffolding in Complex Classroom Systems, J. Learn. Sci., 13(3), 265–272.
  17. DeChurch L. A. and Mesmer-Magnus J. R., (2010), The cognitive underpinnings of effective teamwork: a meta-analysis, J. Appl. Psychol., 95(1), 32–53 DOI:10.1037/a0017328.
  18. de Jong T., (2010), Cognitive load theory, educational research, and instructional design: some food for thought, Instr. Sci., 38(2), 105–134 DOI:10.1007/s11251-009-9110-0.
  19. Deng J. M. and Flynn A. B., (2021), Reasoning, granularity, and comparisons in students’ arguments on two organic chemistry items, Chem. Educ. Res. Pract., 22(3), 749–771 10.1039/D0RP00320D.
  20. Desautel D., (2009), Becoming a Thinking Thinker: Metacognition, Self-Reflection, and Classroom Practice, Teachers College Rec., 111(8), 1997–2020 DOI:10.1177/016146810911100803.
  21. Draskovic I., Holdrinet R., Bulte J., Bolhuis S. and Leeuwe J. V., (2004), Modeling small group learning, Instruct. Sci., 32(6), 447–473 DOI:10.1007/s11251-004-2276-6.
  22. Ford M., (2008), ‘Grasp of Practice’ as a Reasoning Resource for Inquiry and Nature of Science Understanding, Sci. Educ., 17(2), 147–177 DOI:10.1007/s11191-006-9045-7.
  23. Ford M. J., (2012), A Dialogic Account of Sense-Making in Scientific Argumentation and Reasoning, Cogn. Instruct., 30(3), 207–245 DOI:10.1080/07370008.2012.689383.
  24. Forman E., (1989), The role of peer interaction in the social construction of mathematical knowledge, Int. J. Educ. Res., 13(1), 55–70 DOI:10.1016/0883-0355(89)90016-5.
  25. Fredricks J. A., Blumenfeld P. C. and Paris A. H., (2004), School Engagement: Potential of the Concept, State of the Evidence, Rev. Educ. Res., 74(1), 59–109 DOI:10.3102/00346543074001059.
  26. Gencer S., (2023), Development and Use of Flowchart for Preservice Chemistry Teachers’ Problem Solving on the First Law of Thermodynamics, J. Chem. Educ., 100(9), 3393–3401 DOI:10.1021/acs.jchemed.3c00224.
  27. Hao X., Xu Z., Guo M., Hu Y. and Geng F., (2023), The effect of embedded structures on cognitive load for novice learners during block-based code comprehension, Int. J. STEM Educ., 10(1), 42 DOI:10.1186/s40594-023-00432-9.
  28. Harvey M., Coulson D. and McMaugh A., (2016), Towards a theory of the Ecology of Reflection: Reflective practice for experiential learning in higher education, J. Univ. Teach. Learn. Practice, 13(2), 1–22 DOI:10.53761/1.13.2.2.
  29. Heisterkamp K. and Talanquer V., (2015), Interpreting Data: The Hybrid Mind, J. Chem. Educ., 92(12), 1988–1995 DOI:10.1021/acs.jchemed.5b00589.
  30. Hoffman M. M. and Richardson S., (2019), Team Construction and Accountability, POGIL, Routledge.
  31. Holme T., Luxford C. and Murphy K., (2015), Updating the General Chemistry Anchoring Concepts Content Map, J. Chem. Educ., 92(6), 1115–1116 DOI:10.1021/ed500712k.
  32. James N. M. and LaDue N. D., (2021), Pedagogical Reform in an Introductory Chemistry Course and the Importance of Curricular Alignment, J. Chem. Educ., 98(11), 3421–3430 DOI:10.1021/acs.jchemed.1c00688.
  33. Janssen J., Kirschner F., Erkens G., Kirschner P. A. and Paas F., (2010), Making the Black Box of Collaborative Learning Transparent: Combining Process-Oriented and Cognitive Load Approaches, Educ. Psychol. Rev., 22(2), 139–154 DOI:10.1007/s10648-010-9131-x.
  34. Jensen J. L. and Lawson A., (2011), Effects of Collaborative Group Composition and Inquiry Instruction on Reasoning Gains and Achievement in Undergraduate Biology, LSE, 10(1), 64–73 DOI:10.1187/cbe.10-07-0089.
  35. Johnson D. W., Johnson R. T. and Smith K., (2007), The State of Cooperative Learning in Postsecondary and Professional Settings, Educ. Psychol. Rev., 19(1), 15–29 DOI:10.1007/s10648-006-9038-8.
  36. Kararo A. T., Colvin R. A., Cooper M. M. and Underwood S. M., (2019), Predictions and constructing explanations: an investigation into introductory chemistry students’ understanding of structure–property relationships, Chem. Educ. Res. Pract., 20(1), 316–328 10.1039/C8RP00195B.
  37. King A., (2008), Structuring Peer Interaction to Promote Higher-Order Thinking and Complex Learning in Cooperating Groups, in The Teacher's Role in Implementing Cooperative Learning in the Classroom, Gillies R. M., Ashman A. F. and Terwel J. (ed.), US: Springer, pp. 73–91 DOI:10.1007/978-0-387-70892-8_4.
  38. Kirschner F., Paas F. and Kirschner P. A., (2009), Individual and group-based learning from complex cognitive tasks: effects on retention and transfer efficiency, Comput. Human Behav., 25(2), 306–314 DOI:10.1016/j.chb.2008.12.008.
  39. Knight J. K., Wise S. B. and Sieke S., (2016), Group Random Call Can Positively Affect Student In-Class Clicker Discussions, LSE, 15(4), ar56 DOI:10.1187/cbe.16-02-0109.
  40. Kraatz E., Nagpal M., Lin T.-J., Hsieh M.-Y., Ha S. Y., Kim S. and Shin S., (2020), Teacher Scaffolding of Social and Intellectual Collaboration in Small Groups: A Comparative Case Study, Front. Psychol., 11 DOI:10.3389/fpsyg.2020.587058.
  41. Kreps A., Brown I., Wenzel T. J. and Cole R., (2024), Structuring Materials to Support Student Learning: Analysis of Instructional Materials from a Professional Development Workshop, J. Chem. Educ., 101(11), 4603–4613 DOI:10.1021/acs.jchemed.4c00783.
  42. Kumpulainen K. and Kaartinen S., (2003), The Interpersonal Dynamics of Collaborative Reasoning in Peer Interactive Dyads, J. Exp. Educ., 71(4), 333–370 DOI:10.1080/00220970309602069.
  43. Leupen S. M., Kephart K. L. and Hodges L. C., (2020), Factors Influencing Quality of Team Discussion: Discourse Analysis in an Undergraduate Team-Based Learning Biology Course, LSE, 19(1), ar7 DOI:10.1187/cbe.19-06-0112.
  44. Marzano R. J. and Kendall J. S., (2006), The New Taxonomy of Educational Objectives, Corwin Press, A SAGE Publications Company.
  45. McHugh M. L., (2012), Interrater reliability: the kappa statistic, Biochem. Med., 22(3), 276–282.
  46. McNeill K. L., Lizotte D. J., Krajcik J. and Marx R. W., (2006), Supporting Students’ Construction of Scientific Explanations by Fading Scaffolds in Instructional Materials, J. Learn. Sci., 15(2), 153–191 DOI:10.1207/s15327809jls1502_1.
  47. Mercer N., Dawes L., Wegerif R. and Sams C., (2004), Reasoning as a scientist: ways of helping children to use language to learn science, Br. Educ. Res. J., 30(3), 359–377 DOI:10.1080/01411920410001689689.
  48. Michaels S., O’Connor C. and Resnick L. B., (2008), Deliberative Discourse Idealized and Realized: Accountable Talk in the Classroom and in Civic Life, Stud. Philos. Educ., 27(4), 283–297 DOI:10.1007/s11217-007-9071-1.
  49. Moon A., Stanford C., Cole R. and Towns M., (2016), The nature of students’ chemical reasoning employed in scientific argumentation in physical chemistry, Chem. Educ. Res. Pract., 17(2), 353–364 10.1039/C5RP00207A.
  50. Moon A., Stanford C., Cole R. and Towns M., (2017a), Analysis of inquiry materials to explain complexity of chemical reasoning in physical chemistry students’ argumentation, J. Res. Sci. Teach., 54(10), 1322–1346 DOI:10.1002/tea.21407.
  51. Moon A., Stanford C., Cole R. and Towns M., (2017b), Decentering: A Characteristic of Effective Student–Student Discourse in Inquiry-Oriented Physical Chemistry Classrooms, J. Chem. Educ., 94(7), 829–836 DOI:10.1021/acs.jchemed.6b00856.
  52. Moreira P., Marzabal A. and Talanquer V., (2019), Using a mechanistic framework to characterise chemistry students’ reasoning in written explanations, Chem. Educ. Res. Pract., 20(1), 120–131 10.1039/C8RP00159F.
  53. Nennig H. T., States N. E., Macrie-Shuck M., Fateh S., Gunes Z. D. K. and Cole R. et al., (2023), Exploring social and cognitive engagement in small groups through a community of learners (CoL) lens, Chem. Educ. Res. Pract., 24(3), 1077–1099 10.1039/D3RP00071K.
  54. Newton D. W., LePine J. A., Kim J. K., Wellman N. and Bush J. T., (2020), Taking engagement to task: the nature and functioning of task engagement across transitions, J. Appl. Psychol., 105(1), 1–18 DOI:10.1037/apl0000428.
  55. Nokes-Malach T. J., Richey J. E. and Gadgil S., (2015), When Is It Better to Learn Together? Insights from Research on Collaborative Learning, Educ. Psychol. Rev., 27(4), 645–656 DOI:10.1007/s10648-015-9312-8.
  56. Nonaka I., von Krogh G. and Voelpel S., (2006), Organizational Knowledge Creation Theory: Evolutionary Paths and Future Advances, Organization Studies, 27(8), 1179–1208 DOI:10.1177/0170840606066312.
  57. Noyes K., Carlson C. G., Stoltzfus J. R., Schwarz C. V., Long T. M. and Cooper M. M., (2022), A Deep Look into Designing a Task and Coding Scheme through the Lens of Causal Mechanistic Reasoning, J. Chem. Educ., 99(2), 874–885 DOI:10.1021/acs.jchemed.1c00959.
  58. Noyes K. and Cooper M. M., (2019), Investigating Student Understanding of London Dispersion Forces: A Longitudinal Study, J. Chem. Educ., 96(9), 1821–1832 DOI:10.1021/acs.jchemed.9b00455.
  59. Osborne J., (2013), The 21st century challenge for science education: assessing scientific reasoning, Think. Ski. Creat., 10, 265–279 DOI:10.1016/j.tsc.2013.07.006.
  60. Palincsar A. S., (1998), Keeping the metaphor of scaffolding fresh--a response to C. Addison Stone's “The metaphor of scaffolding: its utility for the field of learning disabilities”, J. Learn. Disabil., 31(4), 370–373 DOI:10.1177/002221949803100406.
  61. Parker Siburt C. J., Bissell A. N. and Macphail R. A., (2011), Developing Metacognitive and Problem-Solving Skills through Problem Manipulation, J. Chem. Educ., 88(11), 1489–1495 DOI:10.1021/ed100891s.
  62. Patton M. Q., (2014), Qualitative Research & Evaluation Methods: Integrating Theory and Practice, SAGE Publications.
  63. Pecore J. L., Kirchgessner M. L., Demetrikopoulos M. K., Carruth L. L. and Frantz K. J., (2017), Formal Lessons Improve Informal Educational Experiences: The Influence of Prior Knowledge on Student Engagement, Visitor Studies, 20(1), 89–104 DOI:10.1080/10645578.2017.1297134.
  64. Phillips A. M., Watkins J. and Hammer D., (2017), Problematizing as a scientific endeavor, Phys. Rev. Phys. Educ. Res., 13(2), 020107 DOI:10.1103/PhysRevPhysEducRes.13.020107.
  65. Premo J., Wyatt B. N., Horn M. and Wilson-Ashworth H., (2022), Which Group Dynamics Matter: Social Predictors of Student Achievement in Team-Based Undergraduate Science Classrooms, LSE, 21(3), ar51 DOI:10.1187/cbe.21-06-0164.
  66. Reid J. W., Gunes Z. D. K., Fateh S., Fatima A., Macrie-Shuck M. and Nennig H. T. et al., (2022a), Investigating patterns of student engagement during collaborative activities in undergraduate chemistry courses, Chem. Educ. Res. Pract., 23(1), 173–188 10.1039/D1RP00227A.
  67. Reid J. W., Gunes Z. D. K., Fateh S., Fatima A., Macrie-Shuck M. and Nennig H. T. et al., (2022b), Investigating patterns of student engagement during collaborative activities in undergraduate chemistry courses, Chem. Educ. Res. Pract., 23(1), 173–188 10.1039/D1RP00227A.
  68. Reynders G., Lantz J., Ruder S. M., Stanford C. L. and Cole R. S., (2020), Rubrics to assess critical thinking and information processing in undergraduate STEM courses, Int. J. STEM Educ., 7(1), 9 DOI:10.1186/s40594-020-00208-5.
  69. Rojas-Drummond S. and Mercer N., (2003), Scaffolding the development of effective collaboration and learning, Int. J. Educ. Res., 39(1), 99–111 DOI:10.1016/S0883-0355(03)00075-2.
  70. Russ R. S., Coffey J. E., Hammer D. and Hutchison P., (2009), Making classroom assessment more accountable to scientific reasoning: a case for attending to mechanistic thinking, Sci. Educ., 93(5), 875–891 DOI:10.1002/sce.20320.
  71. Scager K., Boonstra J., Peeters T., Vulperhorst J. and Wiegant F., (2016), Collaborative Learning in Higher Education: Evoking Positive Interdependence, CBE Life Sci. Educ., 15(4), ar69 DOI:10.1187/cbe.16-07-0219.
  72. Sevian H. and Talanquer V., (2014), Rethinking chemistry: a learning progression on chemical thinking, Chem. Educ. Res. Pract., 15(1), 10–23 10.1039/C3RP00111C.
  73. Sheppard M. A. W. and Bauer C. F., (2023), Student Conceptions of pH Buffers Using a Resource Framework: Layered Resource Graphs and Levels of Resource Activation, J. Chem. Educ., 100(1), 22–33 DOI:10.1021/acs.jchemed.1c01078.
  74. Shvarts A. and Bakker A., (2019), The early history of the scaffolding metaphor: Bernstein, Luria, Vygotsky, and before. Mind, Culture, Activity, 26(1), 4–23 DOI:10.1080/10749039.2019.1574306.
  75. Stanford C., Moon A., Towns M. and Cole R., (2016), Analysis of Instructor Facilitation Strategies and Their Influences on Student Argumentation: A Case Study of a Process Oriented Guided Inquiry Learning Physical Chemistry Classroom, J. Chem. Educ., 93(9), 1501–1513 DOI:10.1021/acs.jchemed.5b00993.
  76. States N. E., Lovig C., Martin K., Nennig H. T. and Cole R. S., (2023), Let Students Work: Analysis of the Role of Differing Facilitation on Student Engagement in a Large Stadium-Style Lecture Hall, J. Chem. Educ., 100(11), 4237–4248 DOI:10.1021/acs.jchemed.3c00750.
  77. Sweller J., (1994), Cognitive load theory, learning difficulty, and instructional design. Learn. Instruct., 4(4), 295–312 DOI:10.1016/0959-4752(94)90003-5.
  78. Sweller J., (2010), Element Interactivity and Intrinsic, Extraneous, and Germane Cognitive Load, Educ. Psychol. Rev., 22(2), 123–138 DOI:10.1007/s10648-010-9128-5.
  79. Sweller J., (2011), Cognitive Load Theory, in Psychology of Learning and Motivation, Mestre J. P. and Ross B. H. (ed.), Academic Press, ch. 2, pp. 37–76 DOI:10.1016/B978-0-12-387691-1.00002-8.
  80. Toledo S. and Dubas J. M., (2016), Encouraging Higher-Order Thinking in General Chemistry by Scaffolding Student Learning Using Marzano's Taxonomy, J. Chem. Educ., 93(1), 64–69 DOI:10.1021/acs.jchemed.5b00184.
  81. Tharp R. G. and Gallimore R., (1988), Rousing minds to life: Teaching, learning, and schooling in social context, Cambridge University Press.
  82. Theobald E. J., Hill M. J., Tran E., Agrawal S., Arroyo E. N., Behling S., et al., (2020), Active learning narrows achievement gaps for underrepresented students in undergraduate science, technology, engineering, and math, Proc. Natl. Acad. Sci. U. S. A., 117(12), 6476–6483 DOI:10.1073/pnas.1916903117.
  83. Towns M. H., (1998), How Do I Get My Students to Work Together? Getting Cooperative Learning Started, J. Chem. Educ., 75(1), 67 DOI:10.1021/ed075p67.
  84. Türkmen G., (2024), A Systematic Review of Group Metacognition Researches on Online and Face-to-Face Learning Environments, J. Learn. Teach. Digital Age, 9(2), 64–84.
  85. van Aalst J., (2009), Distinguishing knowledge-sharing, knowledge-construction, and knowledge-creation discourses, Comput. Supp. Learn., 4(3), 259–287 DOI:10.1007/s11412-009-9069-5.
  86. van de Pol J., Volman M. and Beishuizen J., (2010), Scaffolding in Teacher–Student Interaction: A Decade of Research, Educ. Psychol. Rev., 22(3), 271–296 DOI:10.1007/s10648-010-9127-6.
  87. van de Pol J., Volman M., Oort F. and Beishuizen J., (2015), The effects of scaffolding in the classroom: support contingency and student independent working time in relation to student achievement, task effort and appreciation of support, Instr. Sci., 43(5), 615–641 DOI:10.1007/s11251-015-9351-z.
  88. van Merrienboer J. J. G., Kirschner P. A. and Kester L., (2003), Taking the Load Off a Learner's Mind: Instructional Design for Complex Learning, Educ. Psychol., 38(1), 5–13 DOI:10.1207/S15326985EP3801_2.
  89. VERBI Software, (2021), MAXQDA 2022, computer program, VERBI Software, Berlin.
  90. Vo K., Sarkar M., White P. J. and Yuriev E., (2022), Problem solving in chemistry supported by metacognitive scaffolding: teaching associates’ perspectives and practices, Chem. Educ. Res. Pract., 23(2), 436–451 10.1039/D1RP00242B.
  91. Vo K., Sarkar M., White P. J. and Yuriev E., (2025), Exploring Problem-Solving Scaffolds in General Chemistry: A Systematic Review of Scaffolding Goals and Instructional Approaches, J. Chem. Educ., 102(3), 1004–1018 DOI:10.1021/acs.jchemed.4c01327.
  92. Weinrich M. L. and Talanquer V., (2016), Mapping students’ modes of reasoning when thinking about chemical reactions used to make a desired product, Chem. Educ. Res. Pract., 17(2), 394–406 10.1039/C5RP00208G.
  93. Wilson K. J., Brickman P. and Brame C. J., (2018), Group Work, CBE Life Sci. Educ., 17(1), fe1 DOI:10.1187/cbe.17-12-0258.
  94. Wood D., Bruner J. S. and Ross G., (1976), The Role Of Tutoring In Problem Solving, J. Child Psychol. Psychiatry, 17(2), 89–100 DOI:10.1111/j.1469-7610.1976.tb00381.x.
  95. Yuriev E., Naidu S., Schembri L. S. and Short J. L., (2017), Scaffolding the development of problem-solving skills in chemistry: guiding novice students out of dead ends and false starts, Chem. Educ. Res. Pract., 18(3), 486–504 10.1039/c7rp00009j.
  96. Zimmerman B. J., (2002), Becoming a Self-Regulated Learner: An Overview, Theory Into Pract., 41(2), 64–70 DOI:10.1207/s15430421tip4102_2.
  97. Zulu E., Haupt T., Tramontin V. and Tramontin V., (2018), Cognitive Loading Due To Self-Directed Learning, Complex Questions And Tasks In The Zone Of Proximal Development Of Students, Problems Educ. 21st Century, 76(6), 864–880 DOI:10.33225/pec/18.76.864.

Footnote

Electronic supplementary information (ESI) available. See DOI: https://doi.org/10.1039/d5rp00225g

This journal is © The Royal Society of Chemistry 2025
Click here to see how this site uses Cookies. View our privacy policy here.