Ryan S.
Bowen
*a,
Aishling A.
Flaherty
b and
Melanie M.
Cooper
a
aMichigan State University, Department of Chemistry, East Lansing, Michigan, USA. E-mail: bowenrya@msu.edu
bUniversity of Limerick, School of Education, Limerick, Ireland
First published on 4th March 2022
Within chemistry education, there are various curricular and pedagogical approaches that aim to improve teaching and learning in chemistry. Efforts to characterize these transformations have primarily focused on student reasoning and performance, and little work has been done to explore student perceptions of curricular and pedagogical transformations and whether these perceptions align with the transformational intent. To complement our previous work on the Organic Chemistry, Life, the Universe, and Everything (OCLUE) curriculum, we developed this exploratory study to determine if students had perceived the goals of the transformation. As in our previous research on OCLUE, we compared perceptions between OCLUE and a more traditional organic chemistry course. Using inductive and deductive qualitative methodologies, we analyzed student responses to three open-eneded questions focused on how students perceived they were expected to think, what they found most difficult, and how they perceived they were assessed. The findings were classified into three superodinate themes: one where students perceived they were expected to learn things as rote knowledge, such as memorization (“Rote Knowledge”), another where students perceived they were expected to use their knowledge (“Use of Knowledge”), and responses that used vague, generalized language, were uninformative, or did not address the questions asked (“Other”). Students in these two courses responded very differently to the open-ended questions with students in OCLUE being more likely to perceive they were expected to use their knowledge, while students in the traditional course reported rote learning or memorization more frequently. As the findings evolved, our interpretations and discussions were influenced by sociocultural perspectives and other cultural frameworks. We believe this approach can provide meaningful insights into transformational intent and certain features of classroom cultures.
Within the CER and science education literature, there are many studies exploring student perceptions within the affective domain of learning and student experiences across entire courses or programs (Bauer, 2005, 2008; Galloway and Bretz, 2016; Galloway et al., 2016; Flaherty, 2020a). For example, longitudinal studies such as Talking About Leaving (Seymour and Hewitt, 1997) and Talking About Leaving Revisited (Thiry et al., 2019) have leveraged student perceptions and found that students perceive competitive, unsupportive class cultures in many of their STEM courses, including chemistry. According to students in these studies, the class cultures, in conjunction with many other factors, ultimately contributed to their decision to switch out of their STEM majors (Seymour and Hewitt, 1997; Thiry et al., 2019). More recently, studies have explored student perceptions of their chemistry courses following the shift to online instruction during the 2020–2022 COVID-19 pandemic (Ramachandran and Rodriguez, 2020).
Outside of the affective domain, research on student perceptions of learning has also been common. For example, one study explored how students interpreted structure, property, and function relationships across biology and chemistry. The authors found that while students could discuss structure and properties in the context of both courses, students had more difficulty discussing function in the context of chemistry (Kohn et al., 2018). Such work is supported by previous research that has found that students may miss crucial information during instruction that could aid their understanding which causes them to not perform as well as they intended despite their success in earlier chemistry courses (Anderson and Bodner, 2008). In the context of undergraduate laboratories, work with course-based undergraduate research experiences (CUREs) found that students demonstrated gains in their perceived knowledge, experience, and confidence with specific research-related abilities. The authors concluded that such perceptions could help instructors with course evaluation and assessment design (Irby et al., 2020).
Student perceptions have also been leveraged to better understand how students engage in critical thinking. For example, Scott studied student perceptions of critical thinking after they completed a technology course where debate was employed as a pedagogical tool and found that students perceived their critical thinking abilities had been enhanced (Scott, 2008). Similarly, Hammersley-Fletcher and Hanley used student perceptions to explore the ways that international students in the UK viewed critical thinking and concluded that students thought that certain approaches associated with critical thinking silenced their voices (Hammersley-Fletcher and Hanley, 2016). Finally, in a study investigating student, teaching staff, and employer perceptions of the definition of critical thinking in chemistry communities, Danczak and colleagues found that definitions across the groups differed and that students perceived “critique”, “objectivity”, and “problem-solving” were all components of critical thinking (Danczak et al., 2017).
All of these examples highlight the robust and insightful nature of student perceptions in chemistry education, making them a significant area of research. In a study related to this work, co-author AAF employed constructivist grounded theory to investigate student perceptions of the structure and development of scientific knowledge within the transformed organic chemistry course discussed here. After interviewing twelve students in the transformed course, the findings indicated that students perceived memorization of content was not as effective as being able to reason, that students needed to critique information by interrogating prior knowledge, and that students recognized differences in explaining how and why chemical phenomena occur, among others (Flaherty, 2020b). Though this initial study was influential for our work discussed here, it's important to note that it asked fundamentally different research questions and was not comparative. That is, it was focused on student perceptions of the structure and development of scientific knowledge and focused exclusively on students in the transformed course without comparison of their perceptions to students in other organic chemistry environments. Regardless, the findings pushed us to pursue this line of inquiry further.
Although student perceptions had been leveraged in a variety of ways, to our knowledge, they had not been used to further assess transformation efforts and to ascertain whether student understanding of course goals aligned with instructor expectations. Therefore, we found student perceptions of expectations and what was valued to be a significant area of study for four reasons: (1) it complemented our previous research of our transformational efforts at our institution (Crandell et al., 2019, 2020; Houchlei et al., 2021); (2) it afforded another perspective and way to characterize our transformation that did not focus on student reasoning; (3) it allowed us to explore alignment between our transformational intent, expectations, and student perceptions (and ascertain whether there was misalignment); and (4) considering our transformational efforts were informed by research on how people learn and think (National Research Council, 2000, 2012b, 2012a; National Academies of Sciences, 2018), this study would enable us to investigate if student perceptions of what was expected and valued in courses aligned with the evidence base on effective ways of doing and thinking.
With these motivations in mind, we embarked on this exploratory study. However, as we will explain more later, the study evolved as we interfaced with and interpreted the data. Influenced by the Talking About Leaving studies (Seymour and Hewitt, 1997; Thiry et al., 2019), we were reminded that student perceptions can be used to provide insights on elements of the classroom culture. Studies have shown that alignment between course goals and classroom practices can lead to a more productive learning experience and engagement with scientific practices (Sandoval et al., 2019). Furthermore, we recognized that certain classroom norms communicate implicit and explicit messages to students about how to participate, think, and practice (Becker et al., 2013; Chang and Song, 2016; Reinholz and Apkarian, 2018). Considering that we were interested in knowing what students perceived they were doing in these courses and whether they aligned with our transformational intent, our interpretations and discussions evolved to consider the classroom cultures of the two organic chemistry courses in this study.
Our previously published research compared student performance on a variety of tasks, including constructing causal mechanistic explanations and the use of mechanistic arrows, across a two semester sequence of a transformed organic chemistry course (Crandell et al., 2019, 2020; Houchlei et al., 2021). As a result, we have insights on student thinking and their approaches to such tasks. Therefore, we opted to engage in this exploratory study where we investigated student perceptions of two organic chemistry courses (including our transformed course) that, in our opinion and from our previous research, employed different approaches to teaching and learning. One of the courses was transformed using three-dimensional learning (National Research Council, 2012a; 3DL4US, n.d.), and the other embodied a more traditional approach to organic chemistry (as further discussed below). Considering that our previous studies afforded us insights into how students responded to different types of organic chemistry tasks, the work presented here attempted to characterize the course experiences from student perspectives. Our motivations for this work were driven by an interest in complementing this previous work and to characterize our transformation efforts from a different perspective. Just as some CER scholars have argued that student perceptions can inform assessment design (Irby et al., 2020), we assert that student perceptions of what is expected and valued can inform course design and transformational efforts. Furthermore, this study enabled us to explore alignment of student perceptions of what they are expected to do, what is valued, our transformational intent, and the evidence base on effective ways of doing and thinking. As we will note later, given the research on the role that alignment between classroom practices, course goals, and norms of participation and practice have within the classroom culture, we ultimately discuss and situate this work within a sociocultural perspective that is informed by culture scholars (Vygotsky, 1978; Rogoff, 1990; John-Steiner and Mahn, 1996; Carlone et al., 2011; Becker et al., 2013; Chang and Song, 2016; Schein and Schein, 2016; Reinholz and Apkarian, 2018; Sandoval et al., 2019; Zotos et al., 2020; Petterson et al., 2022).
In order to gather this data, we needed an instrument that would help us capture student perceptions in a robust way while minimizing external influences on student responses. Although there are a number of previously developed instruments for use in higher education that address student perceptions, expectations, and other affective states, none of them met the needs of this study; therefore, we opted to develop our own. Our instrument involved three open-ended questions which will be discussed in more detail later. These questions specifically target student perceptions of how they were expected to think in organic chemistry, what they found most difficult in the course, and how they perceived they were assessed. However, first, we find it important to review some of these instruments to justify the development of our own.
The Colorado Learning Attitudes about Science Survey (CLASS), is also a Likert-scale instrument developed for physics (Adams et al., 2005) and adapted for chemistry (Barbera et al., 2008) and biology (Semsar et al., 2011). The CLASS instruments are primarily focused on gathering information from students on their beliefs and attitudes about learning within the specific discipline, the content of the discipline, the structure of the disciplinary knowledge, and connections to the “real world”. In contrast to the MPEX and CHEMX, the CLASS asks about the discipline in general while the MPEX and CHEMX instruments probe student beliefs about a specific course. Just as with the MPEX and CHEMX, results from the CLASS are reported as how well they align with expert-responses, and typically there is no “improvement”. That is, there is no movement to more expert-like responses over a general chemistry sequence. However, these three instruments differ in that the MPEX and the CHEMX cluster responses using confirmatory factor analysis while the CLASS clusters according to exploratory factor analysis.
Although not immediately related to expectations, other instruments have been developed to explicitly measure student attitudes. Some instruments, such as the Chemistry Attitudes and Experiences Questionnaire (CAEQ) (Dalgety et al., 2003) and the Attitude towards the Subject of Chemistry Inventory (ASCI) (Bauer, 2008) have utilized a semantic differential format where students respond on a scale where the extremes include polar opposite adjectives. In the case of the ASCI, the structure of the instrument begins with a sentence stem such as “Chemistry is…” and then students respond to the sentence stem by rating their response on a 7-point semantic differential scale where the extremes represent aforementioned polar opposite adjectives such as “easy/hard”, “comprehensible/incomprehensible”, and “tense/relaxed”, among others. The ASCI has been further developed, producing the ASCIv2 (Xu and Lewis, 2011) and the ASCIv3 (Rocabado et al., 2019).
While the use of Likert- and semantic differential-scale instruments allow for quick diagnostics and analysis of the data, the questions within these instruments may prompt students to respond in a certain way, do not allow for students to state their experience in their own words, and students may not be given the opportunity to volunteer information that they deem as most important or relevant to their experience. These restrictions signify that more open-ended questions coupled with qualitative methodologies could be helpful discovering themes that capture a more accurate picture of student experience. Though some of the items in previous instruments were investigating similar ideas as we are here, the potential for prompting inherent in the questions and the lack of opportunities for students to use their own words may not accurately capture student perceptions, beliefs, or attitudes. Furthermore, qualitative approaches to investigate perceptions, expectations, and other constructs is scarce (Flaherty, 2020a), and we believed this to be a great opportunity to explore student perceptions of their organic chemistry courses in an open-ended way.
With this said, previously published instruments in CER and other fields tended to address how students experienced a course or whole discipline and did not align with our study objectives. Our goal was rather different in that we were interested in how students perceived course/instructor expectations and what was valued. Since we wanted to use students’ own words and perceptions to guide our investigation, minimize prompting in the questions, and use qualitative methodologies to analyze the responses, we opted to use our own instrument. Such an approach allowed for a combination of inductive and deductive coding, highlighted student voices, and provided students the opportunity to identify what they believed to be most important and relevant to their experiences.
Considering that this study uses student perceptions of what is expected and valued and ways of practicing, it is important to acknowledge that we (and students, for that matter) have assumptions and ideas about what it means to know and do. Broadly, our epistemological beliefs are informed by constructivist and sociocultural views of learning where we believe that students construct their own knowledge and are influenced by the contexts in which learning occurs and the interactions they have (Vygotsky, 1978; Bodner, 1986; Rogoff, 1990; John-Steiner and Mahn, 1996; National Research Council, 2000; Carlone et al., 2011; National Academies of Sciences, 2018; Zotos et al., 2020; Petterson et al., 2022). With this, we also ascribe to the resources perspective which asserts that students have knowledge that is connected in various ways which may or may not be activated when prompted depending on their knowledge structure and how the task is scaffolded. Furthermore, it is acknowledged that the resources students have may be more or less productive on a given learning task which offers a way to understand how students are connecting and applying concepts (Hammer, 2000). With all of this said, we ascribe to the idea that people learn best when they are in environments that provide them consistent opportunities to apply and use their knowledge (and resources). Considering that three-dimensional learning engages students in scientific practices around fundamental ideas in chemistry, it resonates with our epistemological beliefs and is the foundation for our course transformations as will be discussed in the Methods section (National Research Council, 2012a; 3DL4US, n.d.). Coupled with our previous work on student reasoning in the context of the courses in this study, we acknowledge these beliefs and previous research influenced our analysis and interpretations of student perceptions.
Lectures in OCLUE are somewhat interactive. New topics are introduced by having students discuss what they already know, clicker questions are posed, students are encouraged to discuss the answers, and occasional group activities are incorporated (for example, groups build molecular models and compare them together). Students work in groups in OCLUE recitation sections to complete scaffolded worksheets which include a mixture of three-dimensional and more traditional questions, such as draw a reaction mechanism or determine the identity of an unknown compound from spectroscopic data. Homework is assigned twice a week for credit upon completion rather than accuracy to encourage students to try and practice without penalty and also includes three-dimensional prompts similar to the recitation activities. Therefore, a considerable proportion of a students’ grade in OCLUE is determined by participating, practicing, and trying with “good faith effort” and explaining how and why something happens, thus allocating a smaller proportion of the students’ grade to high stakes testing. Examinations in OCLUE employ a mixture of multiple-choice and open-response items, some of which mirror traditional questions in an organic chemistry course (such as predicting products and drawing mechanisms). However, frequently students are asked to provide an explanation of how and why a given chemical phenomena is occurring with about 50% of the points on exams focusing on having students use core ideas in the context of scientific practices.
In contrast, traditional courses are usually organized by functional group. Rather than connecting a few types of reaction mechanisms to the core ideas, a traditional course tends to treat each type of reaction and functional group separately. By agreement between all instructors, the same topics are covered, the course is primarily taught in a traditional expository lecture format. While students can (and do) ask questions, there is no expectation of peer interactions either in the lecture or in the recitation sections. Instead, the recitation sections for the traditional course typically consist of a quiz, followed by a question-and-answer period, or another short lecture from the graduate teaching assistant. Students may complete online homework, typically multiple-choice questions, however the homework is not completed for a grade. The examinations consist of open-response items where students must fill in the reactant, reagent, or products, draw a mechanism for a reaction, or design a synthesis, and are typical high stakes summative assessments. These items are similar to those that we have found are prevalent in sophomore organic chemistry courses, and our prior analysis of these items indicates that students are typically not required to explicitly show evidence of reasoning, but rather can they answer questions by recall or pattern recognition (Stowe and Cooper, 2017).
It's worth noting that the overall assessment strategies for the two courses also differ significantly. In OCLUE, between 45–50% of the points are allocated through formative assessment strategies. That is, group work in recitation and homework are not graded for accuracy but on completion with a “good faith effort”. The rest of the overall grade in OCLUE comes from three mid-terms and a final exam. In contrast, in the traditional sections all of the points towards the class grade come from summative exams (midterms and final). This difference may have significant consequences for students since there is emerging evidence that allocating parts of the course grade to completion of formative assessments is a more equitable strategy that can address differences in outcomes among various demographic groups (Tashiro and Talanquer, 2021).
In summary, the two types of courses cover the same material, but they have different pedagogical approaches, course requirements, and approaches to assessments. Given that the purpose of this study was to complement previous research by characterizing our transformational efforts from the student perspective and to generate insights on how student perceptions aligned with our transformational intent and subsequent theories of learning, we found this to be an informative study. As will be discussed in more detail later, our interpretive frame of classroom culture clarifies these purposes by helping us acknowledge that the implicit and explicit messages sent by the course and instructors communicate what are valued ways of knowing and doing. Alignment between classroom practices, course goals, messages instructors send, and the interpretations of those messages by students as well as classroom norms have been shown to be important for engaging students in learning practices (Becker et al., 2013; Chang and Song, 2016; Schein and Schein, 2016; Sandoval et al., 2019).
Gender | First-generation | Transfer | |||
---|---|---|---|---|---|
Female | 693 | Yes | 202 | Yes | 152 |
Male | 310 | No | 801 | No | 851 |
Total | 1003 | Total | 1003 | Total | 1003 |
Major | Ethnicity | ||
---|---|---|---|
Life Sciences | 733 | American Indian/Alaskan Native | 1 |
Lab Sciences | 61 | Asian (non-hispanic) | 91 |
Physical Sciences | 24 | Black or African–American (non-hispanic) | 66 |
Engineering | 4 | Hawaiian/Pacific Islander (non-Hispanic) | 1 |
Animal Sciences and Veterinary | 58 | Hispanic | 37 |
Food and Nutritional Sciences | 36 | International | 49 |
Social Sciences | 27 | Not reported | 11 |
Other | 60 | Two or more races (non-hispanic) | 37 |
White (non-hispanic) | 710 | ||
Total | 1003 | Total | 1003 |
The first question stated: “If you met a student who is thinking about enrolling in (traditional or OCLUE) organic chemistry next year, how would you describe the ways students are expected to think about reaction mechanisms in organic chemistry?” By mentioning mechanisms, we intended to scaffold student responses and help them reflect on course expectations. Furthermore, there is ample research to show that students have great difficulty with thinking about mechanisms and often resort to memorization as a way to succeed (Bhattacharyya and Bodner, 2005). This question also related to our previous work where we have shown that students in OCLUE are more likely to engage in causal mechanistic reasoning, to use mechanisms appropriately, and are significantly more likely than traditional peers to correctly predict products for unknown reactions (Houchlei et al., 2021).
The second question was the following: “What would you tell them is the most difficult thing about organic chemistry?” Previously published instruments often asked students about the difficulty of the overall course or specific content, and we thought this question would provide students the opportunity to identify aspects that they deemed most difficult without constraining their response. The CER literature has detailed various aspects of organic chemistry that students have difficulty with and by investigating student perceptions on the most difficult aspect of the course in this open-ended way, we can gather insights into which facets of a course students struggle with the most, such as a certain way of thinking, a course policy, or instruction in general.
The third question was “How would you describe to them what is assessed in organic chemistry?” This question was designed to elicit if students perceived that assessments aligned with how they perceived they were expected to think in the course. It is well recognized that assessments send strong messages to students about what is valued in a course (Momsen et al., 2013; Stowe et al., 2021). Considering that the approaches to summative and formative assessments were quite different between the two courses in this study, we believed this question would be insightful.
The design of these questions aligned with our goals of the study. We were interested in complementing our previous research on student reasoning and characterizing OCLUE from the student perspective which all three questions would address. In addition, we were interested in investigating whether student perceptions aligned with our course goals and expectations and the underlying theories of learning that informed the OCLUE curriculum. We anticipated that questions 1 (expectations of thinking) and 3 (assessment) would be an open-ended way to explore this alignment while question 2 (most difficult thing) provided additional information on how the enactment of the course impacted difficulties encountered by students.
Inductive thematic analysis was deemed most suitable in the beginning because: (1) we wanted categories to emerge from student experiences at first to guide our analysis and highlight their voices; (2) we had highly open-ended questions; and (3) our initial stance toward this project was exploratory in nature and we believed beginning with an inductive approach was appropriate. Therefore, analysis was conducted as described by Thomas (Thomas, 2006). First, the raw data files were formatted to promote ease of comprehension. Second, we familiarized ourselves with the nature of the data by reading the student responses. Third, categories were identified and defined based on actual phrases or meanings in specific text segments. Finally, each category was continually revised based on the ongoing analysis of data. To establish the codebook, responses from 248 traditional students and all 248 OCLUE students were analyzed. Taken together these 496 students answered the three open-ended questions mentioned earlier, yielding 1488 responses across all three questions, and representing over 58% of the total data. Two of the authors (RSB and AAF) went through several rounds of independent coding of the 1488 responses, developing and revising the codebooks for each question each time. Upon settling on a semi-finalized codebook, the authors then calculated percent agreement and found they had an 86.1% agreement. After discussing the coding discrepancies and sharpening the code dimensions to yield the finalized codebook, the authors settled on a 99.4% agreement. With such a high percent agreement, the authors concluded that any additional measure of inter-rater reliability would not be necessary. The remaining set of data was then split in half between RSB and AAF and coded to yield the full set of analyzed data (all 2556 responses). Throughout the coding process, mutually exclusive codes were identified and used. The decision to use mutually exclusive coding was based on the following reasons:
(1) The overall majority of the responses could only be categorized by a single code.
(2) An analysis by author RSB using non-mutually exclusive coding yielded almost identical overall patterns. This is provided in the ESI,† Fig. S1–S3.
(3) The use of mutually exclusive codes allowed for quicker and more efficient coding of the 2556 responses.
Code | Dimensions | Example quotes |
---|---|---|
Apply and reason | Student responses that include one or more of the following dimensions in regard to their perceptions of how they are expected to think about mechanisms: (1) understanding “why” a reaction proceeds; (2) the use of knowledge, specifically with the use of fundamental or basic ideas; (3) the transfer of knowledge to new problems; (4) making connections between concepts, especially in order to apply them; (5) making predictions in order to solve a problem | Traditional: “I would tell them not to memorize them but to actually think through each of them and the reasoning behind why what happens, happens.” |
OCLUE: “Should expect to understand the molecular interactions of reactions and WHY these occur.” | ||
Identify and describe | Student responses that include one or more of the following dimensions in regard to their perceptions of how they are expected to think about mechanisms: (1) understanding the “what” and “how” reactions proceed, particularly without mentioning the use of knowledge to understanding “why” reactions proceed; (2) mentions understanding at the scalar or one scalar below levels, particularly through the recognition of concepts such as polarity and electronegativity and their significance to understanding; (3) when a student explicitly mentions understanding the mechanism instead of memorization; (4) when a student mentions “differentiating” between reactions with any further explanation; (5) responses includes a discussion of forces, charges, or stabilization | Traditional: “They have to think about the polarity of bonds and the nature of atoms when reacting with other atoms in regard to electronegativity and polarity” |
OCLUE: “The reaction mechanism are meant to show the transfer of electrons from one compound/atom to another. This helps show how these reactions occur.” | ||
Apply heuristics | Student responses that include one or more of the following dimensions in regard to their perceptions of how they are expected to think about mechanisms: (1) focuses on the approach to solving problems rather than thinking about a problem; (2) mentions explicit use of arrow pushing without mentioning how knowledge is used to engage in the formalism; (3) provide descriptive statements without causal or mechanistic knowledge such as “negatives attack positives” or “source goes to sink”; (4) mention of identifying patterns and trends without expanding on the significance of identifying these patterns and trends; (4) explicit mention of the movement or flow of electrons without further explanation of how the movement or flow of electrons influence reactions | Traditional: “think about it in terms of Nu- attacks E +” |
OCLUE: “they need to think about mechanistic arrows as the movement of electrons from a source to a sink” | ||
Memorization | Student responses that include one or more of the following dimensions regarding their perceptions of how they are expected to think about mechanisms: (1) memorization, remembering, recalling, or regurgitation of reactions, products, reagents, and/or mechanisms; (2) knowing reactions, products, reagents, and/or mechanisms, particularly with no explicit mention of understanding the “what”, “how”, or “why” a reaction proceeds | Traditional: “Memorize, memorize, memorize.” |
OCLUE: “memorize what reacts with what” | ||
Generalities | Student responses that include one or more of the following dimensions in regard to their perceptions of how they are expected to think about mechanisms: (1) the total absence of any chemistry in the response; (2) thinking of reactions and/or mechanisms as “puzzles”; (3) thinking of reactions and/or mechanisms on a “step-by-step” basis; (4) using generic and unclear descriptors for thinking such as thinking critically, conceptually, creatively, thoroughly, or differently, particularly if the student does not expand on what they mean; (5) stating general facts about reactions and mechanisms such as “reactants go to products” or that there are many mechanisms for a given reaction; (6) seeing organic chemistry as a new or different language; (7) mentioning that organic chemistry focuses on the details; (8) mentioning that mechanisms are “like a story” | Traditional: “You have to think about would they would benefit from most if they reacted.” |
OCLUE: “You need to think rationally rather than memorize.” | ||
Not applicable | Student responses that include one or more of the following dimensions in regard to their perceptions of how they are expected to think about mechanisms: (1) the response does not answer the question, such as organic chemistry or mechanisms are “easy” and “straightforward”; (2) when a student provides no response at all; (3) when the response is unclear or interpretation difficult, such as when students say “you must understand/know the material”; (4) mention of actions students must do in organic chemistry such as “study a lot”; (5) mention of exam and/or course aspects such as exam difficulty and the challenging nature of organic chemistry; (6) when a student is venting about the course, professor, or other aspects relevant to the course | Traditional: “To do the practice problems in the text book and make flash cards.” |
OCLUE: “good” |
The “Apply and Reason” and the “Identify and Describe” categories differ with respect to whether students noted the significance of knowing why a mechanism occurs. For example, if a student mentioned the existence of forces and stabilization in their response, this was coded as “Identify and Describe.” However, if the student mentioned the existence of forces and stabilization, and then expanded their response to include a discussion of how this helps explain why a reaction happens, then the response was coded as “Apply and Reason.” While “Apply and Reason” responses were considered more sophisticated than “Identify and Describe”, we acknowledge the complexity and potential understanding exhibited in “Identify and Describe” responses.
Although we did not set out to develop a hierarchical model of categories during the analysis, the progression from “Memorization” to “Apply and Reason” does suggest a greater degree of sophistication in students’ perceptions of how they were expected to think about organic chemistry mechanisms. Though it is possible that a student can memorize a heuristic, the “Apply Heuristics” and “Memorization” categories were considered separately because using a heuristic does require some form of application that simply memorizing does not. The “Generalities” and the “Not Applicable” categories were also identified and included in the analysis. The “Generalities” category included responses where students explained the need to think about organic chemistry mechanisms using general or vague terms such as “critically”, “conceptually”, “creatively”, “thoroughly”, or “differently”; that is, they did not appear to have developed a specific vocabulary for what they were doing. Also included in this category were responses which referred to the need to think about mechanisms in a step-by-step manner, solving puzzles, or telling stories since these responses often did not expand on what was meant. Responses that were answering the question but seemed to not refer to chemistry concepts were also included here because we believed they were answering the question, but their meaning and context was uncertain. The “Not Applicable” category included instances when students did not give any response at all, or when their response was unclear or unrelated to the question posed.
The analysis of student responses from the OCLUE and traditional courses revealed differences in the perceptions of expectations of thinking regarding reaction mechanisms. As noted in Fig. 1, more OCLUE students perceived the need to engage in more sophisticated ways of thinking about mechanisms than students in the traditional course. For example, 30.6% (n = 76) of OCLUE students perceived that they were expected to apply what they knew to navigate their way through new and unforeseen problems and to provide a reason for why these mechanisms proceed. This compares to 13.2% (n = 80) of students from the traditional course perceiving the same. For the category “Identify and Describe”, 21.4% (n = 53) of OCLUE students and 17.4% (n = 105) of students from the traditional course perceived this expectation. While the number of students from both types of courses who perceived the need to apply heuristics were quite similar (12.9%, n = 78, of the traditional students and 10.9%, n = 27, of OCLUE students), there was a much larger difference in extent to which students perceived they had to memorize material. Of the traditional student cohort, 20.9% (n = 126) of students perceived the need to memorize information on organic chemistry mechanisms with just 2.8% (n = 7) of OCLUE students perceiving the same. Finally, more students from the traditional course (24.5%, n = 148) used general terms to describe how they were expected to think about organic chemistry mechanisms than OCLUE students (16.1%, n = 40). The “Not Applicable” category included instances when students did not give any response at all (and there were very few of these responses across all three questions), or when their response was unclear or unrelated to the question posed. More students from the OCLUE course (18.1%, n = 45) had responses categorized to the “Not Applicable” category than students from the traditional course (11.1%, n = 67).
Code | Dimensions | Example quotes |
---|---|---|
Apply and reason | Student responses that include one or more of the following dimensions in regard to their perceptions of what is the most difficult thing about organic chemistry: (1) understanding “why” a reaction proceeds; (2) the use of knowledge, specifically with the use of fundamental or basic ideas; (3) the transfer of knowledge to new problems; (4) making connections between concepts or piecing/linking concepts together, especially in order to apply them; (5) making predictions in order to solve a problem | Traditional: “Understanding why mechanism happen the way they do” |
OCLUE: “Realizing that you are not going to memorize every reaction, you just need to worry about patterns and reasons why things happen a certain way” | ||
Identify and describe | Student responses that include one or more of the following dimensions in regard to their perceptions of what is the most difficult thing about organic chemistry: (1) understanding the “what” and “how” reactions proceed, particularly without mentioning the use of knowledge to understanding “why” reactions proceed; (2) mentions understanding at the scalar or one scalar below levels, particularly through the recognition of concepts such as polarity and electronegativity and their significance to understanding; (3) when a student explicitly mentions understanding the mechanism instead of memorization; (4) when a student mentions “differentiating” between reactions with any further explanation; responses includes a discussion of forces, charges, or stabilization. NOTE: responses that simply mention “knowledge” or “understanding” do not receive this code | Traditional: “For me, it was rotating molecules around in my head and understanding how each reaction condition affects the products.” |
OCLUE: “Understanding how each reagent indicates different mechanisms between structures.” | ||
Specific topic | Student responses that include one or more of the following dimensions regarding their perceptions of what is the most difficult thing about organic chemistry: (1) listing off specific topics, particularly with no reference to understanding or approaches utilized. Most common specific topics mentioned include mechanisms, acid–base reactions, learning objectives, naming, synthesis, and spectroscopy; (2) explicit mention of “concepts” without expanding on what they mean (i.e., “understanding concepts” or “knowing a mechanism”) | Traditional: “the synthesis problems” |
OCLUE: “The most difficult thing is CNMR and HNMR so if you can learn that you can learn anything” | ||
Memorization and/or workload | Student responses that include one or more of the following dimensions in regard to their perceptions of what is the most difficult thing about organic chemistry: (1) memorization, remembering, recalling, or regurgitation of reactions, products, reagents, and/or mechanisms; (2) knowing reactions, products, reagents, and/or mechanisms, particularly with no explicit mention of understanding the “what”, “how”, or “why” a reaction proceeds; (3) mention of the large amount/volume of material, the large amount of studying, and/or the amount of time the course requires; (4) explicit mention of feeling overwhelmed with the course; (5) mention of difficulty with keeping up with the class | Traditional: “The memorization of content. Its a lot of information.” |
OCLUE: “The most difficult thing about organic chemistry is how many mechanism you have to know. It can get a bit overwhelming, but if you try to practice once a day, and keep up with your notes then it won't be as bad.” | ||
Personal, course, and/or exam aspects | Student responses that include one or more of the following dimensions regarding their perceptions of what is the most difficult thing about organic chemistry: (1) mention of personal action and/or behaviors that a student must have such as “staying motivated” or “staying on top of the material” or “being patient”; (2) mention of how a student must regulate actions and behaviors to complete the course; (3) when a student discusses aspects of the course or exams such as overall difficulty or time allotted to an exam | Traditional: “the most difficult part is holding yourself accountable to continue studying throughout the semester” |
OCLUE: “The most difficult part is the self discipline that is required in order to make sure you learn everything that is being offered to you in this course.” | ||
Not applicable | Student responses that include one or more of the following dimensions in regard to their perceptions of what is the most difficult thing about organic chemistry: (1) the response does not answer the question; (2) when a student provides no response at all; (3) when the response is unclear or interpretation of the response is difficult, such as when students say “understanding the material”; (4) when the response falls into no other category; (5) when a student is venting about the course, professor, or other aspects relevant to the course | Traditional: “nothing” |
OCLUE: “Literally all of it” |
The “Apply and Reason”, “Identify and Describe”, and “Memorization” categories align to previous explanations of the same categories in question 1 (expectations of thinking). The only category identified that was unique to the responses to this question was the “Personal, Course, and/or Exam Aspects” category. Responses for this category typically reported personal actions or behaviors such as “staying motivated”, “staying on top of the material”, or “being patient” as well as referring to facets of the course (i.e., the professor and grading schemes) and exams (i.e., format). These types of responses received their own category due to their prevalence in the data, unlike in question 1 (expectations of thinking) where there were so few of these types of responses that they were assigned to the “Not Applicable” category. In question 1 (expectations of thinking), “Memorization” had an entire category of its own; however, for question 2 (most difficult thing) many students, particularly in the traditional course, coupled their perception of memorization with a large workload that was “overwhelming” or included a “high speed of coverage”. Initially for this question, there were separate “Memorization” and “Workload” categories, but since it was difficult to determine whether to classify these responses separately as “Memorization” or “Workload” we decided to combine the codes together as it still allowed us to make a broad comparison of the two courses.
However, to further explain our rationale for the combination of “Memorization” and “Workload”, 355 out of 604 responses in the traditional course discussed “Memorization”, “Workload”, or both. Over 25% of the traditional students mentioned “Memorization” and “Workload” simultaneously. Given this sizable chunk of data in one of the courses mentioned both together, we opted to combine them especially given that the narrative we were interpreting did not change based on combining the two categories and it allowed for noting broad themes and patterns across all 852 responses.
Two other categories, namely “Specific Topic” and “Not Applicable” were also identified. The “Specific Topic” category included responses which listed discrete specific topics that students found difficult. Throughout these responses, students did not make any reference to ways of thinking used to interpret the content associated with these topics. The common topics mentioned by students included “acid–base reactions”, “naming”, “synthesis”, and “spectroscopy”. In contrast to question 1 (expectations of thinking), the “Not Applicable” category for question 2 (most difficult thing) did not include references to the course or instructor as they were coded separately, but it did include instances when students did not give any response at all or when the response was unclear or unrelated to the question posed.
The analysis of student responses from the OCLUE and traditional courses again revealed differences in how students perceived the difficulty of learning organic chemistry. As shown in Fig. 2, more OCLUE students perceived that more sophisticated ways of thinking such as “Apply and Reason” and “Identify and Describe” were the most difficult thing about learning organic chemistry compared to students in the traditional course. For example, 16.5% (n = 41) and 16.1% (n = 40) of OCLUE students perceived that the most difficult aspects of learning organic chemistry were applying and reasoning and identifying and describing, respectively. This compares to the 2.6% (n = 16) and 5.1% (n = 31) for students in the traditional course for those same categories. More students from the traditional course listed specific topics (18.4%, n = 111) and the memorization and/or workload aspect (58.8%, n = 355) as the most difficult part of learning compared to OCLUE students (17.7%, n = 44, and 17.7%, n = 44, respectively).
Code | Dimensions | Example quotes |
---|---|---|
Apply and reason | Student responses that include one or more of the following dimensions regarding their perceptions of what is assessed: (1) understanding “why” a reaction proceeds; (2) the use of knowledge, specifically with the use of fundamental or basic ideas; (3) the transfer of knowledge to new problems; (4) making connections between concepts or piecing concepts together, especially to apply them; (5) making predictions to solve a problem | Traditional: “you don't just memorize; you understand why they are made like that so you can apply it to other reactions” |
OCLUE: “We were expected to know WHY things were happening, not just what was going on but the driving force behind those reactions” | ||
Identify and describe | Student responses that include one or more of the following dimensions in regard to their perceptions of what is assessed: (1) understanding the “what” and “how” reactions proceed, particularly without mentioning the use of knowledge to understanding “why” reactions proceed; (2) mentions understanding at the scalar or one scalar below levels, particularly through the recognition of concepts such as polarity and electronegativity and their significance to understanding; (3) when a student explicitly mentions understanding the mechanism instead of memorization; (4) when a student mentions “differentiating” between reactions with any further explanation. (5) responses include a discussion of forces, charges, or stabilization. NOTE: responses that simply mention “knowledge” or “understanding” do not receive this code | Traditional: “You need to know how to classify and name molecules, know characteristics like acidity and aromaticity, and mostly know how bonds are formed and broken in different situations using different molecules.” |
OCLUE: “You are required to think about reactions more about how electrons are moved in a system rather than what the begining and end products are. you have to know the steps of how to get there.” | ||
Specific topic | Student responses that include one or more of the following dimensions regarding their perceptions of what is assessed: (1) listing off specific topics, particularly with no reference to understanding or approaches utilized. Most common specific topics mentioned include mechanisms, acid–base reactions, learning objectives, naming, synthesis, and spectroscopy; (2) explicit mention of “concepts” without expanding on what they mean (i.e., “understanding concepts” or “knowing a mechanism”) | Traditional: “there is naming, mechanism, nmr, lots of reactions, and some bonus questions” |
OCLUE: “Different types of reactions and the classifications of structures.” | ||
Memorization | Student responses that include one or more of the following dimensions regarding their perceptions of what is assessed: (1) memorization, remembering, recalling, or regurgitation of reactions, products, reagents, and/or mechanisms; (2) knowing reactions, products, reagents, and/or mechanisms, particularly with no explicit mention of understanding the “what”, “how”, or “why” a reaction proceeds | Traditional: “The majority of the exams are memorization of the reactions.” |
OCLUE: “Mechanisms and if you can memorize 20 different types of problems with the same molecule everytime.” | ||
Exam format and aspects | Student responses that include one or more of the following dimensions regarding their perceptions of what is assessed: (1) the course materials leveraged on the exam such as lecture notes, homework, practice exams, and/or recitation materials; (2) the format of the exam, such as stating the types of questions on the exam (i.e., multiple-choice, or short answer); (3) the length, time, or fairness of the exam | Traditional: “You need to go to lecture and take notes, because the exams cover pretty closely what we cover in lecture.” |
OCLUE: “your ability to do them as fast as possible since the exam were only 50 minutes and crammed with material” | ||
Not applicable | Student responses that include one or more of the following dimensions in regard to their perceptions of what is assessed: (1) the response does not answer the question; (2) when a student provides no response at all; (3) when the response is unclear or interpretation of the response is difficult, such as when students say “your understanding of the material”; (4) when the response falls into no other category; (5) when the response focuses on student actions (i.e., “be sure to study hard”); (6) when a student is venting about the course, professor, or other aspects relevant to the course and the response focuses on the course or the professor such as frustrations they have with the course or professor | Traditional: “Everything” |
OCLUE: “don’t take 3 other intens classes with it” |
Codes such as “Apply and Reason”, “Identify and Describe”, and “Memorization” were explained in previous questions. A further three categories, namely that of “Specific Topic”, “Exam Aspects”, and “Not Applicable” were also identified and included in the analysis. The “Specific Topic” category is similar to the category in question 2 (most difficult thing) and included responses where students listed discrete specific topics as what gets assessed in organic chemistry. Once again, throughout these responses, students did not make any reference to ways of thinking used to interpret the content associated with these topics. The common topics mentioned by students here included “synthesis reactions”, “naming”, “NMR”, and “spectroscopy”. The “Exam Aspects” category included responses that referred to the format, length, time, and/or fairness of the exams in response to what gets assessed. These responses were noted in question 2 (most difficult thing), but were subsumed into the “Personal, Course, and/or Exam Aspects” category. Responses which noted perceptions of what course materials are typically assessed (such as lecture notes, homework, practice exams, and/or recitation materials) were also included in the “Exam Aspects” category. The “Not Applicable” category included instances when students did not give any response at all, or that their response was entirely unclear and unrelated to the question posed.
The analysis of student responses from the OCLUE and traditional courses to this question revealed differences in how students perceived what gets assessed in organic chemistry. As shown in Fig. 3, more OCLUE students perceived that more sophisticated ways of thinking were assessed in their course compared to students in the traditional course. For example, in relation to “Apply and Reason” and “Identify and Describe”, 35.5% (n = 88) and 12.5% (n = 31) of OCLUE students perceived these modes of thinking were assessed, respectively. This compared with 4.6% (n = 28) and 6.5% (n = 39), respectively, for students in the traditional course. More students from the traditional course listed specific topics (52.2%, n = 315) and memorizing information (11.8%, n = 71) as how they perceived they were assessed compared to OCLUE students (12.9%, n = 32, and 2.0%, n = 5, respectively). However, more OCLUE students noted responses coded to other categories such as “Exam Aspects” (27%, n = 67) and “Not Applicable” (10.1%, n = 25) than students from the traditional course (18.7%, n = 113, and 6.3%, n = 38, respectively).
For question 2 (most difficult thing) there was an even more marked difference between the two cohorts as noted Fig. 5. OCLUE students were evenly split on what aspects of the course they perceived as more difficult, whereas almost 80% of the traditional students believed that the focus on “Rote Knowledge” was the most difficult. Here, we recall back to Fig. 2 where it can be noted that more students in the traditional course perceived that the memorization, workload, or the workload involved in memorizing a large amount of material was what made the course difficult. For example, one OCLUE student's response categorized as “Use of Knowledge” stated the following: “The most difficult thing in orgo [sic] is the mechanism and understanding where and how different molecules attack each other. If you know them well then it makes writing reactions easier”. This student highlights that when you understand the behaviors of different molecules, then this can make writing reactions more approachable. On the other hand, an example from the traditional course categorized as “Rote Knowledge” noted that: “There is a large amount of material that we have to know and memorize”. In this case, the student is not only perceiving a large workload, but they also perceive they are expected to memorize the material.
The differences in perceptions was continued in responses to question 3 (assessment; Fig. 6), where once again a majority (48%) of OCLUE students perceived an emphasis on the use of knowledge in course assessments, whereas 64% of traditional students perceived that they were being assessed on memorization and rote knowledge. For example, one OCLUE student's perspective on this question was the following: “Your knowledge not only of what is taught in class but your ability to apply it to various situations. Also, you [sic] knowledge of the CONCEPTS [sic] and underlying themes is heavily assessed.” Here, the student describes how OCLUE assesses the student's ability to transfer concepts from one problem to another and to identify underlying themes which correlated with “Use of Knowledge”. On the other hand, responses correlated with “Rote Knowledge” included: “how well you can memorize the reactions” and “the exams mainly test reactions and naming of molecules”. Here the responses in the traditional course cluster around memorization and the focus on discrete, specific topics, such as knowing reactions and nomenclature, rather than ways of thinking.
To further determine if the differences between OCLUE and traditional cohorts in the qualitative data was supported statistically, we conducted a Pearson's chi-square test of independence using an alpha of 0.05 within SPSS 27 (SPSS, 2020) with the data organized into superordinate themes. According to the chi-square tests, the analysis for each question yielded statistically significant results at an alpha of 0.05 where p < 0.001 for each question. Since all Pearson chi-square tests came back significant, we decided to run post-hoc analyses to further illustrate which theme(s) were primary drivers for statistical significance in the initial chi-square tests. From the post-hoc analyses we found that the “Use of Knowledge” and “Rote Knowledge” themes were strong primary drivers for significance in each question. All of the calculations and a more in-depth write up of these analyses can be found in the supplemental materials (Tables S1 and S2, ESI).
Across all three questions, more students in OCLUE had perceptions aligned with the use of knowledge while student perceptions in the traditional course aligned more with rote knowledge. As we noted these differences in student perceptions, our interpretations and discussions often centered on the classroom cultures of learning in each organic chemistry course. Certainly, learning is a social and cultural activity that is dependent on the context in which it occurs (Vygotsky, 1978; Rogoff, 1990; Calabrese Barton et al., 2008; Carlone et al., 2011). Therefore, our discussion and interpretation of our findings can be situated within sociocultural perspectives (Vygotsky, 1978; Rogoff, 1990; John-Steiner and Mahn, 1996; Carlone et al., 2011; Zotos et al., 2020; Petterson et al., 2022) and is informed by other scholars who have conceptualized culture (Schein and Schein, 2016; Reinholz and Apkarian, 2018). Since the term “culture” can take on a variety of meanings, we find it important to provide a working definition prior to discussing our findings.
To begin, it is important to note that “no one view of culture… represents a thorough and complete understanding” (Parsons and Carlone, 2013). However, throughout this discussion when we refer to culture, we are referring to a micro-level culture, or a subculture, that exists in the context of these organic chemistry classrooms, as opposed to macro-level cultures which represent larger entities such as ethnic groups, nations, and international organizations (Schein and Schein, 2016; Thoman et al., 2017). Aside from sociocultural perspectives, our view of culture draws heavily on Reinholz and Apkarian's four frames for systemic change (Reinholz and Apkarian, 2018) and Schein and Schein's framework for organizational culture (Schein and Schein, 2016). Reinholz and Apkarian's four frames include structures, symbols, people, and power which exhibits overlap with Schein and Schein's framework of artifacts, espoused beliefs and values, and taken for granted assumptions, both of which inform our working definition and are further explained in our working definition below.
For us, our working definition of culture includes a constellation of visible structures and artifacts which encompass the visible course policies, course practices, expectations, and assessments, among other features. These structures are given meaning by an underlying system of symbols that include beliefs, values, and assumptions. Socializing mechanisms in a context enculturate people by encouraging them to adopt the symbols and participate in or interact with the structures and artifacts. These socializing mechanisms are mediated by people and power that directly and indirectly impact how people talk, act, and think (Rogoff, 1990; Miller and Goodnow, 1995; Lemke, 2001; Gutiérrez and Rogoff, 2003; Nasir and Hand, 2006; Calabrese Barton et al., 2008; Schein and Schein, 2016; Reinholz and Apkarian, 2018; Deng et al., 2021). While our findings cannot speak to all frames (structures/artifacts, symbols, people, and power), this definition helps us suggest potential explanations for our findings.
For this study, we found the cultural frames of structures/artifacts and symbols most useful particularly because most student responses were related to these frames given the questions asked. Structures, or artifacts, within a classroom could be elements such as the practices used, the learning and assessment tasks, and the established norms. That is, they are the visible features of the culture that are informed by the underlying symbols that give them meaning. The symbols could include the implicit and explicit messages that students receive and interpret that communicate valued ways of knowing and doing (Schein and Schein, 2016; Reinholz and Apkarian, 2018). The other frames mentioned by Reinholz and Apkarian, such as people and power are important, but were difficult to address with this data. Therefore, we aimed to use the frames of structures/artifacts and symbols to discuss how students perceived they were expected to practice learning and what was valued, both of which will be linked to elements of their respective cultures of learning. Other studies have used sociocultural perspectives to explore different classroom cultures and found that when the use of certain practices, such as argumentation, align with the course goals, then students engage more productively in the practice (Sandoval et al., 2019) while others have demonstrated how classroom norms (and their interpretation) can impact how students respond to learning tasks (Becker et al., 2013; Chang and Song, 2016). Thus, this suggests that better alignment between course goals, the practices students engage in, and clear and universally understood norms can lead to a more productive learning experience. Therefore, by investigating student perceptions of what is valued through the lens of classroom cultures we can help identify potential mismatches between instructor expectations and what students are doing that may perturb learning.
Constructing explanations and engaging in argumentation are important classroom practices in OCLUE (National Research Council, 2012a; Cooper et al., 2019; 3DL4US, n.d.; Flaherty, 2020b). Their incorporation coupled with the expectation students will engage in them act as structural features of the overarching culture. It has been suggested that by implementing the practice of constructing explanations that students will have a better idea of how scientific knowledge is developed (McNeill et al., 2017) and that argumentation can move the focus of learning away from memorization (Berland and McNeill, 2010). Certainly, this data corroborates this claim. Structural features of the classroom culture, such as the incorporation and consistent use of scientific practices, may have helped students in OCLUE perceive expectations of how to think on a deeper level relative to students in the traditional course.
For question 2 (most difficult thing), an overwhelming majority of students in the traditional course perceived that “Rote Knowledge” (such as memorization, the workload, or the workload associated with memorizing) was the most difficult part of the course (as noted in Fig. 5). In contrast, OCLUE students had a more even distribution of perceptions of which facets were most difficult, though more OCLUE students perceived that the “Use of Knowledge” was the most difficult aspect when compared to the traditional course. The OCLUE curriculum was designed in such a way to discourage rote memorization of content (Cooper et al., 2019), and far fewer students in OCLUE perceived memorization and workload as being the most difficult aspect of the course when compared to students in the traditional course. This highlights that student perceptions in OCLUE exhibit alignment with the transformational intent and implies that memorization and workload are stronger driving influences or forces within the culture of the traditional course.
Students have perceived that organic chemistry requires a great deal of memorization (Moran, 2013) which has also been noted as an approach that students take on organic chemistry exams (Webber and Flynn, 2018). Furthermore, instruments focused on gathering student perceptions have sought to collect information on whether students are memorizing in their courses (Grove and Bretz, 2007), yet, considering the previous exploration of the association of memorization with organic chemistry, a course centered on rote knowledge is almost certainly not the intent of the instructors. In a qualitative study on student reasoning in organic chemistry Anderson and Bodner (Anderson and Bodner, 2008) found that students did not appreciate that mechanisms were used to understand how and why phenomena occur despite the fact that this was the intent of the instructor in that course. In the same study, students in their interviews stated they wanted to understand the material on a deeper level but also mentioned that this was difficult given the volume and pace of the material (Anderson and Bodner, 2008), a perception we noted in our study for the students in the traditional course. That is, structural components of the traditional classroom culture, such as the amount of material covered, and pace of coverage may coalesce with perceived expectations to drive the perception that students need to memorize large amounts of material.
If instructors want students be able to explain how and why chemical reactions happen, the findings from Anderson and Bodner and our study make it clear that the purpose of mechanisms needs to be made explicit and leveraged consistently throughout the course and that courses need to slow down and connect content back to fundamental principles so that students can develop a robust understanding which may not have been clear in the traditional course. Both points are addressed in OCLUE by leveraging the scientific practices, crosscutting concepts, and core ideas and is further evidenced by the shift in perceived difficulties of students in the course toward use of knowledge, relative to students in the traditional course.
Finally, the responses for question 3 (assessment) yielded similar patterns to question 1 where, in general, OCLUE students perceived they were assessed more on their “Use of Knowledge” while students in the traditional course perceived they were assessed more on “Rote Knowledge” such as memorization and discrete, specific topics (as can be seen in Fig. 6). It is important to reiterate that assessments play a large role in the culture of a learning environment and send strong messages about valued ways of thinking and participating in the course (Snyder, 1973; Crooks, 1988; Entwistle, 1991; Scouller and Prosser, 1994; Scouller, 1998; Momsen et al., 2013; Stowe et al., 2021). Within OCLUE, much work is done to ensure alignment between learning goals, expectations, and assessments with regard to the use of knowledge. Student perceptions imply that this transformational goal may be accomplished (at least partially) since 52% of students in OCLUE perceived they were expected to use their knowledge and 48% perceived they were assessed on their use of knowledge.
In terms of culture, assessments act as one mechanism through which instructors reflect what is valued in the learning culture and what students are expected to do. As shown in this study, these messages can be perceived by students and influence how they participate in learning. If the goal of the learning environment is to engage students in reasoning and disciplinary practices, then the culture and instructor expectations must support that goal (Bain et al., 2020), as Cooper and Stowe note: “…it is important for students to receive and respond to the message that both knowledge and the ways that knowledge is used are crucial aspects of learning chemistry,” (Cooper and Stowe, 2018). Instructor expectations and assessments are intricately linked, and the ways in which courses communicate expectations, emphasize particular ways of doing, and place value on those ways of doing (by assessing them) become structures and symbols of the learning culture. The alignment of expectations and assessment in OCLUE, as noted in student perceptions, was an important component of the transformation effort to ensure that what was expected of students was valued in the form of points on assessments.
When considering the culture of a classroom, it becomes important to also consider the ways that culture socializes people. Instructional practices that focus on rote memorization and solving exercises will likely not introduce students to the authentic disciplinary culture nor encourage them to engage in “science-as-practice” (Nasir and Hand, 2006; Stroupe, 2014). Instead, if students are immersed in an environment where they are encouraged to use their knowledge, particularly with unfamiliar problems, and given the chance to make mistakes and learn from them, then students may develop perceptions of learning which are more aligned with authentic disciplinary ways of thinking (Brown et al., 1989). Furthermore, if a class culture's goals align with the practices students are expected to engage in, then it can lead to more productive engagement and learning (Sandoval et al., 2019). By leveraging scientific practices in the context of fundamental core ideas, instructors can shift the culture of learning to expect, emphasize, and value the use of knowledge and provide students a route to connect their knowledge and make sense of a phenomena rather than relying on memorization (Cooper, 2015).
One of the goals of our study was to explore the alignment between student perceptions and instructor expectations. Previous research has found that organic chemistry instructors do not list rote memorization as an important facet of learning organic chemistry (Duis, 2011); yet, students in the traditional organic chemistry in this study largely perceived they were expected to and assessed on their ability to memorize. We imagine that the goal of the instructor was not to have students rely solely on rote memorization. Therefore, there seems to be a disconnect and misalignment between what the instructor values and expects students to do and their learning and assessment task design (Stowe and Cooper, 2017). That is, though instructors may expect and value students to use their knowledge, students are still able to complete prompts and learning tasks by memorizing the material. In a recent interview study on student perceptions of “critical thinking” in organic chemistry courses, some students mentioned they saw memorization as an “easier” method to achieve the results they wanted (a better grade) and that they were accustomed to memorizing in school (Bowen and Cooper, manuscript in preparation). Thus, this highlights that more attention should be given to the questions and prompts being asked of students and that learning and assessment task design be intentional and reflective.
While there are a variety of ways to engage students in meaningful learning, pedagogical approaches such as those informed by A Framework for K-12 Science Education and three-dimensional learning have advocated for engaging students in authentic disciplinary practices in science (National Research Council, 2012a; Laverty et al., 2016; Matz et al., 2018; 3DL4US, n.d.). From these perspectives, learning involves introducing students to the disciplinary cultures of science by engaging them in the practices that scientists actually use, such as constructing explanations and using models to predict and explain. Our previous work on OCLUE, along with the findings here, demonstrate how three-dimensional learning can impact student performance and communicate clear expectations and values that are explicit for students and align with more expert-like practice.
Certainly, there are many factors at work in a learning culture, and our study did not, and could not, address them all. However, what we have highlighted is how instructor expectations, whether implicit or explicit, along with what is emphasized in a course and on assessments are related to the overarching classroom culture and how these features send strong messages about how people should think and practice. While our previous work on student reasoning was certainly insightful, we needed evidence to better understand if students perceived what they were doing aligned with the goals of the course. Put simply, the enactment of a course, the elements of its culture, such as instructor expectations, emphasis, and valued of ways of doing influence how students participate in learning and more attention should therefore be given to these influences when designing and enacting instructional practice.
A third limitation is that our analytic approach relied on interpretation of written student perceptions. With any qualitative study, we must acknowledge that our interpretations are our own and contextual. Through the use of multiple coders, a unified codebook, and multiple cycles of coding and revisions, we aimed to characterize student perceptions the best way we could through noting broad and communicable themes across these two different organic chemistry course experiences. Finally, it could be that these perceptions may not be stable over time. Like other affective constructs, perceptions are subject to social and cultural influences and therefore may change throughout the semester; however, we do believe perceptions offer a snapshot into the student experience and worth considering in our transformation efforts.
With regard to research implications, the use of our instrument occupies a methodological middle ground in that it is less constrained than previously published quantitative instruments, and it does not take as much time as conducting interviews, yet the instrument still provides rich descriptions of student perceptions. By studying perceptions in this way, we can provide valuable insight into how students are engaging with curricula and learning environments without relying on our assumptions, and more qualitative studies on student perceptions of learning and affective states would help expand the CER literature base.
With all of this said, various cultural frameworks informed our interpretation and communication of the results. We posit that more work should be done within the realm of classroom cultures in chemistry courses. Culture influences how students talk, think, and act, and research focused on characterizing the different classroom cultures that support learning and foster student engagement would be productive and insightful.
Footnote |
† Electronic supplementary information (ESI) available. See DOI: 10.1039/d2rp00010e |
This journal is © The Royal Society of Chemistry 2022 |