Eshani N.
Lee
*a and
MaryKay
Orgill
b
aPennsylvania State University, Hazleton, PA, USA. E-mail: EGL51@psu.edu
bUniversity of Nevada, Las Vegas, Department of Chemistry & Biochemistry, Las Vegas, Nevada, USA
First published on 9th August 2024
Multilingual learners face significant challenges when navigating the linguistic complexities of chemistry assessments. This study, employing the Equitable Framework for Classroom Assessment, identified these specific challenging features in general chemistry assessment items on the topics of limiting reactant and percent yield. Through in-depth, semi-structured interviews with multilingual students, we discovered critical barriers to comprehension: lack of metacognitive support, complex vocabulary and syntax, dense text layout, and extraneous information. These findings emphasize the need to better understand and alleviate these types of linguistic features in assessment items to more accurately measure chemistry knowledge, rather than linguistic proficiency. By addressing these challenges, instructors can design more accessible assessment items for a diverse group of students. The results also offer valuable insights and practical guidance for writing equitable assessment items.
Recently, there has been a growing emphasis on STEM education because of its pivotal role in driving innovation, problem-solving, and global competitiveness (Honey et al., 2020). As such, broadening access to STEM for a diverse range of students, including those who are linguistically diverse, is crucial for ensuring equitable learning opportunities. Given that general chemistry acts as a foundational course for aspiring STEM students, understanding the diverse learning experiences within its context is essential. While efforts have emerged within chemistry education to foster inclusivity and equity, prompted by rising awareness of systematic barriers and discrimination (Wilson-Kennedy et al., 2022), there's a pressing need for deeper exploration of language's role in enhancing accessibility. Multilingual learners often confront language barriers that place them at a disadvantage as they navigate the dual challenges of mastering both English and the discipline-specific language of chemistry, particularly in the context of rigorous, high-stakes assessments (Lee and Fradd, 1998).
Research focusing on how multilingual learners engage with STEM content has seen a recent surge. However, exploration into assessing multilingual learners remains notably limited, primarily situated within K-12 education contexts (NASEM, 2018; Pun et al., 2023). This study is part of a broader project investigating the perspectives of multilingual students about (1) typical assessment items used in their general chemistry courses and (2) modified assessment items designed to reduce linguistic complexity while maintaining academic rigor. Previously, we discussed features of assessment items identified as supportive by multilingual learners (Lee and Orgill, 2021). Here, we report on features of the assessment items that the multilingual learners found particularly challenging.
Despite their being a significant—and growing—proportion of learners in the U.S., multilingual learners are underrepresented in STEM fields of study and careers (NASEM, 2018). Increasing multilingual learner students’ access to STEM content, fields of study, and jobs is crucial for multiple reasons. First, STEM knowledge is essential for understanding and participating in the world around us, as well as for addressing the global challenges we face as humans. According to the U.S. National Science Foundation publication STEM Education for the Future: A Visioning Report, “Whether or not they become scientists or engineers, all Americans should have the access, opportunity, encouragement, and tools to participate in the innovation economy, and to succeed amid technological progress and change” (Honey et al. 2020, p. 8). Thus, it is essential that all have access to quality STEM education, including multilingual learners. Second, the requirement for STEM workers has grown significantly, outpacing all other job sectors (NASEM, 2018; Honey et al., 2020). Multilingual learners can be a valuable resource for jobs in STEM-related fields. Third, because of their cultural and linguistic resources, multilingual learners can bring unique insights, approaches, perspectives, and foci to STEM research and innovation that can address the needs of their communities (Honey et al., 2020) and benefit society and research as a whole (NASEM, 2018).
“STEM subjects offer the potential for membership in the communities of mathematicians, scientists, engineers, and other technical experts—communities with their own ways of conceptualizing, representing, evaluating, and engaging with the world. In turn, STEM students from a wide range of backgrounds bring the potential to contribute to shaping STEM fields in critical ways that transform and remake focal topics, practices, and contributions (NASEM, 2018, p. 56).”
Finally, and most importantly, because of their economic value to society, STEM-related jobs generally have a high earning potential. By not increasing multilingual students’ access to STEM-related content, fields of study, and jobs, we reproduce and perpetuate social inequities (Fine and Furtak, 2020a). “Opening avenues to success in STEM for the nation's [multilingual learners] offers a path to improved earning potential, income security, and economic opportunity for these students and their families” (NASEM, 2018, p. 9).
Another shift has occurred in the way that researchers view language acquisition itself. Previously, researchers and scholars identified two types of language proficiencies that multilingual learners needed to negotiate in order to succeed in STEM classes: basic interpersonal communication skills and cognitive academic language (Cummins, 2000). Cognitive academic language was seen as much more difficult, requiring seven to ten years for a typical multilingual learner to develop (Collier, 1995). As a consequence of this dichotomous view of language acquisition, instructors might have seen multilingual learners as lacking in STEM content knowledge if they did not present their understanding in the expected form of cognitive academic language (NASEM, 2018; Garcia et al., 2021; Deng and Flynn, 2023; Grapin et al., 2023). Now, however, researchers recognize that there is not a dichotomy of language proficiencies that multilingual learners develop, but, rather, that students develop different language proficiencies in multiple areas (e.g., reading, writing, speaking, listening, etc.) along a continuum of registers (NASEM, 2018; Buxton and Lee, 2023).
In the context of language, registers refers to “the variation in language choices that people make in engaging in a range of activities throughout the day” (NASEM, 2018. p. 12). The language register that one uses depends on the content being communicated, the people with whom one is communicating, and the modality of communication. We all use multiple registers each day. For example, the type of language and skills (the register) one uses to text a friend is very different than the register that the same person would use to write a letter to a politician or to interpret poetry. A typical student in a STEM course will negotiate multiple registers during a single class period as they interact with peers to discuss weekend plans, work in a group to solve a content-related question, listen to an instructor's lecture, and ask a question of a peer instructor.
Importantly, multilingual learners often encounter hurdles on high stakes STEM exams because of these linguistic features, which can negatively impact their ability to showcase their true understanding of the content (Buono and Jang, 2021). In other words, challenging linguistic features can become a systematic barrier when they cognitively interfere with students’ abilities to demonstrate content knowledge in a competitive, fast-paced, and mandatory science course. Even English-as-the-first-language students (EngL1) struggle with understanding words in the chemical context of technical terminology and discourse structures (Rees et al., 2019). For example, words such as volatile and state are well understood in everyday life, but poorly understood in the context of chemistry (Cassels and Johnstone, 1983). However, to decode these types of terms in the correct context, multilingual students must also navigate multiple language domains effectively and quickly. Experiencing these types of challenges can create substantial obstacles, especially during “high stakes” course assessments, which can have long-term consequences on grades and academic persistence.
A large body of research has identified language as a source of “construct irrelevant variance—variation in test scores due to factors that do not have to do with the knowledge or skill being assessed. Many of these factors have to do with linguistic complexity; for example, complexity due to the use of unfamiliar and morphologically complex words, words with multiple meanings, idiomatic usages, and long or syntactically complex sentences in texts and accompanying test items and directions” (NASEM, 2018, pp. 213–214).
It is essential that STEM assessments be presented and carried out in as equitable a manner as possible because high-stakes assessments often determine who is allowed to learn, do, and be STEM (Fine and Furtak, 2020a). High-stakes STEM assessments carry not only educational consequences, but social consequences, such as determining who has access to STEM-related employment, the disproportionate economic value of STEM-related jobs, and the material goods and positive reputation that can result from employment in a STEM-related field (NASEM, 2018; Fine and Furtak, 2020a; Honey et al., 2020).
Initial attempts at making assessments more equitable for multilingual learners included giving students access to certain accommodations during an assessment. For example, multilingual learners might have been provided with extra time to complete the assessment or a dictionary to translate words written in English into terms in the learners’ native language. Overall, these accommodations have not been shown to be specifically helpful to multilingual learners (NASEM, 2018; Cardozo-Gaibisso et al., 2020). In fact, some accommodations, such as extra test time, benefitted both multilingual and monolingual learners and, thus, did not reduce the performance gap between these learner groups.
More recent efforts at making assessments equitable suggest that multilingual learners’ STEM content knowledge is better demonstrated through alternate assessments that are multimodal, open-ended and dynamic, as these allow multilingual learners to demonstrate their understandings in a variety of ways (Lopez et al., 2017; NASEM, 2018; Buxton et al., 2019; Fine and Furtak, 2020a; MacDonald et al., 2020; Buxton and Lee, 2023; Grapin, 2023; Grapin et al., 2023). Other scholars recommend that multilingual learners be allowed to use trans-languaging practices during assessments (Lopez et al., 2017; MacDonald et al., 2020; Pierson and Grapin, 2021; Pun and Tai, 2021; Grapin et al., 2023). These scholars assert that multilingual learners do not work or think in one language and then switch to working and thinking in a different language. Instead, they have a set of integrated skills, knowledge, abilities, and resources that result from the synthesis of their multiple linguistic and cultural backgrounds. Trans-Languaging involves the multilingual learners’ use of any of their linguistic resources when communicating (Garcia et al., 2021). Assessments that encourage the use of trans-languaging by multilingual students might, for example, provide assessment items in multiple languages or include an explicit statement that learners are allowed to use drawings, words in different languages, etc., in their assessment responses. Trans-Languaging is a potentially transformative equity practice (Grapin et al., 2023), in that it changes what counts as evidence of STEM learning (i.e., it does not require that learners demonstrate their STEM content knowledge through written answers that use perfected discipline-specific academic language).
Other efforts at making STEM assessments more equitable for multilingual learners have focused on modifying assessment items to reduce their linguistic complexity (see, for example, Siegel, 2007; Siegel et al., 2008; Siegel et al., 2014; NASEM, 2018; Fine and Furtak, 2020a, b; Guzman-Orth et al., 2021). For example, the National Academies of Science, Engineering, and Mathematics (2018) and Guzman-Orth et al. (2021) recommend that instructors follow the universal design for assessment principles (Thompson et al., 2002). Among these are two elements that specifically address linguistic complexity: (1) ensuring that “the language in the items is accessible and promotes maximum readability and comprehensibility (i.e., minimize construct-irrelevant variance for nonreading items)” (Guzman-Orth et al., 2021, p. 7) and (2) ensuring that “presentation details such as font, style, spacing, contrast, and white space do not become sources of construct irrelevant variance” (Guzman-Orth et al., 2021, p. 7). As another example, Fine and Furtak (2020a, 2020b) developed the Science Assessment for Emergent Bilingual Learners (SAEBL) checklist, which includes five categories of features to consider when designing assessments for multilingual learners. One of these is the “integration of scaffolds.” Within this category, Fine and Furtak recommend using sentence starters, sentence frames that connect ideas, graphic organizers, visuals, shorter sentences, bulleted items, active voice, and definitions of key terms within sentences, among others when constructing assessment items to meet the needs of multilingual learners.
It is not surprising that reducing the linguistic complexity of assessment items and integrating scaffolds is included on each of these checklists for making assessments more equitable for multilingual learners. Linguistic complexity has been shown to contribute significantly to the performance gap between multilingual learners and monolingual students on assessments in STEM classes (Fang, 2006; NASEM, 2018; Buono and Jang, 2021; Neri and Retelsdorf, 2022). Interestingly, according to the NASEM report, “There is evidence that even changing one word or slightly rephrasing an expression in science items may make a difference on whether an item is or is not biased against linguistic minority students” (NASEM, 2018, p. 215).
While the need for more research on assessing multilingual learners has been established (Bergey et al., 2018; NASEM, 2018), particularly in the context of higher education, there's a crucial gap in understanding the specific challenges they encounter during these assessments. Current research often focuses on broader assessment practices for multilingual learners (Buxton et al., 2014). However, to effectively modify assessments for greater equity, we need a deeper understanding of the specific linguistic and cognitive challenges multilingual learners face on exams.
As discussed elsewhere (Lee and Orgill, 2021), the EFCA's primary goal is to modify assessments linguistically to more accurately measure multilingual learners’ content knowledge without compromising the inherent difficulty of the items. We operationalized the EFCA modifications outlined by Siegel et al. (2008) as described in Lee and Orgill (2021) (for a detailed explanation, see Gandhi-Lee, 2018).
However, current research does not establish which specific EFCA modifications are most useful for undergraduate chemistry assessments. To bridge this gap, we need to go beyond the existing framework and incorporate student voices directly. By asking multilingual learners themselves to identify the features they find challenging in typical assessments and those modified using the EFCA, we can gain crucial insights into which modifications are most effective in the particular context of tertiary chemistry courses.
• What features do multilingual students think make it difficult for them to comprehend general chemistry assessment items?
In total, we recruited 10 multilingual learner participants, all of whom were enrolled in a first-semester general chemistry course at a large university in the U.S. Southwest. The multilingual learner participants represented diverse backgrounds, including two Filipino students, two Thai students, two Hispanic students (from Colombia and Venezuela), one Asian–Indian student, one Chinese student, one Russian student, and one student from Guam. All participants identified as a Generation 1.5 student; however, systematic background information about their prior formal and/or informal exposure to the English language was not collected. To safeguard the anonymity of participants, we assigned pseudonyms to each participant.
The assessment items we selected for this study encompassed a combination of mathematical and chemical concepts, as well as phrases that demanded thoughtful reflection. In other words, these problems could not be solved by mere application of formulas. All the assessment items used in this study were sourced from general chemistry textbooks and the test banks of general chemistry instructors. In the process of item selection, we sought the expertise of experienced chemistry instructors and researchers in chemical education. Four items were selected and then modified according to the EFCA guidelines (Lee and Orgill, 2021), resulting in four original items and four revised items for use in the study. These items will be presented, along with their accompanying discussions, in the Results section.
The interview protocol comprised four sections. The first section involved obtaining informed consent and collecting demographic information through a series of questions. In the second section, participants were asked rapport-building questions to establish a comfortable and open dialogue. The third section presented both the original and modified versions of each assessment item in a randomized order. Participants were instructed to read, work through the problems on paper, and share their thoughts about the challenging aspects of the items that impeded them in the problem-solving process. Participants were also encouraged to jot down notes directly on the assessment items during the interview. The final part of the interview served to address any clarifying questions from participants and to implement member checking as a means to validate the findings (Creswell, 2007) and enhance the study's credibility.
Following the five phases of thematic analysis (Braun and Clarke, 2006), we transcribed (Phase 1) the audio-recorded interview data verbatim and embarked on an iterative coding process based on the transcript content. During this analysis, we closely examined the interview transcripts, searching for recurring themes related to students' perceptions of the challenging features of the assessment items. These features encompassed various elements such as visual representations and wording.
To generate initial codes (Phase 2), we coded the transcripts based on the challenging features that participants identified across each assessment item. These individual codes were then organized into different types, such as sentence structure and contextual cues, grouping similar categories together. Subsequently, we developed overarching themes (Phase 3) that encapsulated these categories, describing each group of challenging features. Additionally, the notes that participants made while working through each item were cross-referenced to validate the emergence of new categories and maintain the authenticity of the evolving themes. To ensure the reliability of our analysis, multiple coders were involved in the data analysis process (Phase 4). In instances where discrepancies arose in our interpretations, we engaged in discussions and reached a consensus on any necessary modifications to the category descriptions.
Based on our analysis, the following four themes emerged (Phase 5) that describe the challenging features of the assessment items as perceived by students:
1. Lack of metacognitive support
2. Use of complex vocabulary and syntax
3. Use of dense text layout
4. Inclusion of irrelevant information
Our team comprises two researchers with diverse professional and personal backgrounds. The first author, serving as an assistant professor and a general chemistry instructor, brings a unique perspective as a first-generation student and woman of color, who identifies as cisgender. Her background as a first-generation student, along with her proficiency in non-English languages, enables a level of connection with the study's participants. The second author is a professor and a general chemistry instructor who identifies as a cisgender woman. She is conversationally proficient in Spanish and spent a year and a half working with Spanish-speaking people. Her struggles to communicate during that time period give her some insight into the challenges her multilingual students face every day. With doctoral degrees (PhDs), both authors acknowledge the disparities within the academic system that may hinder equal opportunities for all individuals.
We emphasize that our distinct backgrounds, experiences, and expertise inevitably influence our interpretations of the data. Our commitment lies in maintaining transparency and reflexivity throughout our endeavors to ensure an impartial and meaningful exploration of the participants' experiences.
An inductive analysis of interview transcripts revealed the following four primary challenges, ranked by frequency of mention:
1. Lack of metacognitive support
2. Use of complex vocabulary and syntax
3. Use of dense text layout
4. Inclusion of irrelevant information
Notably, some of these challenges identified by multilingual learners directly align with the modifications advocated by the EFCA. For instance, the lack of metacognitive support corresponds to the EFCA's emphasis on providing clear task instructions and scaffolding strategies. Similarly, the difficulties with complex vocabulary and dense text layouts mirror the EFCA's recommendations for using simple language and structuring the text in a user-friendly manner. This alignment between the identified themes and the EFCA modifications suggests that adopting the framework has the potential to enhance the assessment experience for multilingual learners. In the discussion that follows, items are referred to by a number followed by either the letter “O” or the letter “R” to represent the original and revised versions of the items, respectively.
1. Disconnected wording: Assessment items lacked internal wording that typically guides readers in understanding the overall flow of information and the key concepts being tested. These internal words function as signals that help readers navigate the text and grasp connections between ideas and include logical connectives (“because,” “however”), sequence words (“first,” “next”), emphasis words (“indeed,” “significantly”), and summarizing words (“overall,” “in conclusion”).
2. Insufficient problem-solving guidance: This element emphasizes the absence of specific and explicit support structures that guide students towards successful problem-solving. Support structures can include breaking down the question into smaller steps, providing prompts to guide students towards the correct approach, or offering sentence starters to initiate equation formulation. The absence of such guidance in assessment items left multilingual learners without crucial support, making it difficult for them to grasp the connections between different parts of the item and identify the core objective of the question.
![]() | ||
Fig. 1 The original version of the assessment item 1 (Zumdahl and Zumdahl, 2012, p. 113). |
Lupe stated, “Well, for me, you won't really know at first where to start. They just give you a bunch of numbers. It takes a while to figure out these numbers from this number and from this molecule, and then this one's from this. It's just hard to organize it at first. You actually have to work on it first before you really work on it or else you'll forget the details.”
Item 2-O (Fig. 2) presented a similar challenge. Although the initial sentences described the experimental set up of the containers with chlorine gas, the lack of linguistic signals made it unclear how that information was connected to Parts A through E. Students like Carlos reported spending significant time re-reading the question because of the difficulty in applying the information provided. Carlos noted, “…it took a couple of times to read because they asked for a lot of things so I just took a couple of times to read through it.”
![]() | ||
Fig. 2 The original version of the assessment item 2 (Zumdahl and Zumdahl, 2012, p. 241). |
Naima stated that, “I think [we] would have a tough time trying to think about what the y-axis is, the numbers and values for the y-axis, and that would just throw us off [from] figuring out B, C, D parts, and eventually towards E.”
Students also highlighted the importance of clear connections within multi-part items. Item 2-O, with parts A through E, lacked clear transitions between these sections. This made it difficult for students to understand how each part built upon the previous one and ultimately solve the problem.
Rohan mentioned that “[item 2 original] is the hardest because it looks simple but the wording makes you think a lot, and people would get confused especially since you don't have any guidelines…confusing steps. There's no y-axis value and stuff.”
Rohan stated, “I noticed that a lot of the word choices here are elevated and then in general seems like it has a longer sentence structure. ‘Potential’ and ‘manufactured’ are words that are sort of especially because some of this information is not super needed especially when it's talking about being simplest alcohol and like potential replacement for gasoline.”
In another case of ambiguous wording, item 4-O (Fig. 3) employed the term “roasting” to the process of making sodium chromate. This choice of language led to confusion among students, who were unfamiliar with the term and/or questioned its relevance to the item's purpose.
Hector stated, “I don't understand how [roasting] would help me with the question. It's just a weird word…because when you think of roasting you think of like heat. Then, you had to think about if heat is lost or gained in the equation. So, it might confuse people, it's just not a great word to put in there.”
The inclusion of “roasting” in item 4-O introduced an unnecessary detail that did not contribute to the overall understanding of the problem. Instead, it served to distract and confuse students, detracting from their comprehension of the content.
For example, item 1-O (Fig. 1) contained sentence structures that began with “Suppose…” and “If…” Conditional structures may not exist or be as commonly used in the multilingual student's native language(s). Navigating abstract concepts within hypothetical scenarios compounded the students’ challenges in this case.
Anastasia expressed her uneasiness when encountering these types of sentences during exams or quizzes, “Because people are really unsure and scared going through this and you hear, ‘Suppose,’ and couldn’t it just be stated like, ‘Yes, this is how much was put in.’ […] I’m already nervous and scared and your ‘suppose’ makes me nervous and scared for the rest of it. That's just my own personal feeling, [especially] if my test anxiety is already shooting.”
These types of sentences found in items 4-O (Fig. 3), 1-O (Fig. 1) and 4-R (Fig. 4) often utilized lengthy, compound sentences featuring coordinating conjunctions (“and,” “with”) to connect multiple clauses and phrases. These complex structures, compounded by the use of unfamiliar terminology (“source of chromium,” “roasting chromium”) and chemistry formulas of sodium carbonate and sodium chromate, can present significant challenges for multilingual learners. Simplifying the language and using shorter sentences with smaller chunks of information can make these assessment items much more accessible.
This issue was also seen in item 4-R (Fig. 4) where students were confounded with processing the last statement containing dense information. The question part starting with “What is the percent yield if 1.2 kg of… 1.0 kg…?” was particularly challenging to follow for some students. Naima discussed how this type of sentence can lead to confusion,
“You put two values in the same question, but they are not supposed to be used in the same step, that gets confusing. Even if it was just 1.2 kilogram of sodium chromate was produced, period. What is the percent yield from that or contains – just separating those [values] even in the slightest bit. Because when you put in the same question, an average student wouldn’t probably know [how to start].”
Students articulated that grappling with multiple numerical values within a single convoluted sentence induced stress and confusion. Ina discussed, “I know most chemistry equations are written like this. It's just when there are two numbers back-to-back like this, I don’t know which one to use first on the conversion factor thing. Yes, it's just a bit confusing.”
This was most evident in how students responded to item 2-R (Fig. 5). Item 2-R (Fig. 5) was a modified version of item 2-O (Fig. 2), developed using the EFCA framework. These modifications aimed to improve the original item (Fig. 2) by adding visual representation (containers with Cl2 and a list of the masses in Na in each container) and targeted contextual support for parts A through E. However, these modifications resulted in a noticeable increase in item length compared to the original version.
Students perceived the item 2-R (Fig. 5) to be too lengthy and reacted negatively to its appearance. They expressed concerns about the increased reading and translation time required during exams or quizzes. Interestingly, despite their initial reservations, students found item 2-R (Fig. 5) to be easier to follow and solve. Elara exemplified this sentiment, stating, “Okay, this one looks more overwhelming but [it's] easier to solve than the other one (2-O).”
Visual presentation also impacted students’ perceptions of difficulty. Item 1-O (Fig. 1) lacked spacing or separation between questions, appearing as a single dense block of text. Students, like Carlos, found this challenging: “Because with this one it looks like a paragraph of information. I feel that's how some of my chem questions are like, and this is when it starts to get hard for me because when I have so much information to sift through, I don't know what I'm looking for.”
These findings suggest that clear spatial organization can make assessments more accessible for multilingual students. On the other hand, while the length of an item can be a deterrent for multilingual students, they ultimately appreciate well-structured items with clear instructions that aid comprehension and problem solving.
Seojun mentioned, “So I don't know if it's really needed to include this mess [pointing at the first two lines] that it is also called a methyl alcohol just because throughout the rest of the problem it's referred to as methanol. So it really doesn't matter what else you call it. Then I guess the same thing with this one with simplest alcohol it's not very relevant.”
Even though item 1-O was revised to include simpler terminology (item 1-R, see Fig. 6), students still carried the same sentiment toward the background information. The overwhelming consensus was that this information was unnecessary and should be eliminated.
Sometimes, students were unable to determine if background information was simply provided for context or if it was needed to solve a problem. For example, in item 3-O (Fig. 7), information about “Haber process” and “high pressure” was perceived as relevant but led students astray. Ina expressed uncertainty about whether the Haber process information guided her in setting up the problem.
![]() | ||
Fig. 7 The original version of item 3 (Brown et al., 2015, p. 238). |
Ina said, “I found really interesting where it describes the Haber process. I've never heard of this before but if people know or if the students know what this process is then would [they] know how to set up this problem?”
The inclusion of this additional information in the item derailed students’ thinking as “high pressure” made some consider using the ideal gas law equation because in PV = nRT, P is related to pressure and CO and H2 were said to be in gaseous state. Lupe discussed, “There's not really any areas that is helpful but the ‘high pressure’ in the first part, it throws you off […] For this one, you would think of whether you have to use the ideal gas law?”
Item 3-R (Fig. 8) was developed to improve readability of the original version 3-O (Fig. 7). While students found the first three sentences about making ammonia and the Haber process easier to read and comprehend, they still questioned the relevance of this information. Sheela expressed, “The Haber Process, I don't even know what is that…I didn't really need to know that.”
Similarly, in item 4-O (Fig. 3), students were inundated with an abundance of information, leading to prolonged reading and translating times. Despite initially seeming relevant (the background information mentioned chromite and the formation of sodium carbonate), students ultimately perceived the first sentence to be unnecessary for problem-solving. Many students invested time in deciphering and annotating this information in both their first language (L1) and English, only to later express frustration upon realizing its lack of utility. Anastasia discussed that this information was too much to process and questioned the use of it. “I think it's very wordy or dense, especially in the beginning. I read through and it said, ‘Use a source of chromium in chromium compounds,’ they really just put chromium and chromium back together, but how do I use this?” Even when the same information was presented using a shorter and simpler first sentence, as shown in Item 4-R (Fig. 4), students like Anastasia still felt this information was extraneous and should be eliminated.
In alignment with literature, this finding also suggests that although rich contextual information can enhance student engagement during assessment, it can also hinder performance if overly complex or irrelevant. The Handbook of Accessible Achievement Tests for All Students (Abedi, 2011, p. 219) emphasizes the need to carefully balance contextual details with the core assessment needs for multilingual learners “…test items presented in rich language context would be more interesting to test takers. However, sometimes the language use for context in the assessment questions may be excessively complex and may cause misunderstanding and confusion. Therefore, it is of paramount importance to distinguish between the language that is a natural part of the assessment and essential to the assessment process and the language that is unrelated to the assessment process.”
The most prevalent challenge was the lack of metacognitive support within items, which posed a massive entry barrier for multilingual students to linguistically access the item. Our previous study showed that multilingual students looked for extralinguistic features within an item to aid their comprehension (Lee and Orgill, 2021); these extralinguistic features are based in contextual cues and scaffolding. These results align with Jerome Bruner's theory in cognitive psychology, which emphasizes the importance of scaffolding for students in facilitating language acquisition, particularly when it coincides with learning new subject-specific language (Bruner, 1983). Scaffolding, as Bruner defines it (1983, p. 60) is a “a process of ‘setting up’ the situation to make the child's entry easy and successful…” Scaffolding, therefore, emerges as a powerful pedagogical tool for enriching the linguistic and extralinguistic contexts on assessment items. The absence of clear connections, emphasis markers, and step-by-step guidance forces multilingual learners to spend a disproportionate amount of time deciphering the meaning of the questions rather than focusing on the identifying the key information to solve the actual concepts being tested.
The findings also underscored that the use of complex vocabulary and syntax is perceived to be cumbersome and confusing. Elevated vocabulary words, not necessarily discipline-specific, may impede comprehension of the assessment items. Consequently, employing simpler alternatives (e.g., ‘made’ instead of ‘manufactured’) to convey the same meaning is advisable if the information is pertinent to setting up and solving the problem. It was also particularly challenging for students to read ambiguous wording in the main part of the question, leaving them more uncertain about how to respond. As instructors, we value our students’ ability to think critically and elicit the next steps to solve a problem; however, this finding shows that for multilingual learners, including ambiguous terms such as “explain” or “describe” seems to raise more confusion about what is being asked. Similarly, students were confused by terms that might have different meanings in everyday registers and chemistry registers. The use of the word “roasting” elicited interesting reactions from students as it made them think of heat or cooking; however, students did not know how to apply this term to the chemistry context. Overall, these findings align with research suggesting that complex language features can impede content comprehension for multilingual learners (Gibbons, 2006). Therefore, multilingual learners could benefit from clear and concise language, allowing them to focus on the core scientific concepts rather than deciphering complex sentence structures or unfamiliar/uncommon vocabulary.
The study also revealed a noteworthy finding among multilingual learners about item presentation: they often equated item length with difficulty. Students initially judged longer items (items 1-R and 2-R) as presenting greater difficulty in both comprehension and problem-solving. This was evident despite the improved visual and spatial organization within the modified item 1-R (Fig. 6). Similarly, the seemingly shorter original version of item 2 initially appeared easier than its modified counterpart, item 2-R (Fig. 5). However, upon delving into the content, students recognized the enhanced clarity and support offered by the revised versions. This finding suggests that for multilingual learners, the initial assessment of difficulty may not always reflect the actual cognitive load required of an item. Therefore, designing items that prioritize clear presentation with visual and spatial arrangement, even if it results in a slightly longer item, can ultimately benefit student comprehension.
Students’ perceptions about the inclusion of irrelevant information were clear if the information does not directly help with problem solving, it should be eliminated from the item. Item 2-R (Fig. 5) attempted contextualization through a narrative involving a student working in a factory—a strategy inspired by an earlier study using the EFCA in life science items among middle school students (Siegel, 2007). However, the students in this study found this approach unconventional and unnecessary for a timed exam solely aimed at finding answers. Some students expressed that they might have reacted differently if such contextualization had been introduced in homework items or if this style of questions were shown in their courses. Evidently, students hold different expectations regarding item formats encountered in homework assignments versus those in exams. This finding suggests that multilingual students prioritize identifying key information that leads to solving the problem. Any additional contextual details, intended to be helpful, can become a burden, especially in time-sensitive assessments.
Effective assessment design for multilingual learners in chemistry requires careful consideration of information inclusion. Extraneous details can impede comprehension, while contextualization strategies must be age-appropriate, relevant to the subject-matter, and aligned with student expectations for the assessment format.
By examining the challenges faced by multilingual students in specific assessment items, this research has identified key features that significantly impact their understanding of what is being asked. The findings underscore the importance of crafting assessments that transcend language barriers and highlight the value of accessible, well-structured assessments in fostering equitable demonstration of content knowledge. The insights gleaned here offer a valuable vantage point for educators to create assessment items that better reflect multilingual students’ knowledge and abilities while mitigating language-related barriers.
1. Identify the linguistic feature that poses the greatest barrier to understanding
2. Evaluate the potential consequences of each modification
3. Strive for a balance between simplifying language and maintaining the level of difficulty of the item
4. Test modifications with a small group. Refine and consider alternative approaches to address the accessibility issue.
5. Seek input from language specialists if possible and chemistry educators.
The following sections discuss some actionable strategies to design more inclusive, accessible and equitable items for students, particularly multilingual learners. While we offer practical strategies for linguistically simplifying assessment items based on the findings of the current and previous studies (Lee and Orgill, 2021), it's important to recognize that there is more than one way to linguistically simplify an item, which may vary based on the topic and/or subject area.
For example, the sample original item in Fig. 10 does not provide any prompts or steps to guide students towards the concept of conservation of mass. It does not ask them to identify knowns and unknowns or set up an equation to solve for the missing mass. There is no mention of the chemical equation for the reactions, which might be crucial for understanding the formation of products and potential gaseous byproducts. The question lacks a clear connection between the law of conservation of mass and the observed discrepancy in mass. A more effective way to present this item would be to rearrange the background information to present smaller bits of information and scaffold the main questions so that students are first asked what the law of conservation of mass is before applying it to the chemical reaction.
For example, the sample original item below (Fig. 11) has words like “constituent ions,” and “electrical neutrality,” which might be unfamiliar or pose difficulty for multilingual learners. The sentence structure is complex, particularly “separation of charges” which could be challenging to parse for students still developing their English language skills. In the revised version, simpler vocabulary has been used such as “breaks apart” instead of “dissociation” and “tiny, charged particles” instead of “constituent ions.” It also has simpler sentences. Please note that we are not suggesting the removal of all challenging chemistry terms from the assessment items. We believe that if these terms are essential to the evaluation and students have been adequately prepared to understand and use them, they should be included.
![]() | ||
Fig. 11 Original item (with complex vocabulary and phrasing) on the left. Revised item (with simpler vocabulary and phrasing) on the right. |
Overall, language barriers can significantly hinder students’ ability to demonstrate their knowledge. Therefore, designing equitable assessments requires careful consideration of language use in order to create a level playing field for students with diverse linguistic backgrounds. Our findings, along with a similar study involving EngL1 students (Gandhi-Lee, 2018) suggest that most of the modifications recommended by the EFCA framework can benefit all students by promoting clearer interpretation of assessment items. The only EFCA recommendation that students in the current study did not find useful was the addition of a storyline as contextualization, which the students found to require significant reading time—time that they did not have available on high-stakes, timed chemistry exams. Other EFCA recommended modifications, such as linguistic simplification of vocabulary and syntax and the use of bold type for emphasis, will help students overcome some of the challenges they identified in assessment items. In fact, based on the results of this study, the EFCA framework has the potential to narrow the performance gap between multilingual learners and EngL1 students in tertiary chemistry courses. However, future research should explore the broader impact of EFCA modifications and how they benefit all learners in more nuanced ways. This focus will be crucial in achieving equitable assessment practices in chemistry, ensuring that all students have a fair chance to demonstrate their understanding of the subject matter.
This journal is © The Royal Society of Chemistry 2025 |