Alice
Lu
,
Joy B.
Ghrayche
,
Reyne
Pullen
* and
Stephen R.
George-Williams
*
School of Chemistry, University of Sydney, Sydney, Australia. E-mail: reyne.pullen@sydney.edu.au; stephen.george-williams@sydney.edu.au
First published on 25th January 2025
The COVID-19 pandemic prompted the shift to online learning, including a change from paper-based (Type-P) exams to online, open-book (Type-O) exams. This study investigated the impact of the sudden shift from Type-P to Type-O exams on the nature of exam questions, and the students’ and academics’ experiences. Type-P and Type-O exams were analysed, focussing on question type, and Bloom's Taxonomy classifications. Type-O exams had a decrease in drawing questions in favour of short answer questions, and a slight shift to lower-order thinking was seen. Other changes were mostly insignificant. Semi-structured interviews with students revealed the main origins of stress for Type-O exams related to technological failures. Students noted they prepared notes for ease of searching for Type-O exams, and optimal memorisation for Type-P exams. Students who had taken both exam types revealed a preference for Type-P exams, as they preferred drawing answers. Semi-structured interviews with academics revealed that writing questions for Type-O exams required more thought to avoid questions with searchable answers. However, academics enjoyed the conveniences of the online format, which includes the automatic marking of multiple-choice questions. Academics appeared to have an astute awareness of the students’ experience with the differing exam types. Our findings suggest that Type-O exams could be successfully integrated into university systems, however there are important considerations that should be addressed.
• identifying knowledge gaps allowing students to re-evaluate their learning strategies and teachers to adapt their teaching strategies;
• providing evidence of learning to hold universities accountable;
• and certifying the student's attainment of knowledge and skills.
As such, it is important that assessments accurately reflect a student's understanding of course content in order to cater to these purposes. In general, there are two main types of assessments:
• formative assessments such as ongoing weekly quizzes;
• and summative assessments such as final examinations (Williams and Wong, 2009).
Traditionally, final examinations are closed-book and paper-based and are usually completed in-person under invigilator supervision (herein referred to as Type-P) (Williams and Wong, 2009; Nsor-Ambala, 2019). However, due to the COVID-19 pandemic, students could no longer attend in-person exams. Instead, exams shifted to formats that did not require face-to-face contact, such as online written exams (Dicks et al., 2020). Due to the remote nature of these exams, many were transitioned to online open-book (herein referred to as Type-O), granting students access to notes and textbooks, with some allowing open-internet access. The risk for academic dishonesty prompted universities to implement online invigilation services, such as ProctorU (Wuthisatian, 2020; Vazquez et al., 2021). However, students often raise concerns for safety and privacy around the use of external proctoring companies (Petillion and McNeil, 2020).
Type of questions | Description |
---|---|
a Only possible in online examinations. | |
Multiple choice | Question with multiple answers to choose from |
Multiple choice with calculation | MCQ requiring numerical calculations |
Fill in the blank | Filling in tables, blank spaces; one- or two-word answers |
Short answer | Answers is a few sentences to a paragraph |
Calculation | Numerical calculations including working out |
Drawing | Drawing diagrams, structures, arrows, labelling, etc. |
Equation | Requiring formulas, laws, equations, expressions etc. |
Drop-down boxa | Box with a select number of answers to choose from |
Hybrid | Two or more types of questions combined |
Hybrid with drawing | Hybrid questions requiring drawings |
Proof | Deriving or proving laws, expressions, etc. |
Further to this, each question can be assessed using the revised blooms taxonomy (RBT) (Anderson and Krathwohl, 2001) which classifies questions into six categories. Based on this, questions can be further classified as either lower- or higher-order of thinking (Agarwal, 2019; Stringer et al., 2021), as shown in Table 2.
RBT level | Order of thinking | Description |
---|---|---|
Remember | Lower | Recognising and recalling facts |
Understand | Lower | Explaining what facts mean |
Apply | Higher | Using formulas, rules, or ideas |
Analyse | Higher | Breaking down information and finding relationships between parts |
Evaluate | Higher | Judging the value, validity, or reliability of information |
Create | Higher | Combining parts to make new wholes |
Students have also reported differing sources of stress depending on the type of exam taken. Pre-exam, Type-O students reported lower levels of pre-exam anxiety when compared to closed-book exams, stating access to notes as the main reason (Myyry and Joutsenvirta, 2015; Nsor-Ambala, 2019). During exams, those taking Type-O exams reported additional stresses about technological failures such as battery issues, and difficulties downloading or uploading sections of the exam (Washburn et al., 2017; Shraim, 2019; Petillion and McNeil, 2020). One study that surveyed 342 students showed that 75% of participants noted online exams increased their stress levels (Shraim, 2019).
Additionally, a study by Iannone and Simpson (2014), showed that most students surveyed from two universities stated a preference for closed-book paper-based exams for mathematics, citing reasons such as less opportunity for cheating. Those who disagreed felt that closed book exams favoured those with a better memory (Iannone and Simpson, 2014). It also appears that preference for computer-based and paper-based exams may depend on the discipline being examined (Okocha, 2017).
Chemistry exams often require students to draw structures and mechanisms. As such, transitioning to Type-O exams introduces various complexities, particularly in integrating drawing questions in an online format. Therefore, we sought to investigate the changes on the nature of exam questions between Type-P and Type-O exams. Additionally, this was an unexpected change for both students undertaking exams and academics writing them. Therefore, we also aimed to examine the impact on the experience of students and academics to gain insights into the suitability of Type-O exams in future. To this end, we sought to answer the following questions:
(1) How did the shift to Type-O exams impact the exam questions, if at all?
(2) How did the shift to Type-O exams impact the experiences of students, if at all?
(3) How did the shift to Type-O exams impact the experiences of academics, if at all?
Analysis of one exam was performed by three different researchers independently. Discussions were had until all category guidelines were agreed upon by all three researchers. The remaining 51 exams were analysed by one researcher, with regular meetings to ascertain consistency.
Categorisation of exam question types were assigned according to Table 1. Classification of questions was performed by utilising the RBT (Anderson and Krathwohl, 2001), and further separated into higher- or lower-order according to divisions used in other studies (Agarwal, 2019; Stringer et al., 2021) according to Table 2.
The percentage of exam items assigned to each category in an exam was calculated using the marks allocated to each question type.
A confidence interval of 95% was applied, with a Bonferroni correction to prevent Type I error as comparisons were performed on all four categories. Cramer's V was utilised to determine the effect size of any significant results (Kim, 2017).
Audio and video recordings were taken during the interviews, but only audio was analysed. Audio recordings were transcribed using the AI Transcribe tool in Microsoft Word before being cleaned by listening to recordings and correcting the generated transcript.
Inductive thematic analyses were conducted on the interviews (Braun and Clarke, 2006). Three researchers independently analysed one interview to develop the first set of codes which were discussed and condensed until 100% agreement was achieved. To validate this codebook, two researchers independently coded a second interview and compared agreement. 90% and 77.8% agreement were reached for the student and academic interviews, respectively. To reach the 80% agreement threshold, the academic codebook was further refined and condensed until 100% agreement was reached. A third researcher then audited the coding of the second interview to validate the new codebooks. Using these codebooks, one researcher coded all interviews and generated three and six additional codes for the student and academic interviews, respectively. Another audit was conducted on all interviews to determine the validity of the coding. The final codebooks contained 14 and 36 codes for the student and academic interviews, respectively.
First-year | Second-year | Third-year | ||
---|---|---|---|---|
a Significant p-value (<0.05). b Large effect size. | ||||
2019–2020 | p-Value | <0.001a | <0.001a | <0.001a |
Cramer's V | 0.406b | 0.435b | 0.337b | |
df | 8 | 10 | 8 | |
2020–2021 | p-Value | <0.001a | 0.997 | <0.001a |
Cramer's V | 0.242b | 0.038 | 0.224b | |
df | 8 | 8 | 10 | |
2019–2021 | p-Value | <0.001a | <0.001a | <0.001a |
Cramer's V | 0.242b | 0.446b | 0.405b | |
df | 8 | 10 | 10 |
The percentage of exam items coded to each question appears to change between Type-P and Type-O exams (Fig. 1). A decrease in drawing questions and an increase in short answer questions in 2020 were attributed to university mandates to avoid file uploads during exams.
![]() | ||
Fig. 1 Percentage of exam items coded to different question types for (A) first-year (B) second-year and (C) third-year chemistry exams from 2019–2021. * = significant p-value <0.05. |
After the lifting this mandate in 2021, some drawing questions reappeared in first- and third-year exams but not in second-year exams, with no significant difference being shown between second-year exams in 2020 and 2021. Despite the return of some drawing questions, there remained a difference in question types between Type-P and Type-O exams, possibly owing to decisions made by the second-year academics.
The increase in drop-down questions between 2019 and 2020 remained between 2020 and 2021. First-year exams favoured drop-down questions more than second- and third-year exams, likely due to the increased exam difficulty in later years which necessitates longer text responses.
There was a notable increase in multiple-choice questions (MCQs) between 2019 and 2020, due to the lack of MCQs in some 2019 units. In 2020 and 2021, MCQs were included in these units, possibly due to ease of implementation or changes in university guidelines on the number of MCQs allowed per exam.
Similarly, other universities also reported resorting to non-handwritten exam questions, such as “pick the correct statement”, for questions that would traditionally require hand-drawn responses (Dicks et al., 2020). Those who used file uploads to accommodate drawing questions reported several technical problems (Clark et al., 2020; Rodríguez Núñez and Leeuwner, 2020), with Clark et al. (2020) resorting to the use of chemical drawing tools, whereby students were able to convert chemical structures into SMILES strings to denote their final responses. Along with our findings, this suggests that drawing questions could be replaced with non-handwritten equivalents.
In summary, there was a difference in common question types between Type-P and Type-O exams. Type-O exams inhibited drawing questions in favour of short answer questions, a trend that persisted even after file uploads were permitted in 2021.
Type-O exams saw a shift from ‘apply’ to ‘understand’ questions, indicating lower-order thinking (Fig. 2). Significant differences were observed between 2019–2020 and 2019–2021, though first- and second-year changes had small and medium effect sizes, respectively (Table 4). This may have been due to the difficulty of including calculation and drawing question in Type-O exam types, which are classified as ‘apply’ questions (Crowe et al., 2008; Pugh and Gates, 2021). It is also possible that the academics aimed to reduce student stress and anxiety by lowering the RBT item type. However, these changes were relatively minor, as indicated by their effect sizes.
![]() | ||
Fig. 2 Percentage of exam items coded to each of the RBT levels for (A) first-year (B) second-year and (C) third-year chemistry exams during 2019–2021. * = significant p-value <0.05. |
First-year | Second-year | Third-year | ||
---|---|---|---|---|
a Significant p-value < 0.05. b Large effect size. c Medium effect size. | ||||
2019–2020 | p-Value | 0.026 | 0.001a | <0.001a |
Cramer's V | 0.104 | 0.145c | 0.173c | |
df | 4 | 5 | 5 | |
2020–2021 | p-Value | 0.024 | 0.131 | <0.001a |
Cramer's V | 0.115 | 0.093 | 0.235b | |
df | 4 | 4 | 5 | |
2019–2021 | p-Value | 0.005a | <0.001a | <0.001a |
Cramer's V | 0.12 | 0.197c | 0.230b | |
df | 4 | 4 | 5 |
Third-year exams showed a significant shift in RBT item type with a large effect size, both between 2019–2020 and 2020–2021, possibly due to syllabus changes in 2021. A significant difference was also noted from semester 1 of 2019 to semester 1 of 2020 (Tables S2 and S3, ESI†), likely due to the rapid transition to Type-O exams. However, no significant difference was found between semesters 2 of 2019 and 2020, suggesting that academics rapidly adapted to the new format (Tables S2 and S3, ESI†).
A study by Davies et al. (2022) also found that the rapid shift to online exams during COVID-19 resulted in exams with minimal changes to question RBT type. However, their questions were limited to “single best answer” questions in the first three levels (i.e. Remember, Understand, and Apply).
Of interest to note, there was a tendency toward lower-order thinking questions in Type-O exams, despite the open-book format. This could pose a problem for long-term learning retention, as higher-order questions are associated with better retention (Johanns et al., 2017). The rapid transition to Type-O exams likely resulted from time constraints for academics, limiting their ability to write higher-order questions that are more suited to an online open-book format. However, this tendency toward lower-order thinking questions seemed to pervade even at higher levels of chemistry (i.e., through second- and third-year units). As such, this tendency could further be explained by a lack of formal training in writing effective exam questions. While there is no consensus on how Type-O exams should be written, the literature suggests benefits from using higher-order questions carefully designed to assess genuine critical thinking skills and avoid easily searchable answers, for example, through utilising authentic assessment principles (Eilertsen and Valdermo, 2000; Brightwell et al., 2004; Williams and Wong, 2007; Johanns et al., 2017).
Overall, questions in all exam types favoured those of a lower-order thinking. Any statistically significant results were found to have a small effect size, despite the minor shift toward ‘understand’ type questions.
‘perform better as well because. this is a more formal event, this is something that's important and they’re having everyone come in and people watch you.’ – Student 2 (Group P/O).
One student revealed that memorising information for closed-book Type-P exams was stressful, mirroring findings seen in the literature (Nsor-Ambala, 2019). Another source of stress raised by students taking Type-P exams was organising transport and locating the exam venue which, cited by three of the nine students:
‘I gotta make sure I catch a few trains early just in case there's delays, and then finding the room’ – Student 4 (Group P/O).
The second-most prominent stressor raised by these students was the time limit. Nine of the 17 students interviewed believed that the time allotted for Type-O exams was insufficient, which resulted in feelings of stress. For example, students noted that typing out answers for calculation questions required more time than it would have for a written exam:
‘I think there was a bit of a push for time, especially with the calculations. It obviously took a bit longer than just writing it down ‘cause we had to type it up.’ – student 9 (Group P/O).
Seven of the 17 students mentioned auditory distractions in their domestic environments as a source of stress during Type-O exams, with one student recounting:
‘One exam, my neighbours, it was Friday afternoon or Thursday afternoon, it was 3:00 PM, they started having a dance party, like an absolute raging party, and I sat there and listened to house music for the most of my exam. And it just made me so angry, like why? Why do I have to deal with this?’ – Student 5 (Group P/O).
Other stressors raised by the students were primarily related to technology. Most prominently, concerns about internet failure were raised by six students, aligning with findings from other studies (Elsalem et al., 2020). Additionally, four students noted that uploading handwritten responses was a source of stress when taking Type-O exam including a lack of appropriate technology, failure to upload, and one student stating they ‘honestly nearly forgot.’ Two students also commented on issues with the exam interface, with one stating:
‘You couldn’t flick through the questions very easily, so… you didn’t really know the overview and how to pace the exam properly.’ – Student 17 (Group O).
These concerns of technological failures align with findings seen in multiple other studies (Washburn et al., 2017; Shraim 2019; Clark et al., 2020; Elsalem et al., 2020).
In addition, four students further noted that they were initially unsure of the expectations of a Type-O exam, citing the lack of past papers and sample exam questions as a significant concern:
‘Cause you basically have no clues about how the online exams will look like… when it transfers to online, the uni teaching staff, they also don't have the past exam paper of online exam, so they can't provide you any clues. Sometimes they don't provide you the sample. The sample they give you, they put nonsense questions there, just gave you a look of the online exam, how online exam feels like. But the questions aren’t really related.’ – Student 7 (Group P/O).
However, after this initial adjustment period, 12 out of the 17 students stated that they had eventually adjusted to the Type-O exams. These findings are in line with what have been reported in the literature, whereby students who had only been exposed to paper-based exams initially found online exams more stressful, but believed that familiarising themselves with the exam format would decrease their stress (Wibowo et al., 2016; Afacan Adanır et al., 2020).
Therefore, it appears that students experience differing origins of stress depending on whether they are taking Type-O or Type-P exams.
‘Because it's online exam and is mostly open book, I study less 'cause I can have all the information on my computer.’ – Student 14 (Group O).
As such, it appears that the closed-book nature of Type-O exams affects the strategies used by students to prepare for their exams.
Eight out of nine students in Group P/O reported skipping questions they were unsure of during both Type-P and Type-O exams. Similarly, all eight students in Group O who had only taken Type-O exams reported the same strategy. However, their methods for flagging skipped questions varied due to the exam format. Four of the nine students in Group P/O mentioned circling the question number, leaving it blank, or holding the page during Type-P exams, while six out of eight students in Group O either left questions blank, used the LMS flagging function, or wrote question numbers on paper during Type-O exams:
‘When we used to do the exams on paper… I had these like little tabs I would attach that would be different colours that meant different things… like, this is a question I just haven't done… But that's not an option with online… Like I would write down on a piece of paper that I need to definitely go back to this question 'cause I haven't done it, so that was slightly more convoluted.’ – Student 3 (Group P/O).
Three additional exams strategies were identified by students who took Type-O exams. Because of the online, open-book, open-internet (herein referred to as open-access) nature of Type-O exams, students had access to external materials for use in their exams. All 17 students interviewed mentioned looking at their notes at some point in the exam – either for factual recall, or to check their answers after writing a response – with one student remarking:
‘I think there were some questions which are like, sort of fact checking… And these are the sort of questions I would probably look up or check a book, 'cause I feel like they don't really test your sort of understanding of the topic.’ – Student 1 (Group P/O).
Our interviews further revealed that students’ Type-O exam strategies were influenced by two conflicting motivations: the desire to finish on time vs. the desire for legibility. Five out of nine students in Group P/O and five out of seven‡ students in Group O mentioned that they used full sentences, and believed that full sentences conveyed their ideas more clearly despite taking longer to write. Four students stated that they used a mix of full sentence and dot point responses depending on time constraints. The remaining two students (one from each group) mentioned they used dot points for efficiency, especially if they were running out of time. For questions that involved calculations, 11 of 17 students remarked omitting lines of working for sake of time, while the remaining six noted including all lines of working for the sake of clarity. Regarding the tension between these conflicting motivations, one student noted:
‘I think in general I did full sentences except for when I was really short for time or if it was a question that I thought was one of the sorts of questions where they really were just wanting to see one or two essential points.’ – Student 3 Group (P/O).
Considering the complexities of typing chemistry responses, three students from Group P/O and five students in Group O mentioned they skipped formatting as they found it time consuming. The remaining nine students chose to include formatting for the sake of legibility, despite acknowledging that it was time-consuming:
‘It probably was just a visual thing, because I can see something in my brain, it can mean like something to me, but it might not mean the same thing to the marker… I have to make sure that it's right and that what I'm trying to convey is coming out.’ – Student 2 (Group P/O).
Furthermore, four students in Group P/O revealed that the convenience of typing allowed them to edit their exam responses. However, some remarked that this slowed them down during the exam:
‘…we might do a lot more self-editing and rewriting stuff which we didn't even need, but it's more just, if you give them the option, then we naturally go for it, so that could be a factor that slows down our speed in the online exams.’ – Student 8 (Group P/O).
Therefore, it appears that students maintained some common exam strategies between Type-P and Type-O exams. However, students also revealed adopting new strategies in order to enhance their performance in Type-O exams, which was mainly influenced by the open-access format, and the conflicting desire to either complete exams on time or maintain legibility.
‘when teachers see your paper, they can see like your process of working out’ (Student 12, Group O).
Previous studies have revealed that students who were taking chemistry exams showed lower acceptability of computer-based exams when compared to those taking other courses (Okocha, 2017). This could be attributed to the need for chemical drawings, reaction diagrams, calculations, and special symbols in the field of chemistry, which lends itself better to paper-based responses.
Regarding Type-O exams, all 17 students stated that the potential for technical failures, and issues with uploading responses made this exam format more stressful. Despite this, six of the nine students in Group P/O still enjoyed having the option to draw and upload, suggesting the potential to successfully implement this feature in Type-O exams if technical issues could be mitigated. Interestingly, only one of the students from Group O preferred having drawing questions in Type-O exams:
‘…it can be really helpful to some students who need that other way to show “I do know what you're talking about let me show you.”’ – Student 11 (Group O).
All 17 students also expressed preference for un-proctored Type-O exams due to both privacy concerns and technical issues. However, five students recognised the need for proctoring, as Type-O exams were perceived to be more susceptible to academic dishonesty. This has been noted in the literature, whereby cheating increased with the shift to online exams (Newton and Essex, 2024).
Despite the issues students expressed about Type-O exams, five students in Group P/O and six students in Group O enjoyed the open-access format, and appreciated not having to memorise large amounts of information. Additionally, four students in Group P/O and two students in Group O enjoyed the convenience of the remote nature of Type-O exams:
‘Uhm, in the paper-based, I think definitely the drawing questions good, 'cause then it's just so nice just to draw straight onto the exam papers… I still prefer doing it at home, in the comfort of my own home, it still outweighs going into uni and doing it in-person.’ – Student 4 (Group P/O).
In light of this, it appears that if technological errors were eliminated, Type-O exams may have the potential to be successfully implemented. This is also suggested by Wibowo et al. (2016) who found that students would enjoy online exams if technical issues were alleviated, and mock exams were provided for preparation.
‘In organic chemistry we use mechanisms a lot, and so obviously you're restricted to be able to do that in an essay-type based question, or you would have to frame it so that students gave descriptions on what is happening rather than drawing it.’ – Academic 1.
Academics also mentioned the importance of writing questions that were not readily searchable online or within students’ notes, to effectively gauge students’ understanding. These findings have been reflected in other studies, where academics mentioned checking if search engines would identify solution to their questions (Johnston and O’Farrell, 2020).
One academic mentioned the need to write multiple Type-O exams to reduce the risk of academic dishonesty and prevent unauthorised disclosure of questions online. However, the academic noted this created an additional workload that is not typically encountered when writing Type-P exams, because the replacement exam is usually recycled and never published.
In addition, two academics stated they felt less supported when writing Type-O exams, noting that for Type-P exams, the education support team oversaw the formatting exams. However, for Type-O exams the academics were assigned the extra requirement of formatting and editing questions for use in the LMS themselves and found it time consuming.
After being presented with Fig. 1, all six academics commented on the university mandates restricting their use of drawing questions – a factor also raised prior to being shown the results. Academics also raised that it was the online delivery of Type-O exams that made it more difficult to implement drawing and calculation questions.
Additionally, two academics raised that changes in the curricula may have contributed to the changes in question types, irrespective of the change in exam type.
‘Well, so I'd say the online exam wasn't any different in a sense to the paper-based exams as there were both types. And of course the higher order questions are the ones that I use as the HD and distinction barrier questions. And the low order ones for the pass and credit.’ – Academic 2.
This is consistent with other studies that have shown that at least 50% of exam questions were lower order, allowing students to pass and higher performing students to be distinguished (Dicks et al., 2020). Two academics believed that Type-O exams contained more higher order questions because of the open-access format, which would allow students to answer ‘recall’ questions by searching their notes or the internet. In contrast, one academic believed that all exams contained mostly lower order questions, regardless of format:
‘I feel many of our students struggle with anything higher order. So if there's a bias, it would be more towards lower order questions. Uhm, yeah, genuinely applying knowledge in an unfamiliar context is something that happens very seldom in USYD exams in my experience…it's to do with how our students are trained, it's to do with their expectations coming in.’ – Academic 4.
Academics were then shown Fig. 2. All academics suggested that the restrictions caused by Type-O exams and university mandates limited their ability to incorporate higher-order question types, which contributed to the decline in RBT item type:
‘I think again, consistent with the modes of delivery. So clearly yeah, 2020 where we went online and only had the short answer response perhaps changed… Uhm, we couldn't probe a lot of things because it was all sort of descriptive. That sort of looks like it made a little bit of a comeback in 2021 where we could have some file uploads. And so yeah, higher order questions could be slotted back in again.’ – Academic 1.
Five academics also raised that the introduction of new curricula may have also contributed to the lower RBT item types. Three suggested that changes in exam writers may have had an influence, suggesting that different examiners may write questions with different levels of thinking:
‘Different people teach different courses. Different people ask different types of questions.’ – Academic 4.
Four academics raised that the main advantage was the automatic marking of questions by the LMS. However, another four academics noted that negative aspects of marking Type-O exams were due to technological issues such as manually verifying answers to check for errors by the LMS, and that markers could not access an exam at the same time without the risk of overwriting each other's marking.§
Two academics enjoyed the convenience not handling physical papers and being able to mark from home. However, two academics believed online marking was more time-consuming as they could not write on the exams. One found it challenging as they preferred marking one question at a time to ensure the marking scheme is applied consistently across students:
‘So when I mark, if I've got to mark six questions in the paper, I like to mark all of question one first and then all of question two first and then all of question three so that I can be consistent more easily across the papers. But that was much more difficult in an online format, so I didn't do that in an online format. So I did spend a lot more time going back and checking that I had applied my marking scheme equally to the first marked papers and the last marked papers. ‘ – Academic 3.
‘the trepidation of exams will be the same regardless’ – Academic 6.
Regarding the students’ time management strategies, four academics believed that this was specific to the individual student.
Three academics mentioned the chronological approach as they noticed questions being left empty towards the end of the exams. Two mentioned it was possible that the questions toward the end of the exam may have been due to students spending time formatting their responses, or looking through their notes.
Overall, it appears that most academics had astute perception of the students’ experience of Type-O exams.
Regardless, it is crucial to consider the impact of exam type on the experiences of both students and academics. Student interviews revealed a shift in the origins of stress when comparing both exam types. The stressors for Type-O exams were mainly related to technological failures, such as internet dropouts. Despite the mentions of stress undertaking Type-O exams at home, it was also noted that the Type-P environment was also a cause of some students’ stress. It may be possible for universities to offer alternative locations for students with difficult home environments to take their Type-O exams. Furthermore, this study identified the potential for reducing students’ stress regarding Type-O exams by providing them with adequate preparatory material to help familiarise them with the delivery and format of Type-O exams. These findings align with literature reports where students who had only been exposed to paper-based exams found online exams more stressful but believed that familiarising themselves with the exam format would decrease their stress (Wibowo et al., 2016; Afacan Adanır et al., 2020). This could be easily implemented by providing students with mock exams on the same LMS used in final exams, allowing students the opportunity to familiarise themselves with the new format. While they voiced a preference for Type-P exams overall, students demonstrated their ability to quickly adapt to the unfamiliar exam format, by adjusting their note preparation style, modifying their response style to suit the Type-O format, and largely maintaining exam strategies between the two exam types.
As for academics, the two main setbacks that academics raised about their experiences with Type-O exams, were due to the open-access nature of the exams and their perceived lack of support. Academics attributed changes in exam questions to changes in curricula, and the inability to add drawing questions due to a lack of upload capabilities in the first year. Despite this, our analyses of the exam questions revealed that academics continued to write assessments at similar level of RBT item type. However, since current literature suggest the benefits of writing higher-order questions for Type-O exams to enhance learning retention and ensure answers are not easily searchable, it is important to ensure that academics are adequately trained in writing such questions.
As such, this study has shown that Type-O exams have the potential to be implemented in the university system without significantly impacting the student and academic experience; however, it is essential to consider key factors to ensure they uphold exam integrity.
A major area for consideration is the rise of artificial intelligence (AI), such as large language models or generative AI, which presents both challenges and opportunities in the post-COVID landscape. For instance, unauthorised use of AI tools by students could pose a significant risk for academic dishonesty. Online proctoring tools like ProctorU offer potential solutions by providing remote invigilation services, but their effectiveness and associated privacy concerns require careful consideration (Hussein et al., 2020; Selwyn et al., 2023; Kuleva and Miladinov, 2024). On the other hand, AI technology itself could be employed to minimise the likelihood of academic dishonesty through the use of AI-based proctoring or AI-generated adaptive exams (Nigam et al., 2021; Surahman and Wang, 2022). Additionally, AI could be used to automate marking and to provide efficient feedback, potentially reducing the workload for academics.
As such, balancing the benefits and potential risks associated with the rapid evolution of AI in relation to Type-O exams is a critical area for future research.
Furthermore, it is important to consider technological solutions that could address several of the issues raised in this study. For instance, the challenges with incorporating drawing questions or uploading answers in Type-O exams could be mitigated by permitting the use of digital drawing devices, such as tablets. However, this approach raises concerns about equitable access, as not all students may have personal access to such devices. To address this, universities could establish digital testing centres equipped with digital drawing devices. Such centres could ensure that students have equal access to necessary technology, thereby maintaining fairness in the Type-O examination process. These testing centres could potentially reduce students’ stressors relating to technological failure, however, may reintroduce the stressors of travelling to the required testing venue. The implementation of digital testing centres could also facilitate the reinstatement of in-person invigilation, thus combining the benefits of Type-O exams while maintaining the academic integrity of the Type-P exam setting. Further investigation would be required to understand the implications of this strategy.
Footnotes |
† Electronic supplementary information (ESI) available. See DOI: https://doi.org/10.1039/d4rp00290c |
‡ During an interview with a student in Group O, the researcher forgot to ask this prompt. Thus, responses were only obtained from seven students instead of eight. |
§ Of note, the LMS marking issues were eventually addressed with the implementation of Gradescope. However, this did not occur until after these interviews were conducted. |
This journal is © The Royal Society of Chemistry 2025 |