Anna
Bertram
*a and
Carmen
Tomas
b
aSchool of Chemistry, University of Nottingham, University Park, Nottingham, UK. E-mail: anna.bertram@nottingham.ac.uk
bHonorary Assistant Professor, School of Education, University of Nottingham, Nottingham, UK. E-mail: carmen.tomas@cantab.net
First published on 28th October 2022
Engaging students actively in assessment, using a range of formative activities, consistently over sustained periods is a common recommendation in the assessment for learning literature. Despite this, practice still lags behind. Our case study aims to bridge the widening gap between theory and practice by illustrating the application of recent theoretical concepts in practice. The literature contains many examples and research on isolated events to engage students in assessment. Cases that explore engagement in different formative practices, over extended periods of time, are scarce and challenging for practitioners to implement in the absence of examples and evidence. Consequently, whilst adoption of theory informed practices remains limited, research also remains limited. Our case study aims to bridge the widening gap between practice and theory by elaborating a case example for practitioners. The redesign of a third year laboratory module, in which students undertake research projects, is presented. Our case illustrates how practitioners can incorporate assessment for learning principles, outlined in contemporary frameworks (evaluative judgement) considering process and a learning sequence over an entire year. The learning design of a module, before and after, is fully described to exemplify how practitioners can implement theoretical principles in practice. Students’ perception of the value of the new tasks were gathered to inform reflections for practitioners in the implementation of evaluative judgement.
Assessment should be used strategically in the learning process. The literature lends support to the statement that ‘students see assessment as the curriculum’ (Gibbs, 1999) and that whether a task is assessed or not shapes students’ perceived value of the task. There is considerable discussion in the literature about the value of formative alongside summative assessment in the curriculum and how these can be successfully balanced to optimise learning. The proportion of formative assessment in relation to summative assessment remains low (Jessop and Tomas, 2017; Wu and Jessop, 2018). This is perhaps due to the evidence in the literature that students will focus most effort on what ‘counts’ (Harland et al., 2015) and therefore course designers are reluctant to change assessment practices. For example, many frameworks propose more time is allocated to involve students actively in understanding quality but it is, at times, hard to see how these activities may be prioritised over other important practices or activities. Reducing summative assessment, increasing formative, is very challenging in practice.
Bridging the gap between theoretical models and practice requires careful consideration. Good practice in assessment is available but evolves slowly. The changes required, in line with models that we explore in sections below, demand a transformation that goes beyond single interventions. Practitioners often lack the confidence that many of the ideas proposed will be effective and deliver positive outcomes. Creating assessment for learning designs that incorporate multiple interventions and over a period of time poses multiple challenges for practitioners. There is a paucity of models or case examples situated in real contexts for practitioners. Illustrating how theory can be implemented in practice is the first step.
Secondly, institutional cultures still do not promote the reduction of summative and the increase of formative assessments, over sustained periods. Recent research evidences that dominant discourses and culture do not facilitate these changes for individual practitioners (Winstone et al., 2021). Whilst theoretical proposals suggest increasingly sophisticated ideas, institutional cultures do not promote the practices. The gap between practice and theory is widening. Consequently, whilst practice remains so far behind theoretical proposals, building up an evidence base at the scale and scope required, remains compromised.
Our study, in a third-year practical chemistry module, explores how to develop assessment for learning designs in practice. Designing learning and student engagement, over an entire year, is the primary focus of our study. Literature with case examples that illustrate the application of theory into practice is limited. Consequently, the effectiveness of some of the principles, widely recognised as central, is yet to be evidenced. Our case study aims to advance the field by proposing a theory-driven redesign of an existing module in practice. An example, situated in a real context, will show practitioners how they can design learning in practice taking account of the proposals in most recently formulated theoretical frameworks.
Below, we review the existing frameworks for assessment for learning, the evidence and recent reformulations of assessment for learning. In this context, we present the aims and relevance of our study.
Self-regulated learning defines learning as students’ understanding of their own abilities and themselves (metacognition). This has replaced former thinking about learning based on constructs such as intelligence or individual differences (Zimmerman and Kitsantas, 2002). Self-regulation is central to success in life and achievement in academic settings alike (Dignath et al., 2008; Zimmerman, 2008; Sitzmann and Ely, 2011; Clark, 2012).
Peer, self and co-assessment are methods of engaging students that can promote autonomy and self-regulation. Peer learning is defined by Boud (1991) as the “use of teaching and learning strategies in which students learn with and from each other without the immediate intervention of a teacher.” Peer and self-assessment are very closely linked; Boud (1991) described self-assessment as “the involvement of students in identifying standards and/or criteria to apply to their work and making judgements about the extent to which they have met these criteria and standards.” This was later reiterated by O’Donovan et al. (2007) who said that “by involving students in the marking process, they can expand the assessment of learning into an effective learning tool and generate a technique of assessment for learning. Peer and self-assessment techniques thus enable a student to become more autonomous in their learning and help to develop a student's ability to identify their own learning needs.”
Evidence is available on various strategies that address aspects of self-regulation. For example, the formative use of rubrics to engage students in self and peer assessment promotes the development of self-evaluation and reflection skills (Dochy et al., 1999; Jönsson and Svingby, 2007; Reddy and Andrade, 2010; Panadero and Jönsson, 2013; Brookhart, 2018). These are essential skills for self-regulation, autonomy and ultimately life-long learning (Struyven et al., 2005; Carless, 2007; Boud et al., 2018). Comprehensive reviews of the literature (Winstone et al., 2015, 2017; Evans, 2013) also reinforce this essential aspect of instructional environments and how students’ active participation could be promoted. An essential element of active participation is to develop the students’ ability to understand requirements and how their performance aligns with the criteria for assessment.
In chemistry education, the same principles have also been applied and similar types of activities implemented, for example peer review of reports (Jones and Seybold, 2016), peer review and digital badges for laboratory skills development (Seery et al., 2017), engaging students in creating assessment questions using PeerWise (Ryan, 2013; Galloway and Burns, 2015), self-assessment and cooperative learning in mini-lab activities (Branan and Morgan, 2010).
Formative assessment has been shown to be effective in supporting the development of experimental techniques, note taking, data analysis and report writing in a second year chemistry practical module (Seery et al., 2019). Significantly the formative feedback was integrated throughout the whole laboratory module, this was in contrast to previous models which involved the majority of feedback occurring towards the end of the laboratory sessions.
A common limitation, to both practice and research, is that much of the existing literature and evidence base has been drawn from single interventions. All existing frameworks emphasise the need for a sustained engagement in such activities or deploying different activities over time. Recent revisions of theoretical frameworks, to guide practice, have taken account of the integration of these various techniques, their sequencing and process over a period. Below, we present a recent theoretical proposal reframing self-regulation, practice, and the process of engaging students in assessment.
Student self-regulation is an important driver in instructional design. Despite this, instructional designs and assessment practices do not fully embrace the tenets of student self-regulation (Winstone et al., 2021). Evaluative judgement provides a framework to advance our understanding of how instruction can take account of self-regulated learning (Panadero and Broadbent, 2018). Self-regulation is not an individual trait or ability. Self-regulation is self-generated thoughts, feelings, and behaviours that are oriented to attain goals (Zimmerman, 2000; Zimmerman and Kitsantas, 2002).
The components of self-regulation are cognition, metacognition and motivation. Self-regulation involves self-awareness of one's own abilities, the ability to self-assess in relation to a task in order to self-correct and, finally, motivation to engage. Motivation is rooted in the belief and perceptions about one's own abilities (self-efficacy), (Bandura, 1982). Self-efficacy is a predictor of motivation. The beliefs one holds about one's own abilities can determine the time spent on a task and intrinsic motivation (Zimmerman and Kitsantas, 1997) which are, again, associated with success in tasks.
Self-regulation is not an ability but a process. Four levels of self-regulation exist (observation, emulation, self-control and self-regulation) (Zimmerman and Kitsantas, 2002). These levels relate to each other and progression between levels is non-linear. Panadero and Broadbent (2018) propose that various assessment-related interventions would be appropriate at different levels and stages of this process. The proposal remains theoretical but considers instruction and learning, drawing attention to the process elements. This proposal is summarised below (Panadero and Broadbent, 2018):
Despite theory and evidence advancements, adoption in practice lags behind. Recent research reminds us that culture and dominant notions in practice pose barriers to the adoption of the proposed concepts (Adachi et al., 2018; Winstone et al., 2021). Evaluative judgement, despite its advantages as a coherent framework, poses challenges to practitioners. Whilst theoretical developments become more sophisticated, these outpace the developments in practice.
Our study explores a recently formulated theoretical framework, and its implementation in practice. The aim is to provide an important direction and example for practitioners. The model proposed above involves integrating, in practice and over a period of time, a range of support mechanisms to foster evaluative judgement (Boud et al., 2018). This adds questions in practice concerning design, timing, frequency, and sequencing of these activities. The framework requires practitioners to deploy multiple strategies (rubrics, self and peer assessment, reflection), most of which are new to many practitioners. Adding a consideration of process, presents new challenges to most practitioners to begin to visualise what this really means in practice. Moreover, making time for these additional activities requires reducing summative assessment which is another well-known challenge in practice.
Taking into consideration the challenges in practice, providing example cases and stimulating reflection with concrete examples is an important step when considering the widening gap between theory and practice. Translating theoretical frameworks into practice and real contexts requires cases and examples in the first instance. Enabling and inspiring others to embrace the model and principles is a requisite for growing these practices. Growth of cases and at scale, will pave the way for larger scale research into the impact of the theory described on student learning and self-regulation. Our primary aim is to stimulate the discussion amongst practitioners by providing an example.
We have referred to a growing body of evidence, on the effectiveness of proposed techniques for learning. In addition to impact on learning, previous research also shows that students’ perceptions of different modes of engagement also matter for practitioners to understand how best to implement a given technique (Patton, 2012). Understanding students’ experience is an important aspect of practice as, in addition to learning gains, the experience of different tasks poses important considerations for practitioners to implement the framework.
Our aim is to support the advancement of practice within the direction set in the aforementioned theoretical framework of evaluative judgement. Our focus, therefore, is the design of student engagement in assessment. To advance this, as a first step towards exploring the theoretical models in practice, our case first explores the redesign from a traditional approach in chemistry laboratory projects. Secondly, we aim to understand student reactions and experience to the range of techniques deployed to help inform the development and implementation in practice of the initial model. Therefore, our aim is to explore student reactions to and perceptions of the multiple tasks throughout a year of study.
Our study focusses on a theory-based redesign and provides an example for the chemistry education community of how to implement a series of challenging ideas in practice. We consult students on their experience of the initial redesign to gain insights and inform our thinking and revisions of the initial proposal. Therefore a flexible framework, allowing a plurality of perspectives and an emphasis on complex aspects (Stake, 2000; Yin, 2002), is required to capture both the redesign and student experience. The case study, whilst mainly framed within qualitative approaches of enquiry, allows for a variety of information sources and types to be collected (e.g., quantitative and qualitative) throughout the process. The rich descriptions that case studies permit are necessary for other practitioners to learn from this case and extrapolate to their contexts.
A case study offers the possibility of using a range of diverse methods to bring to light aspects described above. This case study is exploratory, aiming to provide insights for chemistry education practitioners of how to approach the design of evaluative judgement in practice. The case study is comprised of two distinct parts to provide insights into the process of redesign by the practitioner and the students’ experience of the new activities proposed.
• Part 1 provides a rich description of the application of principles from theory in a real practical context.
A module redesign, taking into account the theoretical principles and applying them in practice, is the focus. In line with the model, the stages of preparation, practice and self-regulation were included in the redesign. This section provides a rich description of the practitioner's thinking and reflection on theory in practice.
This rich and descriptive case of our module redesign can make a significant contribution for chemistry education practitioners.
• Part 2 explores students’ perceptions of the value of different activities that were included over the course of a year.
The review identified that moving from theory to practice is challenging. One challenge is the additional factors that affect implementation. In addition to the value to learning, students’ perceptions and reactions are also key to the success of embedding new practices. In our initial implementation, it was important to gauge students’ reaction, their ability to engage with the tasks and the value they perceived the tasks having. The effective use of peer assessment in practice requires understanding students perceptions and reactions to the tasks (Patton, 2012). In our first implementation of the design we wanted to incorporate student feedback as a starting point to consider the effectiveness of exposing students to such a range of diverse strategies (rubrics, peer assessment, self-assessment, reflection sessions) most of which are not part of mainstream practice. Certainly, in our context, students might have occasionally encountered some activities to engage them actively. Understanding students’ perceptions of the tasks presented was essential in the first implementation.
We distributed three questionnaires at different points during the course of the year. Each covered a range of activities and stages in line with the proposals of evaluative judgement (see Table 4). The questionnaires captured student perception of the value of the tasks. In order to avoid participant fatigue, three questionnaires were distributed:
• Before preparation workshops (Questionnaire 1, see Appendix 1, ESI†).
• After preparation workshops (Questionnaire 2, see Appendix 2, ESI†).
• At the end of the module (Questionnaire 3, see Appendix 3, ESI†).
Each questionnaire covered a range of the new activities in the module. The focus was on evaluating student perception of the range of tasks experienced throughout the module and as detailed in part 1.
The appendices contain the full questionnaires which were used. This initial case study reports on the questions aimed at capturing the student experience and comments on the activities. Future publications will present fully other exploratory questions aimed to capturing student learning and self-efficacy, which is still under investigation.
The questionnaires included Likert scale type questions to understand the student experience (overall satisfaction) of the sessions. Students’ ratings of their perceived value of the activity using a five-point scale (from not valuable to extremely valuable). Additional questions to understand the value of the sessions in students’ learning and whether they would take part again.
Students also provided open text comments to gain insight into the reasons for their ratings on value of different tasks.
• 116 responded to Questionnaire 1 before taking part in the preparatory workshops.
• 89 responded to the Questionnaire 2, after preparatory workshops.
• 22 students completed Questionnaire 3 (end of the module).
The sample size varies depending on each question. The exact number of students answering each question is detailed where appropriate, as sample sizes vary. See Questionnaires 1–3 in the Appendix (ESI†).
Descriptive numerical summaries are offered in relation to questions where students were asked to declare preferences for future sessions and modes of engagement. Questions where we have quantified answers were either category (yes/no) responses or Likert scales. Percentages, frequency counts and medians or means are used for these.
Likert scale ratings generated a numeric type of response. For students’ ratings of the perceived value of different activities in the preparation stage (see Table 4 workshops 1–3: co-construction of criteria and co-assessment of exemplar activities) we obtained a large sample in the first and second questionnaires. This allowed for us to test students’ ratings of their perceived value of those initial activities during the preparation stage. A Wilcoxon signed ranks test was used given the data derived from Likert scale ratings in Questionnaire 2 (5 point Likert scale). Given the large sample (89 responses to Questionnaire 2) we report summary statistics using parametric measures (mean and standard deviation).
Evaluation at the end of the module (Questionnaire 3), covered activities: self-assessment, feedback review and action planning. This also included other elements like the rubric used throughout the module and tutor feedback to enable us to get a complete picture of the key elements of engaging students. The lower participation as it was the end of the year, resulted in a lower response rate. We could only match eight questionnaires to the responses provided in the previous questionnaire (2). This limits the analyses and we present summary descriptive statistics since sample size of eight responses would not allow for statistical or significance testing. Also, non-parametric measures are used given the low response rate. Medians are used as they are a more resilient measure of central tendency for a small sample.
The third and last questionnaire, also collected students’ justifications for their ratings as open text comments. These open text comments have been quoted directly. The extracts are quoted verbatim to enable readers to interpret the student perception of the various activities. We have used classification of the comments according to positive and negative comments.
Also, we cluster comments in relation to key aspects of self-regulation described in the model: understanding expectations, goal setting, reflection. These are the key concepts referred to in the in the SRL model. Due to the small number of open text comments, we have included the full comments verbatim to allow readers to interpret the comments and views expressed.
The third-year module was broken down into three five-week projects in which students did one project in each of the three areas: inorganic, organic and physical chemistry. In the first week of the module, a series of short workshops were delivered focussing on some of the key transferable skills students would require to successfully complete the projects (Table 1).
Activities | Teaching weeks | |
---|---|---|
Preparation | Transferable skills workshops | 2 |
Introductions to: searching the literature, report writing, presentations | ||
Project 1 | Searching the background literature | 3 |
Preparing a project plan (summative feedback) | ||
Practical project work | 4–6 | |
Data analysis & report writing (summative feedback) | 7 | |
Assessed presentations (summative feedback) | ||
Project 2 | Searching the background literature | 8 |
Preparing a project plan (summative feedback) | ||
Practical project work | 9–11 | |
Data analysis & report writing (summative feedback) | 12 | |
Assessed presentations (summative feedback) | ||
Project 3 | Searching the background literature | 19 |
Preparing a project plan (summative feedback) | ||
Practical project work | 20–22 | |
Data analysis & report writing (summative feedback) | 23 | |
Assessed presentations (summative feedback) |
Many of the challenges described in the literature were visible in this module. For example, module evaluations suggested that students did not understand the assessment criteria and had a poor understanding of what was required in order to perform well. Additional feedback from students also highlighted an issue with the demanding timetable for each project. Very limited time to complete the required work and little time to digest feedback from the previous project was a challenge.
The assessment criteria, with descriptors for different attainment levels, were provided as a document on the Virtual Learning Environment (VLE) with the expectation that students would read and understand the document. When work was assessed, a numerical rubric was attached to the assessment, but these numerical marks did not link with the criteria and descriptors. Students were then expected to cross-reference this with the document containing the criteria and descriptors of attainment levels. However, it was clear from student feedback that this was not happening and the level of engagement with the documentation outlining the assessment criteria was low.
The assessment of the third-year practical module incorporated some laboratory-based elements, a final report and a presentation. Encouraging students to engage with learning outcomes and assessment criteria early in the assessment cycle was a consistent challenge. Many attempts had been made to increase the prominence and signposting of the information via the VLE, but this lack of engagement remained an issue. A lack of understanding of the assessment criteria is likely to diminish the learning gain and may be responsible for the lack of engagement with feedback, and it was therefore critical that this issue was addressed.
There was an expectation that students would develop transferable skills, which are key to the success of student project work, throughout the module. However, the experience of running this module suggested that this was challenging as these transferable skills build on skills previously practiced and knowledge previously acquired, thus requiring students to work with a level of autonomy not previously required. In light of the initial observations and feedback after running the project module, we wanted to expand the transferable skills sessions at the start of the academic year and focus on supporting students in learning independently. In other words, developing self-regulation and autonomy as an important basis for project work.
• Reduction of assessment load to allow for more opportunities to engage in learning.
• Enhanced transparency of assessment (e.g. use of rubrics in marking and feedback).
• Engaging students in understanding quality, criteria, self-assessing and planning.
Following the evaluative judgement as our theoretical model, as already discussed, we wanted to explore in the preparatory stage of the module how Self-Regulation Levels (SRL) 1 and 2 could be integrated. It is very important for the effective running of projects to enhance students’ understanding of what is expected of them (SRL 1 observation) early in the process and foster self-evaluation (linking to SRL 2 Emulation), all of which are linked to motivation (self-efficacy). The steps taken in the redesign are detailed below.
Workshop | Activities | Teaching weeks | |
---|---|---|---|
Preparation | 1–3 | Transferable skills workshops (see details in Table 4) | 2–3 |
Project 1 | Preparing a project plan (formative feedback) | ||
Peer assessed presentations on project aims (formative feedback) | 4–5 | ||
Practical project work | 6–9 | ||
Data analysis & report writing (summative feedback) | 10–12 | ||
Assessed presentations (summative feedback) | |||
Student self-evaluation against assessment criteria for project | |||
Feedback and Reflection | 4 | Feedback reflection, comparison of assessor's feedback with self-assessment. | 19–20 |
5 | Action planning for future development & Project 2. | ||
Meeting with assessor (see details in Table 5) | |||
Project 2 | Preparing a project plan (formative feedback) | 21–22 | |
Peer assessed presentations on project aims (formative feedback) | |||
Practical project work | 23–26 | ||
Data analysis & report writing (summative feedback) | 27–29 | ||
Assessed presentations (summative feedback) | |||
Student self-evaluation against assessment criteria for project | |||
Feedback and reflection | 6 | Feedback reflection, comparison of assessor's feedback with self-assessment | 30–31 |
7 | Action planning for future development | ||
Meeting with assessor (see details in Table 5) |
We redesigned the module learning outcomes, subsequently broke these down into assessment criteria (Table 3) and created a module assessment rubric with descriptors for four levels of attainment for each assessment criterion (Fig. 1). The literature contains a range of papers on the design of assessment rubrics, and the work of Mertler (2001); Dawson (2017) was very helpful. Overall, we had five overarching module learning outcomes: team working and time management; Safety and good laboratory practice; Technical competence; Knowledge, understanding and critical thinking; Technical writing and presentation skills. These were then broken down into a total of 17 assessment criteria. To emphasise alignment and enhance transparency of assessment for students, the rubric represented all the criteria, indicated where they would be assessed and the weighting of each. A small, summarised extract of the rubric is shown in Fig. 1. The full assessment rubric has been attached in Appendix 4 (ESI†).
Module learning outcomes | Assessment criteria |
---|---|
Team working & time management | Plan & devise experimental work for each laboratory session to make best use of the time available |
Communication & team working | |
Safety & good laboratory practice | Identification of hazards, relevant precautions & disposal to ensure safe experimental work |
Use of Good Chemistry Laboratory Practice (GCLP) | |
Technical competence | Accurately record experimental procedures & observations (laboratory notes) |
Quality experimental work | |
Knowledge, understanding & critical thinking | Breadth of knowledge of the background research area of the project |
Understanding of the methodologies/techniques/tools relevant to the project (e.g. synthetic routes, software, equipment, spectroscopic methods) | |
Ability to describe relevant results | |
Technical writing & presentation skills | Report is well structured |
Visual elements (graphs, figures, tables): range of types of visual elements, quality and adequate presentation (labels, captions) | |
Nomenclature, schemes, equations, figures. Style (use of technical vocabulary) | |
Experimental data reported in journal style | |
Degree of consistency of the formatting of the document (font size, references list, use of bold/italic, layout, spacing, etc.) | |
Accuracy of the grammar, structure of sentences and division of text into paragraphs | |
Use of in-text citations to support development of argument | |
Presentation of experimental results to a small audience using PowerPoint (or similar) |
![]() | ||
Fig. 1 Summarised extract from the assessment rubric outlining the information available to students for each criterion. |
The rubric type chosen was analytic (i.e. matrix type of display), as this type of rubric is preferred for learning due to the greater level of detail and feedback they provide than holistic rubrics (Brookhart and Nitko, 2019). It was essential to have a comprehensive assessment rubric like this to enable us to design the previously described activities to engage students with the module learning outcomes. In contrast to previous practice, this same rubric was used for students to self-assess themselves and to communicate levels of performance by the assessor.
Workshops | Outline | |
---|---|---|
1 | Promoting understanding of learning outcomes and self-assessment. | Setting the context – highlighting the differences between approach in 2nd and 3rd year laboratory modules. |
Project description given, tasks requiring teams of students to: | ||
• Summarise skills required to successfully carry out project work, as those: | ||
○ Already developed. | ||
○ Requiring development. | ||
• Generate a list of proposed learning outcomes for the module. | ||
• Propose how these learning outcomes could be assessed. | ||
At the end of this series of exercises the actual learning outcomes for the module were shared and there was very close correlation. | ||
2 | Students generating assessment criteria for project reports and assessing exemplars. | An introduction was given to report writing, including discussion of some exemplars. Students worked in small groups to co-review three different reports. After each report was considered a class discussion took place led by the academic facilitator. |
A comparison was made between student 3rd year project reports and research papers, with the aim being to ensure students could see where their current learning fitted in. | ||
Students were asked to come up with a list of assessment criteria for the report. | ||
The assessment criteria for the report were shown, these were discussed and comparisons with those suggested by students made. We found that students were able to propose criteria which matched were very closely aligned. | ||
3 | Students generating assessment criteria for presentations | Colleagues delivered two contrasting presentations and students were asked to highlight positive and negative elements. This exercise culminated in students suggesting assessment criteria for the presentations. We found that students were able to propose criteria which matched were very closely aligned. |
The workshop would rely on successfully engaging students in this reflective dialogue. As such, we created online boards to which students could anonymously submit contributions. Students were asked to:
• Identify skills already developed in practical modules in earlier years, which they would need for project work.
• Identify skills they still needed to develop to successfully complete project work.
• Propose the module learning outcomes.
• Suggest how the module learning outcomes could be assessed.
Students worked in groups to discuss these questions and then commented using their mobile phones or tablets on the electronic discussion board which received very good engagement. After each activity, students’ comments were summarised in a reflective discussion by the facilitator.
In this module, students are required to write a report on their project in the style of a Royal Society of Chemistry journal, the first activity in this session was for students to work in groups to look at exemplar reports. Students were asked to pick out elements which they thought were good and those which needed improvement. Following on from this, students were asked to suggest the assessment criteria that should be used for the reports.
Our rationale for this approach was that students, given the right support, would be able to identify for themselves what constituted a good report. Orsmond et al. (2002) describe the benefits of using exemplars in increasing students’ understanding of assessment criteria. Interestingly, they also note that peer assessment can be more objective than self-assessment.
Workshop 3 was delivered using the same strategy as workshop 2 but focussing on presentations. Staff presented using two very different styles and students again suggested assessment criteria. Online discussion boards were used to facilitate both of these sessions.
In the newly designed module, students have two weeks to search the literature and write a project plan before commencing practical work. During this period, we introduced peer-assessed presentations. The purpose of the peer-assessed presentations is three-fold: to ensure students have a clear understanding of their project area before commencing project work; to provide an opportunity to practice a presentation which is not formally assessed; to increase understanding of the assessment criteria for the presentations. The peer-assessed presentation gives an opportunity for students to operate in SRL 2 (emulation) following on from earlier sessions on presentations in which exemplars were given (SRL 1 observation).
Students were asked to self-assess themselves against the marking criteria used to assess the projects and this self-assessment was then submitted alongside the project report. The aim of this is to encourage students to self-assess (SRL 2 emulation) more routinely and to work towards a situation where they are able to self-assess as a matter of routine and in the absence of exemplars (SRL 3 self-control).
Workshops | Outline | |
---|---|---|
4 & 6 | Feedback and reflection | These workshops are delivered after each practical project and are designed to encourage students to engage with the feedback provided by the assessors. The aims are to: |
• Facilitate student's understanding of what they’re doing well in their project work & areas they could improve in. | ||
• To help students develop their evaluation skills to enable them to better judge the quality of their work. | ||
This workshop is focussed around the use of the proforma in Appendix 5 (ESI). | ||
5 & 7 | Action plan for future learning | These workshops are designed to follow on from the workshop in the previous week. Having reflected on their feedback and considered the significant feedback they’ve received, students are guided through a process of writing an action plan for their future development. This process involves students considering the following: |
• What can you do to build on the positive feedback you’ve received? | ||
• How can you develop your work in the areas where you received some criticisms? | ||
Students are also asked to develop their own evaluation by considering the following: | ||
• What can you add to those above? | ||
• What do you want to keep doing in the future? | ||
• What do you want to change or improve in the future? | ||
This workshop is focussed around the use of the proforma in Appendix 6 (ESI). |
Activity | N | Mean | SD |
---|---|---|---|
Co-construction of criteria | 87 | 3.46 | 1.06 |
Exemplar co-assessment | 87 | 4.00 | 0.90 |
A Wilcoxon signed rank test was performed to explore whether students rated the two activities differently. Out of the total 89 responses, only 87 were complete so that we could match and compare responses on the two activities. The results show that there was a significant difference in participants’ ratings of the two activities (W = −3.766, n = 87, p < 0.001). The co-construction of criteria having a lower rating (M = 3.46, SD = 1.06) than the co-assessment of exemplars activity (M = 4.00, SD = 0.90). The scale was a five point Likert scale (1 not valuable, 5 extremely valuable). This difference is associated with a medium effect size, r = 0.285. In terms of student preferences, 35 participants rated the activities the same, 12 ascribed higher value to the co-construction of criteria and 40 thought the co-assessment exemplar activity was more valuable (Table 6).
An additional breakdown of students’ ratings is provided in Fig. 2 below. It shows the frequency of students’ ratings of the value of the task engaging them in co-constructing criteria.
![]() | ||
Fig. 2 Frequency count of students’ ratings of the value of the co-construction of criteria task (1 not valuable – 5 extremely valuable). |
Fig. 3 below provides a graph showing the frequency counts for different ratings by 87 students of the co-assessment of exemplars. This activity involved students assessing a range of example reports in class and discussing assessment criteria.
![]() | ||
Fig. 3 Frequency count of students’ ratings of the value of the co-assessment of exemplars task (1 not valuable – 5 extremely valuable). |
Lastly, in addition to understanding differences, we wanted to capture the student experience of the full set of workshops designed to engage them in this very early stage, overall. When asked ‘would you like to take part in future sessions similar to these ones?’, out of 87 respondents, 79 (90 per cent) indicated they would take part in future similar sessions.
Activity | Median | N |
---|---|---|
Self-assessment | 3 | 21 |
Tutor assessment | 4 | 15 |
Reviewing feedback | 4 | 18 |
Assessment rubric | 4 | 20 |
Action planning | 3 | 5 |
The ratings above indicate that students perceived some activities to be more valuable than others. Students were also asked to justify their ratings. A total of twelve students (of 22) provided justifications for their ratings which are summarised below.
“helpful to have a breakdown of the marks”
“excellent guideline”
“excellent for seeing what is required of me”
Other comments showed enhanced ability to identify actions to improve, which is a higher level of self-regulation:
“clear to find how to improve”
“good to know where to improve”
Lastly, important insights about how understanding rubrics is not always straightforward were made:
“useful but sometimes difficult to interpret”
“vague an non-precise categories”
Self-assessment posed challenges for students. We quote justifications that provide insights into how the task was not a valuable experience:
“I would not submit a report I didn't think was excellent, therefore self-evaluation seemed a little pointless,”
“everyone wants to give themselves high marks”
“I feel the rating itself did not benefit me much”
Two of the less positive comments referred to the rubric and limited understanding of standards as the reason for the limited value of this task: “categories vague”; “difficult to judge own work. Lacking understanding of what a good report consists of”
Other comments related to the timing of the self-assessment being submitted prior to receiving feedback.
“do not feel it is useful before feedback is given”
However, on a number of cases, students referred to important thinking processes like reflection and clarifying expectations that were positively impacted.
“good to think about my role in the project”
“the self-assessment helped me to understand the requirements of the lab report”
“I would have never thought back about the progress”
“good to really take in improvements required”
“helpful to plan where marks were coming from”
“useful to know where marks are obtained and distributed”
“getting ratings from a professional is always good”
“they know the project in great detail and can add really detailed comments”
“good for specific questions”
Other positive comments were generic (“very useful”, “good to see others’ opinions of the work”)
Students do not value their own evaluations or peer assessment as highly.
The remaining comments on tutor feedback also revealed common problems of the transparency and fairness of written feedback and marking:
“inconsistent and harsh”
“some marks are not clearly justified so it would be more helpful to know why these marks were awarded”
“helpful to analyse where I went wrong”
“important to understand positive/negative aspects of work”
“give time to reflect and break down feedback”
“very informative”
“good to compare own opinion with assessors”
“have time to reflect on performance and can discuss with people on how to improve”
“good for actually making me self-reflect”
“being told to learn what we’ve done bad and accept it”
“useful to know where marks were lots, and how to improve”
One comment referred to situations where the activity may not be valuable:
“only positive feedback meant not much to add or improve on”
One comment (of the 11) was not positive: “less useful”
A minority of students did not really understand what the benefit of the feedback review was. Comments were mostly positive.
“helped with making a plan of what to do on the second project”
“good to think about how to implement self-evaluation”
There are many examples in the literature on how various elements (e.g., use of rubrics, peer/co/self-assessment, action planning) in isolation, as single interventions for students, impact on learning (Panadero and Jönsson, 2013; Winstone et al., 2017; Brookhart, 2018). Supporting learning over a sustained period, by engaging students in the known elements of self-regulation (understanding the task, expectations, practice and own performance), is less well evidenced given the challenges for practitioners (Winstone et al., 2021) to implement learning design. The aim of our case study was, first, to illustrate how we can inform our practice in line with the evaluative judgement framework. An evaluation of student perceptions of the value of a range of tasks, over the course of a whole year, provides an exploratory insight into what students may value most or perceive more accessible in the context of a year-long exposure to different ways of being engaged in their learning as part of assessment.
In this context, our case contributes to chemistry higher education, but also to other disciplines, in two ways. Our redesign case may inspire module leaders to consider greater alignment of criteria and engagement of students in learning. More importantly, our redesign case illustrates how contemporary theories of student autonomy and self-regulation can be articulated in our instructional design over a sustained period, and this could extend to any discipline. Our project module redesign demonstrates how to address key challenges in the literature regarding assessment over-load and student engagement. The redesign has involved reducing the number of summative assessment points, increasing the number of formative activities with the aim of familiarising students with criteria and descriptors, training them as markers by marking and discussing quality of exemplars in the early stage of the module.
Regarding the student experience, based on their views we gathered, the initial results, overall, are positive and encouraging. These initial findings also present variable levels of student acceptance and perceived value which are an important element to reflect on for future implementations.
The evaluative judgement framework proposes that a range of activities is implemented, as illustrated in our theoretical discussion. Models, such as the one by Panadero and Broadbent (2018), include a whole range of activities (e.g., self-assessment, reflection, evaluation of the work of others). Students’ perceived value was variable for different elements. Students attributed a greater value to the rubric and activities that took place at the preparation stage. In particular, assessing the quality of exemplars, using the rubric and then discussing this with the tutor. Co-construction of criteria was perceived to have a lesser value in comparison with the active engagement in the co-assessment of exemplars. Ninety per cent of students declaring they would like similar sessions to run in the future.
Tutor feedback, and reflection on tutor feedback in the second part of the year, were valued positively by the students. Self-assessment and action planning activities, which took place in the second half of the year, were the least valuable in comparison to the whole suite of activities according to students’ ratings. Many factors may have influenced these outcomes. Students completed self-assessment and action planning individually in a workshop setting. In this initial iteration, we have not yet included opportunities for discussion, debate or input from an expert or in a peer group.
Our initial evaluation suggests that self-assessment and action planning are activities that we need to reconsider. In the broader context, these activities were encountered only once in the programme of study. These new, and more demanding, activities may have played a part in how students were able to appreciate their value. Self-assessment and reflection are advanced skills in self-regulation and considering the level of mastery of other aspects (e.g., understanding quality, SRL level 1) may influence students’ ability to perform these higher-level skills in an explicit manner. This is an important consideration for practice and future research.
Students seemed to rate activities in the preparation stage of the year more highly (e.g., reviewing report exemplars). Some of the other activities like self-assessing, identifying skills and synthesising module learning outcomes were less highly valued. Theoretical models suggest that these activities are all part of promoting self-regulation and consequently, part of a whole. Yet, in practice, the initial feedback provided by third-year students on their experience, reveals that not all the activities are easily accessible.
In our future implementations we will seek to find better ways of implementing self-assessment and action planning. Also, to promote familiarity integrating into a programme-level assessment strategy across all years of the degree would also be important. It is hardly surprising that when applied in a standalone third year module students are not easily able to change their approaches and views on assessment. Introducing more self-assessment activities throughout the degree course may support students to understand their value and be able to honestly self-assess themselves, and in turn develop skills of self-regulation. Theoretical proposals for practice (Panadero and Broadbent, 2018) may incorporate these considerations in their models to continue to support and guide practitioners.
Self-assessment and action planning are challenging to students. We made an assumption that students would be able to self-assess and action plan as routine. In our experience, students need guidance and support. We know from the evaluative judgement framework that self-assessment and action planning require students to operate at SRL levels 3 and 4. The ‘Developing Engagement with Feedback Toolkit’ (DEFT) framework (Winstone and Nash, 2016) suggests that students can do these activities on their own. In our experience, however, we have gained insight into the higher level of demand for students and the need to support students in this process. Self-assessment and planning actions might require either scaffolding with an expert and, possibly, a much greater level of preparation of students than we had anticipated. We have taken on board these initial insights and are introducing changes to these activities to incorporate peer support in the future.
This initial study has focussed on how students perceived the value of a range of tasks during a year. The study provides initial insights about tasks perceived to be more valuable. Evaluations of self-regulation are still necessary to provide additional insights and continue developing the theoretical models and practice guidelines. Our study has drawn primarily from student self-reported perceptions of value. Additional methods should be used to evaluate and verify the perceived value and the impact on learning.
In addition, larger samples are also needed. Our exploratory study shows some differences in how students valued different tasks. Due to limitations of sample size (Questionnaire 3), we have not been able to conduct statistical tests on the full set of student ratings. However, descriptive statistics suggest that this line of enquiry would be worth pursuing further.
Future publications will provide additional insights drawn from parts of our evaluation looking at the impact on student learning.
Reducing the number of projects and, in turn, the amount of summative assessment is quite a bold move in an environment where the culture is to set regular summative assessments. As already discussed, moving towards increasing the number of formative activities is still seen as high risk. A common experience in practice is that students will focus their efforts on activities that ‘count.’ We need to change student perceptions about higher education, assessment, feedback and their role as partners with assessors in the process. Our redesign case study can provide inspiration for practitioners by helping to think of an alternative concept of how to design and support assessment in practice.
In this initial exploration, student perceptions and experiences, show a higher value attributed to the activities in the earlier part of the year. These activities aimed at increasing their understanding of what was required of them to do well. Students commented on the benefit of having a comprehensive assessment rubric and on using exemplar reports in co-assessment exercises. These were relatively small interventions with a high value for students. Practitioners may be encouraged to introduce these types of developments which are of high impact for students.
Students’ perceived value of the self-assessment and action planning activities was lower than we expected. In future implementations we will consider how best to prepare students for these tasks.
To conclude, we recommend the introduction of activities aimed to develop students’ evaluative judgement skills and their understanding of assessment criteria, primarily for student learning and secondly for assessment transparency. Decreasing the summative assessment load may seem challenging, for all the reasons highlighted earlier, but is necessary to create time to introduce more valuable activities aimed to:
• Enhance the transparency of assessment.
• Engage students in understanding quality and criteria.
• Develop students’ self-assessment skills and action planning for future development.
Overall, our case study demonstrates the design of a chemistry module to promote student learning and autonomy. The theoretical proposals are challenging for practitioners, since implementing each of these activities requires preparation for both staff and students. In particular, staff have to learn about each of the proposed techniques and adapt them to their context. Our findings provide an indication for practitioners about which types of activities they could work on introducing first. We have evidence of activities that may be of greater value and more accessible to students.
Colleagues might prioritise the elements that are more accessible and valuable to students. Rubrics and co-assessment of exemplars are good starting points. Introducing these right at the start we also recommend was effective. Reflecting on tutor feedback was another valuable activity for students. Other elements recommended in the literature (e.g. self-assessment and action planning) require more preparation and practice for students to understand and obtain the desired value from them.
Footnote |
† Electronic supplementary information (ESI) available. See DOI: https://doi.org/10.1039/d2rp00213b |
This journal is © The Royal Society of Chemistry 2023 |