Skip to main content

COVID-19 and alternative assessments in higher education: implications for academic integrity among nursing and social science students

Abstract

Background

COVID-19 and its associated restrictions called for innovations in higher education teaching and learning space with many universities resorting to online teaching and alternative assessments. However, little has been done to understand the academic integrity implications in alternative online and non-invigilated assessments.

Aim

This study explored the perceptions of higher education students regarding academic integrity in alternative assessments.

Methods

Cross-sectional mixed method design following the parallel convergent approach was utilised in this study. A convenience sample of 380 undergraduate and postgraduate nursing and social science students completed an online survey on academic integrity behaviours associated with alternative assessments.

Results

High risk (31.7%) of academic misconduct was perceived among young people (18-24 years old). Collusion was common among nursing students (24.5%) and cheating likely to occur in assessments with longer duration—between 2 and 4 hours (18.8%) and between 1 and 2 weeks (46%). Qualitative data resulted in 274 findings and three themes— (i) impossible to cheat; (ii) easy to cheat and (iii) understanding the consequence of cheating. Suggestions for preserving academic integrity in alternative assessments were also made from the qualitative data.

Conclusion

Like other forms of traditional assessments, alternative assessments have increased risk of breach of academic integrity; however, with the right strategies, they could serve as effective means of assessing learning outcomes.

Introduction

Over the years, higher education has been responsive to global issues. In 2020, when COVID-19 was declared as a global pandemic by the World Health Organisation (World Health Organization 2020), all aspects of life, including higher education, were affected (Aristovnik et al. 2020; Toquero 2020). To reduce the spread of the COVID-19 virus, many restrictions including social distancing were put in place. As a result of these restrictions, higher institutions of learning in 185 countries were closed (Marinoni et al. 2020). To prevent the long term impact of closing institutions of higher education, many technological innovations such as online teaching and assessments became the norm for both developed and developing countries (who are less technologically rich) around the world (Adnan and Anwar 2020; Paudel 2020). Although online and non-invigilated assessments are associated with challenges (Alruwais et al. 2018; Guàrdia et al. 2017), they became the global norm during COVID-19 pandemic and may be considered in higher education to curb the potential for academic misconduct associated with sophisticated Artificial Intelligence applications.

Prior to COVID-19 pandemic, higher education has been striving to ensure assessments are fit for achieving learning and graduate outcomes (such as employability or industry readiness) with recommendations to move from test or examination-based assessments to assessments that test higher order cognitive skills (Bloxham and Boyd 2007; Merta Dhewa et al. 2017). Bloxham and Boyd (2007) argue that although examinations are widely used to assess learning outcomes, their disadvantages far outweigh their advantages. Advantages such as reduction of risk of cheating, testing knowledge and increasing understanding of concepts during end of semester revision were reported. On the other hand, disadvantages of examinations/test such as inferior/superficial quality of learning, anxiety provoking, lack of continuous learning (leaving learning to end of semester), promoting memorisation and limiting originality and restriction of opportunity for real-world learning have been explored (Bloxham and Boyd 2007). Although examinations/tests are highly criticised for assessing knowledge and understanding at a superficial level, they can still be used in certain assessments where memorisation is required.

In response to the COVID-19 global pandemic, Australian institutions of higher education, like many countries, rose to the challenge of online teaching and assessments (Bozkurt et al. 2020; Gamage et al. 2020; Reedy et al. 2021). Traditionally, in the pre- COVID-19 era, many end of semester assessments were invigilated/proctored examinations whereby students are required to gather in examination halls and complete assessments within a specified period. Although the strict invigilated examinations caused high levels of anxiety and stress for students (Clutter et al. 2017), it is perceived to preserve academic integrity as opposed to online assessments (Johnson 2018). However, due to COVID-19 restrictions, invigilated examinations were replaced by alternative assessments via various online platforms, where students completed their semester assessments remotely without a proctor.

Alternative assessments became widespread and undertaken throughout the year 2020 as a measure to reduce the spread and manage uncertainties of COVID-19 virus. Alternative assessments comprise various types of assessments such as oral assessment or VIVA via Zoom or similar video conferencing, submission of video recordings of presentations, live psychomotor skill demonstration via Zoom, take-home open book assessments, time-limited online open-book assessments and assignments. Take-home long and short answer assessments were completed in Cadmus (an online assessment platform) or submitted via Turnitin. Using real-world scenarios, the alternative assessments were designed to assess learning outcomes and higher order capabilities such as critical thinking, reasoning, problem solving, skill acquisition, synthesis, conceptualisation, evaluation and communication skills in order to reduce or eliminate the chances of academic misconduct (Guàrdia et al. 2017; Roksa et al. 2016).

For years, online assessments are perceived as being associated with high risk of cheating or academic dishonesty (Costley 2019; Harmon et al. 2010; Kocdar et al. 2018). However, interventions such as appropriate design of assessments that are aligned to learning outcomes; oral assessments to complement online examinations and maintaining culture of academic integrity in higher education have been cited as means of preventing academic dishonesty in online assessments (Akimov and Malin 2020; Harmon et al. 2010; Peterson 2019). Based on experiences of academics and current evidence on assuring academic integrity in online or non-invigilated assessments, the alternative assessments were designed based on current best practice to prevent or reduce the risk of academic misconduct. Relative to the research setting, some strategies were implemented to safeguard academic integrity for the alternative assessments. They include development of problem-based or application of theory to practice questions, presentation of photo ID for all the VIVA’s, a five-minute video providing ID and explaining one answer from the take home assessment, the assessor marked the video alongside the written assessment. Cadmus and Turnitin plagiarism detecting tools were also utilised for the various alternative assessments.

Given the importance of assuring academic integrity and the role of assessments in achieving learning/graduate outcomes in higher education, it is imperative to explore the students’ perceptions of alternative assessment methods in order to gain insights into possible academic integrity issues associated with alternative assessments in higher education. Therefore, at the end of the 2020 academic year, we explored students’ perceptions of academic integrity issues associated with alternative assessments.

Methods and materials

Cross-sectional mixed method design was used for this study. Cross sectional research is effective for gaining a snapshot from a given timeframe (Wang & Cheng, 2020). Surveys are applicable for data collection used in a cross-sectional study. The mixed method approach in this study followed a parallel convergent pattern. Both statistical and narrative views are considered when applying a mixed method approach (Creswell, 2017; Tashakkori, Johnson, & Teddlie, 2020). Strengths of convergent mixed method design consists of it being efficient, collecting both qualitative and quantitative data at the same time, and allowing for the same viewpoint to be obtained (Creswell & Clark, 2017). Creswell and Clark (2017) highlight that the method allows for the analyses to occur separately, and comparisons of inferences occur before merging the information throughout the discussion.

A self-reported online questionnaire (Qualtrics powered) was developed based on various evidence on higher education teaching, assessment and academic integrity behaviours (Adesile et al. 2016; McCabe and Trevino 1993; Ramdani 2018). The survey consisted of both quantitative and qualitative questions; it was made up of 22 items expected to be completed within 15-20 minutes. It consisted of both 5-point Likert scale (1-strongly disagree to 5-strongly agree) on self-reported and perceived academic integrity statements and open-ended questions to provide further information on Likert scale responses. This paper presents an aspect of a larger study that investigated students’ perception of alternative assessment on their learning, academic integrity and mental health during the pandemic. The cross-sectional study explored the perceptions of students regarding the alternative assessments delivered during the 2020 academic year as a result of COVID-19 restrictions (Additional file 1).

To ensure the rigour of the data collection tool, the face validity of the research questionnaire was reviewed by two experts in higher education in the two disciplines. The Associate Dean of Teaching and Learning in both disciplines reviewed and provided feedback on the survey instrument. The feedback was then incorporated into the survey and piloted with four students within one undergraduate course to ensure comprehension and appropriateness of the questions (Bowden et al. 2002) while assessing both face and content validity. Thereafter, appropriate revisions were made to the survey questionnaire before data collection.

Participants

The participants invited to complete the survey were from two schools (School of Nursing and School of Arts and Humanities), across stages within their courses in a public university in Australia. Students who have undertaken alternative assessment in the 2020 academic year were invited to complete the online survey. The survey was available from August to December 2020 for both domestic and international undergraduate and postgraduate students. Students who completed the pilot survey were excluded from the main study. In all, 380 participants completed the survey.

Ethical consideration and data collection

Ethics approval was obtained from the prospective university’s Human Research and Ethics Committee (2020-01533-ADAMA). The study information sheet and informed consent were embedded in the survey link and a forced response applied. All students provided consent prior to completion of the survey.

Convenience sampling was applied as all undergraduate and postgraduate students across two schools and multiple courses were invited to complete the survey. An invitation email was sent to students by the course/programme directors and associate deans of teaching and learning of the two schools to avoid potential for power-imbalance and conflict of interest between students and teachers (authors) and to encourage voluntary participation in the study.

Measures

Demographic data

Biographic data included age and level of education, programme/course of study and number of subjects/units enrolled in one semester. To ensure confidentiality, students’ identity data were not collected.

Academic integrity measures

We used academic dishonesty behaviours (or behaviours that were discouraged during the assessment) such as seeking help from peers, collaborating with others and seeking help from experts/ professionals/ senior colleagues to describe academic integrity concerns. These behaviours were presented on a Likert scale and students agreed or disagreed with them.

Qualitative measures

Open-ended questions were included to expand on the responses from the Likert scale. These also provided participants the opportunity to share their experiences with academic integrity issues associated with the alternative assessments. Participants responded to the specific open-ended question “Do you think the alternative assessment made it easier for students to cheat? Please explain”.

Data analysis

Quantitative data analysis

Statistical analysis of quantitative data was carried out in SPSS version 26. The quantitative segment evaluated the relationship between demographics (age, course, level and stage of study, and number of units enrolled) and academic integrity measures, namely, seeking help from peers, collaborating with others, and seeking help from experts/ professionals/ senior colleagues. Frequencies and percentages highlighting the distributions across levels of the demographics for the measures were then reported. The chi-square test was carried out to establish significance or otherwise of these relationships. A logistic regression was formulated to further quantify any differences that may exist at the various categorical stratification or segmentation of our study participants.

Qualitative data analysis

Thematic deductive and inductive data analysis of open-ended questions was undertaken in NVivo 12 to generate codes that represent students’ perceptions of academic integrity in the alternative assessments (Feng and Behar-Horenstein 2019). Deductive analysis was guided by the research questions and inductive analysis undertaken by reading the raw data back and forth and immersing in the qualitative dataset. Word Cloud visualisation was also used to show the word frequency within the qualitative data (Fig. 1).

Fig. 1
figure 1

Overview of qualitative responses via word cloud

Results

Quantitative findings

The modal age bracket of the respondent was 18-24, and mostly pursuing undergraduate degrees (83.9%). Majority of the respondents were first (29.2%) and second (34.2%) year undergraduate students. Half of the respondents (51.0%) enrolled in three or more units per semester. Variant forms of assessment and timeframes were used to assess students’ competence and understanding of the learning outcomes for their enrolled units. Majority of assessments were completed within a 24-hour window (48.2%), followed by those scheduled to be completed between 2 to 4 hours (38.7%).

The indicator scale used to measure academic integrity for alternative assessments was sufficiently reliable with a Cronbach alpha score of 0.77 (> 0.7). Response rate varied across academic integrity indicators (seeking help from peers, collaborating with others and seeking help from experts/ professionals/ senior colleagues). Although there were academic misconduct concerns, the results show that majority (> 68%) of students did not seek help from their peers, more than three-quarters (> 78%) did not collaborate with others, and more than four-fifth (> 80%) of students did not seek help from experts and therefore were not involved in these possible cheating behaviours (Table 1).

Table 1 Background characteristics of study sample and test of association between academic integrity indicators and study demographics

With regards to the nature of association between study demographics and academic integrity indicators, it was found that seeking help from peers was associated with age of students (χ2(4) = 13.631, p = 0.009) and course of study (χ2(1) = 3.884, p = 0.049). Students in the 18-24 age bracket often sought help from their peers (31.7%). The multivariable analysis revealed that students in this age brackets were at least 38% more likely to seek help from peers compared to the other age brackets (Table 2). Students pursuing nursing (25.5% [n = 45]) were found to seek help from peers compared to their counterparts from social sciences (15.2% [n = 19]). Nursing students were 21.5% more likely to seek help from peers compared to Social Science students. The majority of students who did not seek help from their peers, (84.8% [n = 106]) were enrolled in social sciences and (75.5% [n = 139]) were enrolled in nursing.

Table 2 Multivariable analyses of the association of demographic factors with propensity to commit assessment dishonesty such as seeking unauthorised help and collaboration

Additionally, an association was found between collaboration with others and course of study (χ2(1) = 7.722, p = 0.005). Again, higher proportion of students in the nursing programme (21.6%) collaborated with others to work on the alternative assessments compared to their counterparts from Social Sciences (9.6%). The adjusted odds ratio indicates that nursing students were 2 times more likely to collaborate with others compared to social science students (Table 2). Although, some proportion of students reported to have sought help from experts/ professionals/senior colleagues, no significant associations were found with the study demographics (p > 0.005) (see Table 1).

There was no significant difference between the stages of study and the percentage of students with academic integrity concerns. However, more than one fifth (> 20%) of first year students responded favourable to seeking help from their peers, collaborating with others, and seeking help from experts. Second year students were less likely to collaborate with others (12.9%) than seek help from peers and experts (> 19.7%). Twelve (< 17.9%) students in their third year indicated they asked for help from their peers, collaborating with others, and sought guidance from experts. Postgraduate students (28.6% [n = 6]) were more likely to seek help from their peers than collaborate or ask experts for help (see Table 1).

In terms of the duration of the assessment and academic integrity, it was found that higher proportions of students sought help from others when the duration for the assessment was greater than 4 h. For instance, for assessments of durations between 2 and 4 hours, nearly one-fifth of the students (18.8%) sought help from experts/professionals/senior colleagues compared to 46.0% for durations between 1 and 2 week(s). Notably, the upper bounds of the 95% confidence intervals of the adjusted odds ratio indicate that there is higher propensity for students to seek help from experts and professionals when assessment durations are relatively long (Table 2).

Qualitative findings

Thematic content analysis of the qualitative dataset resulted in aggregation of 274 findings into three themes and word cloud used to visualise the content of the qualitative dataset (Fig. 1). The themes focused on whether alternative online assessments made it easier for students to breach academic integrity. The themes are (i) impossible to cheat; (ii) easy to cheat and (iii) understanding the consequence of cheating. While some students believed that alternative assessment made it easier to cheat, others believed the structure of the ‘questions do not allow for cheating’ and that students are aware of the consequences of cheating. The probability of cheating depended ‘on the ethics of the student’.

Impossible to cheat in alternative assessment

Majority of students perceived cheating was not easier in alternative assessments, they cited; ‘application of theory to practice’, ‘critical thinking requirement’ in the assessments, ‘limited timeframe’ or ‘time restrictions’ there was no time ‘to cheat nor think of ways of how to cheat!’, ‘availability of tracking devices’ ‘ethical standard’, morality as some students ‘value academic integrity and want to personally get the most out of the course and have marks that reflect own efforts’ or ‘comes down to each individual’s own integrity’. Personal circumstance of students was also perceived as a deterrent for cheating in alternative assessments. For instance, ‘being online student and not having friends.

Well-designed assessment questions were indicated as a way of maintaining integrity in alternative assessments; ‘cleverly structured questions’ as ‘without understanding the content it would have been impossible to answer the questions no matter the access to books’ or ‘if you don’t understand the information then you cannot apply the information obtained from course materials’ and ‘the questions required quite in depth answers’. Other students were of the view that plagiarism detector software such as Turnitin and/or Cadmus prevented them from thinking about cheating; ‘Turnitin would ensure if students collaborated it would be discovered’ and ‘as an open book exam you can’t really cheat unless you had someone else log in and do it for you’. Some also felt the alternative assessment was ‘an individualised exam. Not a group exam’ so did not think of cheating.

A few students believed that with appropriate pre-assessment preparation by teaching staff, it was unlikely to cheat; ‘students were well prepared for these assessments, therefore the likely hood (sic) of students cheating is minimal’. On the part of students, ‘proper preparation and organisation required no need for cheating as it was open book’. Perceptions of students who strongly believed that the academic integrity was preserved, is summarised by one student as:

No, as the questions were structured differently so in the viva's you are asked in real time so you either know it or you don't. With my science exam the questions were structured very differently. You could not simply google the answer, I didn't use any of my note all I used was my textbook to give me some background knowledge so I could use that information to answer the question in my own words. It was submitted through turn it in (sic), so everything had to be your own work, the questions were very challenging and not basic questions they actually required researching just in order to answer the question

Easy to cheat in alternative assessment

For those who believed cheating was easier, they cited ‘access to academic resources or notes or books’, ‘internet’ ‘google’ and ‘friendship or familiarity with course mates’ ‘too much allocated time for assessment’ and ‘being an on-campus student with friends’ as the main reasons for cheating. Collusion and collaboration were cited as the main forms of cheating in the alternative assessments. Cheating was perceived to be prevalent among students who ‘believe that they have not had sufficient support in their studies’. Some students noted that it was easy for them to ‘message for answers or get perspective’discussed with colleagues in the library’ ‘a phone call with other students discussing answers’ or ‘easily share information using social media as nobody is invigilating’ or ‘yes because there is no invigilation to stop people from collaboratingoranyone could be undertaking the exam on their [student’s] behalf’. Some students were of the view that cheating was common in ‘generally poorly written’ assessment. One student stated ‘Yep, everyone did. There were a whole range of social media formats to share answers’.

Having friends or close classmates was also considered a facilitator for cheating in alternative assessments: ‘I guess so, if you had a group of friends that where [sic] in the same class you could probably meet up and do it together’ and ‘I don’t know any of my peers in my units[subject or course] so I had no one to collaborate with’. It was perceived that, academic integrity was hard to maintain in alternative assessments that were not well designed:

‘I would say that it's probably better to design an assessment that requires explanation rather than google-able answers.’ ‘For facts you could google or multiple-choice questions that come straight from Quizlet, absolutely.’

‘100% easier for students to cheat. If they wanted students could google answers, read answers straight from the textbook, be facetiming friends asking for answers, I heard someone say they used 'google home' to search for answers for them, getting together in groups to sit the assessments.’

Understanding the consequence of cheating

Students were aware of the consequence of cheating as all students have completed the University’s academic integrity course. ‘Completing the academic integrity course, I believe everyone did their assessments independently without collaborating or contacting others’ as such ‘all students are aware of the consequences involved with academic misconduct and would not want to jeopordise (sic) their studies by doing the wrong thing’. Consequences were also narrated as; ‘cheating eventually catching up on them in future’ ‘cheating on themselves’, ‘the disservice they are doing will come back to haunt them’, Understanding the consequences and outcomes of academic integrity also served as a demotivation for cheating in alternative assessments. The consequence of academic misconduct in future practice was narrated by one student as:

‘I think to a degree there isn't a problem with looking up elements of the assessment as you go, but it is problematic if someone went in to practice without having put in the effort to truly learn and grasp certain elements of the unit/subject content’

Discussion

The COVID-19 crisis has significantly impacted the higher education sector. Social distancing protocols and the conflicting need for continuity of education and the necessity to contain the spread of the virus meant that universities shift their operations online (Rashid and Yadav 2020). The shift to the online mode in teaching and learning presented systemic challenges, allowing little time for university administrators to conduct in-depth risk assessments of academic integrity (Slade and Benson 2020). This study provided a serendipitous opportunity to examine the practicality and implications of alternative assessment and non-invigilated examinations and tests vis-a-vis academic integrity in higher education.

Thus, the purpose of the current study was to explore the perceptions of university students on academic integrity concerns associated with alternative and non-invigilated assessments, implemented during the COVID-19 pandemic. Findings from the study suggest that on the average, more than three quarters (> 75%) of students did not engage in any of the three cheating behaviours associated with alternative assessments. This figure correlated well with the qualitative findings that alternative assessments are hard to cheat. For the nearly 25% students that perceived one or more cheating behaviours in alternative assessments, there was increased risk of cheating among younger people (18-24 years) and nursing students. The study also found that the longer the duration of the assessment, the higher the risk of cheating. Qualitative evidence suggests that while majority of students perceived that alternative and non-invigilated assessments are hard to cheat, few students believed that it is easy to cheat in these assessments but, at the same time, students were aware of the current and future implications of academic misconduct.

Findings from our study showed that majority of students did not engage in cheating behaviours in the alternative assessments. Furthermore, most students did not perceive cheating to be easy in alternative assessments. The data showed that although some students did not cheat out of personal and ethical values, nearly one fourth of students accessed unauthorised resources, colluded and collaborated with their mates or sought help from senior colleagues or experts. This figure is high compared to other studies with traditional invigilated on-campus assessments and online assessments of 8 and 12% respectively (Harris et al. 2020; McCabe 2005). When age is controlled to 25 years or less, our data shows high perception of academic misconduct (24.5%) in online alternative assessments than the 12% previously reported by Harris et al. (2020) but contrary to that of McCabe (2005). McCabe (2005) study shows that, with written assignments, 24% of undergraduate students are likely to “receive unpermitted help from someone on an assignment” while 42% will “work with others on an assignment when asked for individual work” (otherwise known as collusion). Perhaps, the use of academic misconduct tools and design of questions and other strategies in our study may have contributed to the observed reduced risk of collusion.

In addition, the study revealed that the number of units/subjects enrolled per semester is positively correlated to cheating behaviours. This can be explained in the context of the COVID-19 related stress and the increased workload associated with more units. This finding is consistent with previous studies that found stress and increased workload to be predictors of academic misconduct in higher education (Amigud and Lancaster 2019; Tindall and Curtis 2020).

Our study revealed that young people (18-24) perceived high academic misconduct in alternative assessments than older people. This could be because these young people, considered Generation Z, rely heavily on the internet/technology and friends/peers for information and support respectively. Generation Z are fully aware of the different ways cheating can occur in online non-invigilated assessment than older generations (Hernandez-de-Menendez et al. 2020). Our finding is similar to that of Kisamore et al. (2007) and Landa-Blanco et al. (2020) where they reported that older people are less likely to perceive or commit academic misconduct than younger people. Our findings also suggest that many technologies such as mobile phone messaging and social media platforms were used to aide cheating in alternative assessments. These platforms are highly patronised by Generation Z and are frequently used for cheating in alternative assessments. In this regard, education on academic ethics and the repercussion of cheating will be very critical in a post COVID-19 academic integrity policy and praxis (Curtis et al. 2022; Morris 2018). Additionally, academic integrity roles should be well-resourced and supported within the bureaucracy of academic institutions (Eaton 2020; Seeland 2020). Furthermore, existing academic misconduct detection and management architecture will need to be reviewed if not overhauled as most of them appear to cater for face-to-face learning mode, aside from inconsistencies in their application (Amigud and Pell 2021).

Like age, there was a negative relationship between undergraduate academic year and cheating behaviours. Students in first year were more likely to be involved in cheating behaviours than students in second and third years. This could result from first year students’ inability to transition effectively into higher education before the drastic changes in assessments as a result of COVID-19. These students may not have had the full opportunity of academic integrity training compared to their senior colleagues due to the chaos created by COVID-19. With COVID-19 restrictions, academic institutions, the gate keepers of academic integrity, had to juggle between organisational adjustments and learning and delivery approaches. As a result, academic integrity measures were not the priority for universities who had to focus on building their online systems to meet the unprecedented demands for online teaching and learning (Slade and Benson 2020). This finding may be explained in the work of Bretag et al. (2014), in which over 80% of students responded that the first time they learnt about academic integrity was during their university orientation. It is recommended that innovative approaches to academic integrity workshops be implemented for first year students who are likely to be Generation Z to develop a culture of academic integrity.

Of the three academic misconduct behaviours explored in alternative assessments, our study showed that collusion is the commonest form of cheating with more than 31% of students confirming that they have sought help from peers. This behaviour is perceived to be common among on campus students with friends or classmates that they can collaborate through other means such as phone text messaging or social media platforms. As reported in other studies (Harris et al. 2020), this form of cheating is the commonest in project-related or assignment type assessments. Perhaps, the possible reason for this observation is students’ perception of collusion being a form of team work or supporting each other to succeed (Forkuor et al. 2019). In our study, although students were aware of the requirement to undertake the assessment individually, collusion remained prevalent. The emphasis on team and collaborative work in nursing education/profession (Barton et al. 2018), may have contributed to the high rate of collusion among nursing students. Ongoing education on the differences between collusion and unauthorised collaboration must be the focus in designing academic integrity interventions in courses where teamwork is highly regarded.

Furthermore, our study also found positive correlation between assessment time and risk of academic misconduct. The risk of academic misconduct increases by more than double when the assessment duration is 1-2 weeks compared to 2-4 hours. The longer the assessment duration, the more likely students are to collude or collaborate with others. It is recommended that assessment be given limited time duration—not more than 4 hours—to minimise this risk.

The qualitative data in the current study suggested ways in which risk of academic integrity can be minimised if not eliminated in alternative assessments. Aside from the promotion of ethical values among students, making assessments time-bound (less than 4 h) and use of plagiarism detection softwares were mentioned in the data and corroborated in other studies as useful measures for bolstering academic integrity in a non-invigilated assessment setting (Brown and Janssen 2017; Denisova-Schmidt, 2017). Other ideas touched on the need to design assessments that require application of theory to practice as opposed to “google-able” questions and assignments. This finding is consistent with the work of Golden and Kohlbeck (2020) in which they reported decreased rate of cheating when online question banks are paraphrased. Ongoing education on present and future consequence of academic misconduct should be a usual conversation between academics and students as this has proven to minimise the risk of academic misconduct in alternative assessment.

Limitations and future research

In self-reported questionnaires, there is the tendency for fake responses, typically presenting favourable images of phenomenon under investigation, referred to a social desirability bias. This bias can confound study results by creating false relationships or obscure true relationships (Van de Mortel 2008). In this study, we observed an average non-response rate of approximately 15% across the indicator scales for academic integrity. This indicates that students who were not comfortable with the indicator items opted not to answer. It can be inferred that the majority of students responded honestly to the questionnaire items, and thus the conclusions arrived at are meaningful and informative to instigate appropriate measures to safeguard academic integrity in alternative assessments.

Another limitation of the study is related to the sampling method—convenience sampling—which may be reflective of the nonresponsive bias and the inclusion of only nursing and social science students. Students who do not feel strongly for or against the alternative assessment did not feel the need to complete the survey (Wang & Cheng, 2020). This may result in a limitation to generalisability of the results for all students. Future research should explore the academic integrity concerns associated with alternative assessments from the perspective of both academics and all students from various courses to present a holistic picture.

Conclusion

This paper sought to explore the perceptions and experiences of academic integrity concerns of 380 tertiary students in higher education regarding online alternative assessments during the COVID-19 pandemic. The study employed a cross-sectional mixed method design. We conclude that alternative assessment is a viable option for assessing students if measures are implemented to reduce associated academic integrity risks. Like other forms of traditional assessments, alternative assessments have risk of academic integrity breach; however, with the right strategies, they could serve as effective means of assessing learning outcomes. In this study, although cheating was not found to be a generality among participants, some students reported consulting unauthorised sources and colluding with their peers. Addressing such vices in tertiary institutions would require continuous education of students on the importance of academic integrity as well as the costs of academic dishonesty to the university, the students, and their future. Other measures, such as designing time-bound assessments, non-googleable questions, use of plagiarism detection softwares and designing assessments that require application of theory to practice would go a long way to making academic dishonesty less popular with alternative and non-invigilated assessments. COVID-19 and its attendant issues with academic misconduct has brought to the fore the need for universities to resource its academic integrity gate keepers by training academics on developing assessments that can enhance academic integrity.

Availability of data and materials

Data is available upon request.

Abbreviations

COVID-19:

Coronavirus Disease-19

ID:

Identification card for identifying students

References

  • Adesile I, Nordin MS, Kazmi Y, Hussien S (2016) Validating academic integrity survey (AIS): an application of exploratory and confirmatory factor analytic procedures. J Acad Ethics 14(2):149–167

    Article  Google Scholar 

  • Adnan M, Anwar K (2020) Online learning amid the COVID-19 pandemic: Students' perspectives. Online Submission 2(1):45–51

    Google Scholar 

  • Akimov A, Malin M (2020) When old becomes new: a case study of oral examination as an online assessment tool. Assess Eval High Educ 45(8):1205–1221

    Article  Google Scholar 

  • Alruwais N, Wills G, Wald M (2018) Advantages and challenges of using e-assessment. Int J Inform Educ Technol 8(1):34–37

    Google Scholar 

  • Amigud A, Lancaster T (2019) 246 reasons to cheat: an analysis of students’ reasons for seeking to outsource academic work. Comput Educ 134:98–107

    Article  Google Scholar 

  • Amigud A, Pell DJ (2021) When academic integrity rules should not apply: a survey of academic staff. Assess Eval High Educ 46(6):928–942

  • Aristovnik A, Keržič D, Ravšelj D, Tomaževič N, Umek L (2020) Impacts of the COVID-19 pandemic on life of higher education students: a global perspective. Sustainability 12(20):8438

    Article  Google Scholar 

  • Barton G, Bruce A, Schreiber R (2018) Teaching nurses teamwork: integrative review of competency-based team training in nursing education. Nurse Educ Pract 32:129–137

    Article  Google Scholar 

  • Bloxham S, Boyd P (2007) Developing effective assessment in higher education: a practical guide. Open University Press, UK

  • Bowden A, Fox-Rushby J, Nyandieka L, Wanjau J (2002) Methods for pre-testing and piloting survey questions: illustrations from the KENQOL survey of health-related quality of life. Health Policy Plan 17(3):322–330

    Article  Google Scholar 

  • Bozkurt A, Jung I, Xiao J, Vladimirschi V, Schuwer R, Egorov G, Lambert S, Al-Freih M, Pete J, Olcott D Jr (2020) A global outlook to the interruption of education due to COVID-19 pandemic: navigating in a time of uncertainty and crisis. Asian J Dist Educ 15(1):1–126

    Google Scholar 

  • Bretag T, Mahmud S, Wallace M, Walker R, McGowan U, East J, Green M, Partridge L, James C (2014) ‘Teach us how to do it properly!‘an Australian academic integrity student survey. Stud High Educ 39(7):1150–1169

    Article  Google Scholar 

  • Brown N, Janssen R (2017) Preventing plagiarism and fostering academic integrity: a practical approach. J Perspect Appl Acad Pract 5(3):102–109

    Google Scholar 

  • Clutter LB, Potter WT, Alarbi A, Caruso JF (2017) Test anxiety and salivary cortisol levels in nursing students. Nurse Educ 42(1):28–32

    Article  Google Scholar 

  • Costley J (2019) Student perceptions of academic dishonesty at a Cyber-University in South Korea. J Acad Ethics 17(2):205–217

    Article  Google Scholar 

  • Creswell JW, Clark VLP (2017) Designing and conducting mixed methods research. Sage Publications

  • Creswell JW, Creswell JD (2017) Research design: qualitative, quantitative, and mixed methods approaches. Sage Publications

  • Curtis GJ, Slade C, Bretag T, McNeill M (2022) Developing and evaluating nationwide expert-delivered academic integrity workshops for the higher education sector in Australia. High Educ Res Dev 41(3):665–680

    Article  Google Scholar 

  • Denisova-Schmidt E (2017) The challenges of academic integrity in higher education: Current trends and prospects

  • Eaton SE (2020) Academic integrity during COVID-19: reflections from the University of Calgary

    Google Scholar 

  • Feng X, Behar-Horenstein L (2019) Maximizing NVivo utilities to analyze open-ended responses. Qual Rep 24(3):563–571

    Google Scholar 

  • Forkuor JB, Amarteifio J, Attoh DO, Buari MA (2019) Students’ perception of cheating and the best time to cheat during examinations. Urban Rev 51(3):424–443

    Article  Google Scholar 

  • Gamage KA, Silva EK, Gunawardhana N (2020) Online delivery and assessment during COVID-19: safeguarding academic integrity. Educ Sci 10(11):301

    Article  Google Scholar 

  • Golden J, Kohlbeck M (2020) Addressing cheating when using test bank questions in online classes. J Account Educ 52:100671

    Article  Google Scholar 

  • Guàrdia L, Crisp G, Alsina I (2017) Trends and challenges of e-assessment to enhance student learning in higher education. In: Innovative practices for higher education assessment and measurement. IGI Global, USA, pp 36–56

  • Harmon OR, Lambrinos J, Buffolino J (2010) Assessment design and cheating risk in online instruction. Online J Dist Learn Admin 13(3):23–33

    Google Scholar 

  • Harris L, Harrison D, McNally D, Ford C (2020) Academic integrity in an online culture: do McCabe’s findings hold true for online, adult learners? J Acad Ethics 18(4):419–434

    Article  Google Scholar 

  • Hernandez-de-Menendez M, Díaz CAE, Morales-Menendez R (2020) Educational experiences with generation Z. Int J Interact Design Manufact 14(3):847–859

    Article  Google Scholar 

  • Johnson JA (2018) Academic integrity in higher education: is collegiate aviation education at risk? Coll Aviat Rev Int 14(1):1–9

    Google Scholar 

  • Kisamore JL, Stone TH, Jawahar I (2007) Academic integrity: the relationship between individual and situational factors on misconduct contemplations. J Bus Ethics 75(4):381–394

    Article  Google Scholar 

  • Kocdar S, Karadeniz A, Peytcheva-Forsyth R, Stoeva V (2018) Cheating and plagiarism in e-assessment: students’ perspectives. Open Praxis 10(3):221–235

    Article  Google Scholar 

  • Landa-Blanco M, Santos-Midence C, Blanco AL (2020) Academic integrity: attitudes and practices of students of a public university in Honduras. Academia (20-21):202–217

  • Marinoni G, Van’t Land H, Jensen T (2020) The impact of Covid-19 on higher education around the world. IAU Global Surv Rep 23:1–7

    Google Scholar 

  • McCabe DL (2005) Cheating among college and university students: a North American perspective. Int J Educ Integr 1(1):1–11

    Article  Google Scholar 

  • McCabe DL, Trevino LK (1993) Academic dishonesty: honor codes and other contextual influences. J High Educ 64(5):522–538

    Google Scholar 

  • Merta Dhewa K, Rosidin U, Abdurrahman A, Suyatna A (2017) The development of higher order thinking skill (hots) instrument assessment in physics study. J Res Method Educ 7(1):26–32

    Google Scholar 

  • Morris EJ (2018) Academic integrity matters: five considerations for addressing contract cheating. Int J Educ Integr 14(1):1–12

    Article  Google Scholar 

  • Paudel P (2020) Online education: benefits, challenges and strategies during and after COVID-19 in higher education. Int J Stud Educ 3(2):70–85

    Article  Google Scholar 

  • Peterson J (2019) An analysis of academic dishonesty in online classes. Mid-Western Educ Res 31(1)

  • Ramdani Z (2018) Construction of academic integrity scale. Int J Res Stud Psychol 7(1):87–97

    Article  Google Scholar 

  • Rashid S, Yadav SS (2020) Impact of Covid-19 pandemic on higher education and research. Indian J Hum Dev 14(2):340–343

    Article  Google Scholar 

  • Reedy A, Pfitzner D, Rook L, Ellis L (2021) Responding to the COVID-19 emergency: student and academic staff perceptions of academic integrity in the transition to online exams at three Australian universities. Int J Educ Integr 17(1):1–32

    Article  Google Scholar 

  • Roksa J, Arum R, Cook A (2016) Defining and assessing learning in higher education. Improving quality in American higher education: learning outcomes and assessments for the 21st century, p 1

    Google Scholar 

  • Seeland J (2020) Reflection on academic integrity during COVID-19. Can Perspect Acad Integr 3(2):39–40

    Google Scholar 

  • Slade C, Benson K (2020) Reflections on academic integrity and academic development during COVID-19 at. The University of Queensland, Australia

    Google Scholar 

  • Tashakkori A, Johnson RB, Teddlie C (2020) Foundations of mixed methods research: integrating quantitative and qualitative approaches in the social and behavioral sciences. Sage Publications

  • Tindall IK, Curtis GJ (2020) Negative emotionality predicts attitudes toward plagiarism. J Acad Ethics 18(1):89–102

    Article  Google Scholar 

  • Toquero CM (2020) Challenges and opportunities for higher education amid the COVID-19 pandemic: the Philippine context. Pedagog Res 5(4)

  • Van de Mortel TF (2008) Faking it: social desirability response bias in self-report research. Austr J Adv Nurs 25(4):40

    Google Scholar 

  • Wang X, Cheng Z (2020) Cross-sectional studies: strengths, weaknesses, and recommendations. Chest 158(1):S65–71

  • World Health Organization. (2020). WHO Director-General’s opening remarks at the media briefing on COVID-19.

    Google Scholar 

Download references

Acknowledgements

We would like to acknowledge the contribution of Dr. Fiona Foxall, Associate Dean Teaching and Learning, School of Nursing and Midwifery for her valuable feedback on the initial survey questionnaire and Prof Clive Barstow, Associate Dean Teaching and Learning, School of Arts and Humanities, for his support.

Funding

No funding received.

Author information

Authors and Affiliations

Authors

Contributions

Esther Adama: Study conception and design, data collection, manuscript draft, critical review. Amanda Graf: Study conception and design, data collection, manuscript draft, critical review. Kwadwo Adusei-Asante: Study conception and design, data collection, critical review. Ebenezer Afrifa-Yamoah: Data analysis, manuscript draft, critical review. The author(s) read and approved the final manuscript.

Corresponding author

Correspondence to Esther Abena Adama.

Ethics declarations

Competing interests

No competing interest among all authors.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Supplementary Information

Additional file 1.

Survey for academic integrity.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Adama, E.A., Graf, A., Adusei-Asante, K. et al. COVID-19 and alternative assessments in higher education: implications for academic integrity among nursing and social science students. Int J Educ Integr 19, 8 (2023). https://doi.org/10.1007/s40979-023-00129-0

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1007/s40979-023-00129-0

Keywords