Nine out of 10 UK undergraduates who responded to a survey have admitted using generative artificial intelligence in assessments, suggesting use of tools such as ChatGPT has become routine among students.
In a poll of 1,041 undergraduates for the Higher Education Policy Institute, 88 per cent of respondents said they had used generative AI in assessments over the past year – up from 53 per cent in 2024.
Report author Josh Freeman said that this “extraordinary growth” indicated that AI was becoming “an integral aspect in the process of doing exams”.
One in four students said that they had used AI-generated text to help them draft assessments, but students more commonly used ChatGPT and similar tools to explain concepts, summarise relevant articles or suggest ideas.
Students using AI in assessments were not necessarily breaking the rules, with 59 per cent agreeing that their universities had changed how they conducted assessments in response to the rise of generative AI and three-quarters confident that their institution could spot AI use in assessed work.
However, only three in 10 agreed that their institution encouraged them to use AI, with only four in 10 agreeing that AI-generated content would get a good mark.
Students were most likely to say that they used AI tools to save time or improve the quality of their work, and were most likely to say that they would be put off from doing so by the fear of being accused of cheating, or the risk of getting false results – so-called “hallucinations”.
The report recommends that universities should “continually review” assessments to keep up with the growing power of AI tools.
“Exams which are easy for someone to achieve a high grade in if they have good AI skills, but without engaging deeply with the course, should be immediately rewritten,” writes Freeman, Hepi’s policy manager.
“Closed-book examinations where the questions are predictable (and therefore can be prepared in advance using AI tools) are not immune.”
The report further recommends that universities should “adopt a nuanced policy which reflects the fact that student use of AI is inevitable and often beneficial”, and that staff involved in setting exams “should have a deep working understanding of AI tools”.
“If their assessments are robust, institutions need not fear educating students on the responsible use of AI, where it can genuinely aid learning and productivity,” the report says.
The survey found that the proportion of students reporting using an AI tool for any task had risen from 66 per cent to 92 per cent over the course of a year, and Freeman told Times Higher Education that he expected usage to be near-universal by next year.
However, while 67 per cent of students agreed that it was essential nowadays to understand how to use AI effectively, only 36 per cent said that they had received support from their university with their AI skills.
Forty-two per cent said staff were well-equipped to support them to work with generative AI, although this was a significant increase from 18 per cent last year.
Students told Hepi they felt that their university provided them with “mixed messages” over AI, with one student complaining: “It’s still all very vague and up in the air if [and] when it can be used and why. It seems to be discouraged without the recognition that it will form an integral part of our working lives.”
Other students were more sceptical of AI use, and one said “it feels like cheating”, while others said “it will completely ruin what actual work will look like” and “I feel like AI is good but just gives me the answers”.
The report also described how a “digital divide” identified in 2024 appears to have widened, with those from wealthier backgrounds more likely to use AI tools.