Published Feb 24, 2026 · Updated Feb 24, 2026
On 27 January 2026, the Quality Assurance Agency for Higher Education (QAA) published the report of a QAA targeted peer review at the University of Glasgow, commissioned by the Scottish Funding Council under the Scottish Quality Concerns Scheme. [QAA announcement] For universities that collect and act on student feedback, this is a reminder that student voice evidence is often tested in operational terms: whether assessment processes are clear, whether students can navigate extensions and complaints, and whether changes are communicated in a way students trust.
QAA’s announcement states that the review identified systemic risks in relation to quality and standards at the University of Glasgow, with a particular focus on assessment regulations and student communications. The report sets out 21 recommendations. QAA also notes the university is required to submit an action plan within four weeks.
The report says the peer review was commissioned by the Scottish Funding Council after a quality concern was raised in 2025, following an independent investigation by King’s Counsel. The peer review itself focused on systemic risks, and on processes that affect the student learning experience at scale, including: assessment regulations and the award of credit, extension request processes, student communications, risk mitigation and oversight, and student engagement in institutional change.
The announcement also signals a wider sector implication for Scotland:
"commissioning QAA to conduct a national review of the assessment and associated policies and procedures across the sector."
For student experience teams, the theme linking the recommendations is practical. Students need coherent programme and assessment information, a predictable extension process, and clear routes to raise concerns. These are the exact areas where open-text survey comments, complaints themes, and staff to student forum issues often surface problems first.
First, it is a prompt to treat assessment policy and process as part of the student feedback system, not just the academic rulebook. When policies are unclear, applied inconsistently, or hard to navigate, students do not just report dissatisfaction, they lose confidence that raising issues leads to fair outcomes. That makes it harder to sustain engagement in module evaluation and other feedback channels.
Second, this is a reminder to measure student experience at the level of process. A clear improvement plan usually needs more than a policy document. It needs evidence that students can find accurate information quickly, that they can submit and track requests (extensions, deferrals, special circumstances), and that complaints are handled with service standards. The report explicitly recommends using complaint data to identify issues and set expectations, which is a useful pattern for the wider UK.
Third, if you are strengthening your approach to quality assurance, it is worth checking whether your student voice evidence is traceable. Can you show what students said, how you analysed it, what changed, who owned it, and when students were told? This is the same evidence chain that comes up in OfS quality assessment contexts, as in our recent post on OfS quality assessments and student feedback evidence.
At Student Voice AI, we often see assessment and communication problems hidden in plain sight in qualitative comments. Students describe uncertainty about regulations, frustration with extension processes, and confusion when assessment changes are not communicated clearly. Analysing open-text at scale helps teams quantify how widespread a problem is, identify which cohorts are most affected, and pull representative examples for action planning.
If you are tightening governance around these topics, two practical starting points are our student comment analysis governance checklist and NSS open-text analysis methodology. For teams using sentiment views in reporting, our guide to sentiment analysis for UK universities sets out practical interpretation rules.
Q: What should institutions do now in response to this QAA targeted peer review?
A: Treat it as a checklist for your own assessment related feedback loop. Confirm students have a single source of truth for assessment information, test the extension journey end-to-end (including tracking and response times), and review complaints themes alongside module evaluation open text. If you make changes, publish an action log and close the loop with students.
Q: Does this apply UK-wide, and what is the timeline?
A: The targeted peer review relates to the University of Glasgow and the Scottish Quality Concerns Scheme, so the immediate scope is Scotland. QAA’s announcement (published 27 January 2026) also notes a Scottish Funding Council commissioned national review of assessment policies and procedures across the Scottish sector.
Q: What does this signal about how student voice evidence is used in quality work?
A: It reinforces a direction of travel towards operational evidence. Student feedback becomes more powerful when it is timely, analysed consistently, linked to action, and supported by clear communications that show students what changed and why.
[Quality Assurance Agency for Higher Education (QAA)]: "QAA review finds systemic risks to quality and standards at University of Glasgow"
Published: 2026-01-27
[Quality Assurance Agency for Higher Education (QAA)]: "Scottish Quality Concerns Scheme, Targeted Peer Review, University of Glasgow (Review Report)"
Published: 2026-01-27
Source URL: https://www.qaa.ac.uk/news-events/news/qaa-review-finds-systemic-risks-to-quality-and-standards-at-university-of-glasgow
Request a walkthrough
See all-comment coverage, sector benchmarks, and reporting designed for OfS quality and NSS requirements.
UK-hosted · No public LLM APIs · Same-day turnaround
Research, regulation, and insight on student voice. Every Friday.
© Student Voice Systems Limited, All rights reserved.