Student Voice as Partnership, Not Extraction

Updated Apr 03, 2026

If students only appear in a feedback process when it is time to fill in a survey, student voice has already been reduced to data collection. A recent paper in Active Learning in Higher Education by Brown and Kimber (2026) argues universities should treat student voice in higher education as agency and partnership within scholarship of teaching and learning (SoTL) inquiry, not as a one-way extraction exercise.

At Student Voice AI, we help universities make sense of student feedback at scale, especially free-text comments. That work matters most when analysis supports dialogue, co-design, and visible action, not just faster reporting.

Context and research question

SoTL has grown as a way for higher education to improve learning and teaching through evidence, reflection, and research-informed practice. Student feedback is often central to that work, but Brown and Kimber note a recurring pattern: students are asked for perspectives, then excluded from the processes that turn those perspectives into knowledge, decisions, and change. That weakens both trust and the value of the evidence institutions rely on, echoing what students report about teaching evaluations and visible follow-through.

The paper asks a straightforward question with major implications for institutions: what would it mean to expand SoTL inquiry so that students are treated as contributors and domain experts, not passive sources of data? The authors describe this as an "epistemic expansion" of SoTL. In practice, that changes who helps shape the research questions, methods, interpretation, and actions that follow.

For UK higher education teams, the relevance is immediate. The sector is rich in feedback mechanisms (module evaluations, NSS, PTES, PRES, UKES, complaints, rep systems), but many students still experience a gap between giving feedback and seeing change. If student voice is meant to strengthen quality enhancement rather than simply measure it, agency and partnership have to be designed into the process from the start.

Key findings

The paper offers institutions a clear test: are you learning with students, or only extracting answers from them?

First, Brown and Kimber argue that many higher education studies that solicit student feedback are still extractive, even when the intention is improvement. Students are positioned as respondents rather than co-creators of knowledge. That limits both the legitimacy and the usefulness of what institutions learn.

As they put it:

"many are extractive in that they position students as passive sources of data, rather than domain experts"

Second, the authors propose that the solution is not "more feedback". It is changing the epistemic role students play in inquiry, so their expertise shapes research design and scholarship praxis. In practice, that means involving students in framing problems, choosing what evidence to collect, and interpreting what results mean in context. That shift makes the evidence more useful because student experience shapes the inquiry earlier, not just the response at the end.

Third, the paper highlights approaches that are already moving in this direction. Brown and Kimber identify early adopters who place students at the heart of scholarship through co-design, and those who frame students explicitly as partners (often described as Students as Partners, SaP). The benefit is not symbolic. The authors argue this shift supports more authentic, responsive learning and teaching practice because it centres lived experience while improving how evidence is produced and used.

Finally, the paper challenges a common institutional reflex: treating feedback as something to "process" and "report" rather than something to build with. If student voice is supposed to lead to better decisions, then sense-making and action need shared ownership, not just better dashboards. That is the difference between a feedback system that records opinion and one that helps improve practice.

Practical implications

For UK universities using student feedback to improve the student experience, the paper points towards five practical design moves that make partnership more concrete:

  • Map where student voice becomes extractive. Identify where students provide input but have no visibility or influence over interpretation, prioritisation, or action, then fix those gaps first.
  • Co-design what you ask, not just how you analyse. Work with student reps and underrepresented groups to refine evaluation questions and open-text prompts so they surface what students actually want changed, a principle that also matters in student voice in curriculum design.
  • Invite students into interpretation, especially for open text. When you run thematic analysis or text analytics, create structured "sense-making" sessions where students can challenge, contextualise, and refine what themes mean.
  • Make action-tracking a shared artefact. Treat "you said, we did" updates as a working partnership document, not a marketing output. Be specific about what changed, where it changed, and what remains constrained.
  • Use analytics to scale listening, then reinvest the saved time in partnership. Student Voice Analytics can categorise and benchmark comment themes quickly, freeing teams to spend more time on co-design, dialogue, and follow-through.

The paper is a useful reminder that better student voice work is not mainly about collecting more feedback. It is about deciding whose knowledge counts, who helps interpret the evidence, and how that evidence turns into visible action. Institutions that design for partnership will get richer insight and a more credible basis for change.

FAQ

Q: How can universities reduce extractive feedback practices without abandoning sector surveys like the NSS?

A: Keep the surveys, but redesign the process around them. Co-design a small number of local open-text questions with students, share the analysis back in accessible formats, and involve students in prioritising actions. The NSS can remain a benchmark, but partnership is built through what happens before and after the survey: expectations, interpretation, and visible change.

Q: What does "Students as Partners" look like in a student feedback analytics project?

A: It means students help shape the work at multiple points: what questions are asked, what themes matter, how categories are named, how dashboards are interpreted, and what actions follow. It also requires practical commitments such as paying students for their time, widening participation beyond the usual representatives, and being transparent about limits, for example when professional requirements or funding constrain change. That wider partnership model only works when institutions close the loop in student voice initiatives and show students what changed.

Q: What changes if we treat student voice as agency rather than a dataset?

A: You interpret patterns differently. Instead of treating a theme trend as a verdict, you treat it as a prompt for dialogue: what conditions produced it, who is represented, and what improvement would look like to students. That mindset leads naturally to better use of open-text comments because it encourages institutions to treat qualitative feedback as shared meaning-making, not as a box to be ticked.

References

[Paper Source]: Alice Brown, Megan Kimber "Repositioning Student Voice and Agency: A Call for the Epistemic Expansion of Scholarship of Teaching and Learning Inquiry" DOI: 10.1177/14697874261426374

Request a walkthrough

Book a free Student Voice Analytics demo

See all-comment coverage, sector benchmarks, and reporting designed for OfS quality and NSS requirements.

  • All-comment coverage with HE-tuned taxonomy and sentiment.
  • Versioned outputs with TEF-ready reporting.
  • Benchmarks and BI-ready exports for boards and Senate.
Prefer email? info@studentvoice.ai

UK-hosted · No public LLM APIs · Same-day turnaround

Related Entries

The Student Voice Weekly

Research, regulation, and insight on student voice. Every Friday.

© Student Voice Systems Limited, All rights reserved.