The promise is compelling: learning analytics can flag a struggling eighth-grader weeks before a failing grade appears, adapt coursework to individual pace, and give administrators a clearer picture of what actually drives student outcomes. Yet that same instrumentation, deployed without governance guardrails, can quietly erode student privacy, entrench demographic bias in predictive models, and transform classrooms into surveillance environments that undermine the trust on which effective teaching depends.
The gap between promise and peril is not theoretical. As UNESCO's guidance on data governance in education emphasizes, students must be the beneficiaries of analytics, not merely the subjects of them. Schools that fail to draw that distinction risk regulatory exposure, reputational harm, and, most importantly, real damage to the young people they serve.
This guide lays out a governance framework that allows schools to capture the benefits of learning analytics while managing the risks with the rigor they deserve.
What Are Learning Analytics?
The Jisc Code of Practice for Learning Analytics defines learning analytics as the measurement, collection, analysis, and reporting of data about learners for the purpose of understanding and optimizing learning. In practice, that definition spans a wide range of applications: at-risk identification systems that predict which students may struggle (Sclater, 2017), adaptive platforms that adjust content in real time based on performance, engagement trackers that monitor participation and completion rates, performance dashboards that visualize progress for teachers and families, and early warning systems that trigger alerts when leading indicators decline.
These applications draw on an equally broad set of data sources. Learning management systems contribute assignment completion records and grades. Student information systems supply demographics, attendance histories, and longitudinal performance data. Assessment platforms generate test scores and item-level response patterns. Behavioral telemetry captures login frequency, time on task, click sequences, and navigation paths.
The sheer volume and variety of this data is what makes governance both difficult and indispensable.
The Governance Challenge
Every worthwhile application of learning analytics carries a corresponding risk. Early identification of struggling students can shade into labeling those students in ways that become self-fulfilling prophecies. Personalized support interventions can metastasize into a surveillance culture that erodes trust between students and teachers. Data-informed instruction can embed algorithmic bias in the very predictions educators rely on to allocate resources. Optimization initiatives can experience purpose creep, drifting from educational goals toward administrative or commercial ones. And any system that aggregates granular student data creates a surface for privacy violations.
Governance exists to hold these tensions in productive balance, ensuring that analytics serve learning without sacrificing the rights and wellbeing of the learners themselves.
Governance Principles
Principle 1: Purpose Limitation
Every analytics use case should begin with a single question: does this directly serve student learning and wellbeing? If the answer requires qualification, the use case warrants scrutiny. Purpose limitation also demands data minimization. Schools should collect only the information necessary for a stated educational objective and should be prepared to explain that objective to any parent who asks.
Principle 2: Transparency
Students, parents, and teachers deserve a clear account of what data is collected, how it is used, and what predictions or scores are generated. Transparency is not satisfied by a dense privacy notice buried on a website. It requires accessible language, proactive communication, and dashboards that make analytics legible to non-technical stakeholders.
Principle 3: Human Oversight
Analytics should inform human judgment, never replace it. That means requiring teacher review before any predictive output triggers an intervention, ensuring that a human can override algorithmic recommendations at every decision point, and investing in training so that educators interpret analytics critically rather than deferentially.
Principle 4: Fairness
Predictive models must perform equitably across demographic groups. Schools should test for differential impact, investigate disparities when they appear, and accept that a model producing biased outcomes is worse than no model at all. Fairness is not a one-time validation exercise; it requires ongoing monitoring and annual bias audits.
Principle 5: Student Agency
Students should not merely be data points. They should have access to their own analytics, the ability to challenge inaccurate data or predictions, and the opportunity to develop data literacy. Schools that treat analytics as something done to students rather than for them will eventually face resistance from the communities they serve.
Governance Policy Template
The following template provides a starting framework that schools can adapt to their own contexts and regulatory environments.
Permitted Uses
Learning analytics may be used to identify students who may benefit from additional support, personalize learning experiences, inform instructional decisions, evaluate program effectiveness, and support student goal-setting and self-reflection.
Prohibited Uses
Learning analytics shall not be used to make automated decisions about student discipline, share identifiable student data for commercial purposes, create permanent student profiles that follow individuals across schools, make high-stakes decisions without human review, or evaluate teacher performance based solely on student analytics.
Data Minimization
Schools should collect only data necessary for specified educational purposes, review data collection practices annually to eliminate unnecessary collection, and retain analytics data only as long as the educational purpose requires.
Transparency Requirements
Schools should inform parents and students about learning analytics through their privacy notice, make dashboards available to students and parents where appropriate, and explain how predictive models work in language that a non-specialist can understand.
Human Oversight Requirements
Schools should require teacher review before acting on predictive analytics, train teachers on appropriate interpretation and use, and document all decisions materially influenced by analytics outputs.
Fairness Requirements
Schools should conduct bias audits on predictive models at least annually, monitor for differential impact across student groups, and investigate and address any identified disparities promptly.
Student and Parent Rights
Students and parents may access analytics data about them, request correction of inaccurate data, opt out of specific analytics programs with alternative arrangements provided, and challenge decisions influenced by analytics.
Oversight Structure
A designated person or committee should oversee compliance with the policy, conduct an annual review of all analytics programs against its provisions, and report significant concerns to school leadership.
Implementation Checklist
Assessment Phase
The first step is a comprehensive inventory. Schools should catalog every learning analytics system in use, document what data each system collects, identify how analytics outputs influence decisions, and assess alignment with the governance principles outlined above.
Policy Phase
With the inventory complete, schools should formally adopt a learning analytics governance policy, communicate it clearly to staff, parents, and students, and integrate it into the institution's broader data protection framework.
Operations Phase
Policy without practice is performative. Schools should train teachers on appropriate analytics use, establish human oversight procedures with clear escalation paths, create a process for handling parent and student data requests, and schedule annual bias audits with defined accountability for follow-through.
Monitoring Phase
Governance is not a one-time project. Schools should track analytics usage patterns on an ongoing basis, monitor for differential impact across student populations, collect structured feedback from teachers and families, and conduct an annual policy review to ensure the framework keeps pace with evolving technology and regulation.
Next Steps
The most effective starting point is the inventory itself. Schools that map their current analytics landscape, assess each use case against the governance principles above, and address the highest-risk gaps first will build a foundation that scales as their analytics capabilities mature.
Common Questions
Governance frameworks for collecting and using student learning data ethically, including purpose limitation, transparency to students and parents, human oversight of AI decisions, and data protection.
Limit data collection to educational purposes, be transparent with students and families, ensure human oversight of AI insights, protect data appropriately, and build in student agency.
Explain what data is collected, how it's used, what insights are generated, who has access, and how it affects the student. Provide access to their own data and ability to correct errors.
References
- Code of Practice for Learning Analytics. Jisc (2020). View source
- Learning Analytics Explained. Routledge (Sclater, N.) (2017). View source
- Artificial Intelligence in Education. UNESCO (2024). View source
- Guidance for Generative AI in Education and Research. UNESCO (2023). View source
- Youth Privacy — Education and Student Privacy. Future of Privacy Forum (2024). View source
- Advisory Guidelines on Use of Personal Data in AI Systems. PDPC Singapore (2024). View source
- AI and Education: Protecting the Rights of Learners. UNESCO (2024). View source

