Skip to content

glowkeeper/feedbacker

Repository files navigation

Feedbacker

standard-readme compliant

AI-assisted feedback that enables educators to feedback smarter, faster, better.

Live Deployment

Try Feedbacker here: https://feedbacker.education/

Feedbacker interface

Table of Contents

Overview

Feedbacker is designed for responsible AI-assisted feedback in higher education.

Feedback Smarter

  • Blends academic expertise with automation and AI
  • Keeps feedback human-focused, not machine-generated
  • Reduces repetitive work so the educators attention stays on pedagogy
  • Helps deliver feedback students can genuinely act on

Feedback Faster

  • Automates the slow, repetitive parts of marking
  • Helps produce personalised feedback in less time
  • Shortens turnaround without sacrificing quality
  • Frees up time for teaching, discussion, and student connection

Feedback Better

  • Makes it clear how students are performing
  • Highlights where improvement is needed
  • Points students towards actionable next steps
  • Combines insight with smart automation for feedback that is more impactful

Why Feedbacker Exists

Providing high-quality feedback is one of the most valuable — and time-intensive — parts of teaching. Increasing student numbers and administrative pressures often reduce the time educators can spend crafting meaningful responses.

Feedbacker was created to support educators by automating repetitive aspects of feedback while preserving academic judgement and pedagogical intent. The goal is not to replace educators, but to help them scale thoughtful, actionable feedback in a sustainable way.

Responsible Use of AI

Feedbacker is designed to support — not replace — academic judgement.

  • AI generates draft feedback only
  • Educators remain responsible for reviewing, editing, and approving all outputs
  • The system does not automate grading decisions
  • Feedback remains transparent and attributable to the educator

Feedbacker aligns with emerging UK higher-education guidance encouraging AI as an assistive tool rather than an autonomous assessor.

Intended Audience

Feedbacker is designed primarily for:

  • Higher education lecturers and tutors
  • Foundation and undergraduate teaching
  • Educators using rubric-based assessment
  • Institutions exploring responsible AI-assisted marking workflows

Especially Useful When

  • Experimenting with AI in teaching practice
  • Managing large cohorts with heavy feedback workloads
  • Curious about responsible AI adoption but cautious about automation
  • Looking for ways to prototype new feedback workflows

Use in Teaching Contexts

Feedbacker has been developed alongside real teaching practice and is intended to support:

  • Formative feedback workflows
  • Rubric-based assessment
  • Draft feedback generation
  • Feedback consistency across cohorts

It is particularly suited to large-cohort modules where maintaining feedback quality and turnaround time is challenging.

Limitations

Feedbacker assists feedback writing but does not:

  • Replace academic judgement
  • Verify factual correctness of feedback
  • Assess academic misconduct
  • Assign final grades automatically

Educators should always review outputs before sharing with students.

Data and Privacy

Feedbacker does not require institutional integration or student accounts to operate.

The application does not store student submissions unless explicitly configured to do so by a deploying institution. Institutions remain responsible as data controllers for any content entered into the system.

Users should avoid entering personally identifiable student information unless permitted under their institutional policies.

When self-hosted, institutions retain full control over API configuration and data handling practices.

Institutional Deployment

Feedbacker can be:

  • Used via the public deployment
  • Self-hosted by institutions
  • Configured with institution-approved AI providers
  • Adapted to local assessment policies

The application is intentionally lightweight to enable experimentation and pilot adoption.

Running a Pilot

Typical institutional pilots involve:

  • Use within a single module or cohort
  • Formative assessment contexts
  • Educator review and editing of all generated feedback
  • Optional participation by teaching staff exploring AI-assisted workflows

This approach allows institutions to evaluate pedagogical value while maintaining academic oversight and governance compliance.

Design Principles

Feedbacker is guided by the following principles:

  • Human-led feedback — AI assists, educators decide
  • Transparency — outputs remain editable and reviewable
  • Pedagogy first — educational value takes priority over automation
  • Efficiency without dehumanisation — save time without losing voice
  • Institutional compatibility — works alongside existing assessment practices

The interface is designed to work with standard accessibility tooling and screen readers where possible.

Quick Start

  1. Visit the live deployment: https://feedbacker.education/
  2. Enter assessment criteria or rubric information
  3. Provide student work or summary notes
  4. Generate draft feedback and refine as needed

Trying Feedbacker Safely

Feedbacker can be explored without changing existing assessment workflows.

A typical first use is to generate draft feedback for a small number of assignments while continuing normal marking practices. Educators remain fully in control and can adopt or discard outputs as they see fit.

Many users begin by experimenting with anonymised or sample work before using the tool in live teaching.

Technology Stack

Install

This project uses node and pnpm. If you've not done so, please install those first. Then clone this repository, switch to its root directory, and type pnpm install.

Local Development

You will need to create a .env file in the root directory, with the following four variables:

  1. NEXT_PUBLIC_OPENROUTER_URL
  2. NEXT_PUBLIC_OPENROUTER_KEY
  3. NEXT_PUBLIC_OPENROUTER_MODEL
  4. NEXT_PUBLIC_TITLE
  5. NEXT_PUBLIC_HOMEPAGE

Where NEXT_PUBLIC_OPENROUTER_URL is the URL of the OpenRouter completions API, NEXT_PUBLIC_OPENROUTER_KEY is the API key and NEXT_PUBLIC_OPENROUTER_MODEL is the AI model to use. NEXT_PUBLIC_TITLE and NEXT_PUBLIC_HOMEPAGE help define the app's title and public URL.

Once you've defined those variables, you can run a local development server via pnpm dev.

Project Status

Active development.

Feedbacker is currently used in real teaching contexts and continues to evolve based on educator feedback.

Educational Context and Collaboration

Feedbacker is developed as part of ongoing exploration into responsible uses of AI in higher education assessment and feedback practices.

The project welcomes discussion, experimentation, and collaboration with educators, learning technologists, and institutions interested in piloting or studying AI-assisted feedback workflows.

If you are exploring similar questions in teaching practice or institutional innovation, the maintainer is happy to hear from you.

Maintainer

Steve Huckle.

Contributing and Feedback

Feedbacker is evolving through real teaching use.

Bug reports, teaching experiences, pilot stories, and suggestions are all welcome — especially from educators trying the tool in new contexts.

Please contact the maintainer by email.

Citation

If you use Feedbacker in research or teaching practice, please cite:

Huckle, S. (2025). Feedbacker: AI-assisted feedback for higher education. https://feedbacker.education/

License

Creative Commons Attribution 4.0 International Deed (CC BY 4.0)

CC BY 4.0

About

AI feedback tool

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Sponsor this project

 

Packages

No packages published