77
SCIENCE GIVING FEEDBACK ON STUDENT WRITING IN THE SCIENCES: MAKING IT WORK REPORT AND HANDBOOK OF GOOD PRACTICE LORRAINE RYAN, LEARNING CENTRE, THE UNIVERSITY OF SYDNEY

GIVING FEEDBACK ON STUDENT WRITING IN THE SCIENCES…sydney.edu.au/stuserv/documents/learning_centre/Feedback_Handboo… · science giving feedback on student writing in the sciences:

  • Upload
    lamdang

  • View
    215

  • Download
    2

Embed Size (px)

Citation preview

SCIENCE

GIVING FEEDBACK ON STUDENT WRITING IN THE SCIENCES: MAKING IT WORK

REPORT AND HANDBOOK OF GOOD PRACTICELORRAINE RYAN, LEARNING CENTRE, THE UNIVERSITY OF SYDNEY

This document is an outcome of the project funded by the University of Sydney Teaching Improvement and Equipment Scheme:

Feedback in the Sciences: what is wanted, what is given, and how it can be improved amongst a diverse student population

Project Team:

Leader: Meloni Muir, Discipline of Physiology

Team members: Miriam Frommer, Discipline of Physiology; Fiona White, School of Psychology; Vanessa Gysbers, School of Molecular Bioscience; Lorraine Ryan, Helen Drury, Learning Centre.

Research Assistants: Isabel Arnaiz, Andrew Howe, Leonardo Gabales

FEEDBACK: MAKING IT WORK

– It wastes my time to write the same comment on almost every student’s work.

– I don’t think I know enough about giving feedback.

– It would be good to have more training.

– It’s a challenge to give clear, concise and constructive feedback.

– I try not to just correct students’ work but then it’s difficult and takes time to explain why something is wrong.

– Should I give general or specific feedback or both?

WHAT MARKERS SAY ABOUT GIVING FEEDBACK

– Should I give feedback on both language and content?

– If a student is from a nonEnglish speaking background, should I consider this when I’m marking and giving feedback?

– If I understand the content, isn’t that enough?

– How can students use my feedback if they only get the assignment back at the end of semester?

– I wonder if students read and learn from my comments …

– Do students understand my feedback?

– There isn’t enough time to give students individual written or oral feedback

– I’m a casual staff member, so I don’t have enough paid time to give the feedback students want.

SCIENCE

Produced by the University of Sydney, 2013. The University reserves the right to make alterations to any information contained within this publication without notice. ABN 15 211 513 464 CRICOS 00026A

FOR MORE INFORMATION CONTACT

Meloni M. MuirE [email protected]/science

REDUCING FEEDBACK TIME

– Discuss assignment guidelines and criteria with students before the assignment

– Hold a report/essay writing tutorial for students when they are working on their assignment

– Provide exemplars of similar past assignments for students

– Use audio-files instead of written feedback

– Use feedback statement banks

EXPRESSING FEEDBACK – When possible, start feedback with

a positive statement

– Be constructive: note mistakes, but also explain how to improve

– Focus individual feedback on major points, e.g. on two/three themes/concepts

– Give timely, generic feedback on an assignment to the whole class asap after submission

– Markers provide general feedback to coordinator to compile and put online

WHAT COORDINATORS AND MARKERS CAN DO TO IMPROVE FEEDBACKCoordinators helping markers:

– Align assessment objectives with marking criteria and course outcomes

– Involve markers in assessment design and marking criteria development*

– Hold a marker session prior to marking to discuss expectations, feedback provision and grade criteria (e.g. P and HD)*

– Organise and manage a blog for markers’ questions

– Hold meetings to standardise marks before assignments returned*

– Hold post-marking meeting re problems encountered with marking scheme/rubric/criteria etc.*

* Note funding implications for sessional staff

GIVING FEEDBACK ON LANGUAGE

– Ensure that students know that English expression forms part of the assessment mark

– Coordinators and markers agree on the amount and type of literacy feedback, e.g. choose two problem areas and give one example of better phrasing for each

– Coordinators and markers agree on how non-English speaking background students’ work will be marked

HELPING STUDENTS TO USE FEEDBACK

– Set a small, low stakes assessment early in semester so that students can build on feedback in later assignments

– Approve a draft essay/report plan and/or give feedback on a draft and minimal comments on final assignment

– Give students experience as markers by holding an assessment tutorial using exemplars, then students give peer feedback before submission

– Students self-assess their work on submission

– Return assignments with feedback only; students rate themselves, with reasons, before mark given

– Students explain how they have used prior feedback in current assignment

  1  

Table  of  Contents  Introduction:  Using  the  Feedback  Handbook  .................................................................  3  

List  of  Tables  .................................................................................................................  2  

Part  A:   Perspectives  on  Feedback  ..............................................................................  5  1.0     The  Feedback  Project:  Student  Perspectives  ................................................................  5  

1.1   Project  Overview  ............................................................................................................  5  1.2   What  the  students  say/want  ..........................................................................................  5  

1.2.1   Summary  of  outcomes  from  student  questionnaire  .............................................................  5  1.2.2.   Student  Focus  groups  ...........................................................................................................  8  

2.0   The  Feedback  Project:  Markers’  Perspectives  .............................................................  13  2.1   Marker  Interviews  .........................................................................................................  13  

2.1.1   Problems  with  written  feedback  .........................................................................................  13  2.1.2   Marker  comments  re  dealing  with  diverse  language  backgrounds  of  students  .................  13  

2.2   Marker  workshops  ........................................................................................................  14  2.2.1.  Current  practices  and  problems  .............................................................................................  14  

3.0   The  Feedback  Project:  Suggested  Solutions  to  Feedback  Issues  and  Challenges  ..........  16  3.1   Coordinator-­‐centred  issues  and  suggestions  ................................................................  16  3.2   Marker-­‐centred  issues  and  suggestions  .......................................................................  16  

3.2.1   Solutions  suggested  in  Marker  Interviews  ..........................................................................  16  3.2.2   Solutions  suggested  by  Workshop  participants  ..................................................................  17  

3.3   Suggestions  from  the  literature  to  bring  students  into  ‘the  community  of  assessment  practice’  ..................................................................................................................................  18  

Part  B:   What  the  literature  says:  ‘Suggested  Solutions’  Themed  Annotated  Bibliography  (non-­‐exhaustive)  .....................................................................................  20  

4.1:     From  the  literature:  Overview  of  Feedback  ................................................................  23  4.1.1   General  concerns  ................................................................................................................  23  4.1.2   Purposes  of  feedback  ..........................................................................................................  24  4.1.3   Problems  with  the  transmission  view  of  feedback  ..............................................................  25  4.1.4   Student  understanding  of  feedback  comments  ..................................................................  25  4.1.5   Student  engagement  with  feedback  ...................................................................................  26  4.1.6   Student  use  of  feedback  ......................................................................................................  26  4.1.7   Student  perceptions  of  tutors’  written  feedback  ................................................................  27  4.1.8   Effectiveness  of  feedback  ....................................................................................................  27  4.1.9   Importance  of  ‘tacit  knowledge’  .........................................................................................  29  4.1.10    University  Feedback  Websites  (see  others  under  ‘Using  Technology  to  provide  peer  and  marker  feedback’,  below)  ................................................................................................................  29  4.1.11    University  of  Sydney  Feedback  webpages  .......................................................................  30  

4.2   Suggested  Solutions  from  the  Literature  ......................................................................  31  4.2.1   Feedback  practice  ...............................................................................................................  41  4.2.2   Examples  of  feedback  responses  markers  can  use  .............................................................  43  4.2.3   Student  engagement  with  feedback  ...................................................................................  45  4.2.4   Dialogic  approaches  to  feedback  ........................................................................................  49  4.2.5   Using  technology  to  provide  peer  and  marker  feedback  ....................................................  50  4.2.6   Dealing  with  large  classes  ....................................................................................................  54  4.2.7   Feedforward  ........................................................................................................................  55  

5.0:   Glossary  of  key  terms  –  definitions  from  the  literature  .............................................  57  References  ..........................................................................................................................  59  

     

  2  

List  of  Tables    Table  1:  From  the  literature:  Overview  of  Feedback  –  Guide…………………………  20  Table  2:  From  the  literature:  Suggested  Solutions  –  Guide……………………………..32  Table  3:  Suggested  Solutions  –  Feedback  practice………………………………………….36  Table  4:  Suggested  Solutions  –    Student  engagement  with  feedback………………37  Table  5:  Suggested  Solutions  –  Using  technology  to  provide  peer  and  marker  feedback…………………………………………………………………………………………………………39  Table  6:  Suggested  Solutions  –  Dealing  with  large  classes………………………………40      

  3  

 

Introduction:  Using  the  Feedback  Handbook    In  university  Science  teaching  contexts  in  Australia,  markers  are  under  pressure  to  provide  timely,  quality  feedback  to  large  and  diverse  student  cohorts.  We  designed  this  Handbook  to  help  markers  in  their  provision  of  written  feedback  on  student  writing.  Of  course,  there  are  no  ‘easy  answers’  to  all  the  situations  and  questions/issues  related  to  the  giving  and  receiving  of  written  feedback,  so  all  suggestions  need  to  be  considered  in  the  light  of  your  own  teaching  context.  You  can  refer  to  the  Handbook  for  tips  and  practical  suggestions.  Lecturers  and  unit  co-­‐ordinators  responsible  for  designing  curriculum  and  assessment  can  also  use  it  in  staff  training  sessions,  particularly  those  that  address  assessment,  marking  and  giving  feedback.  The  Feedback  Flyer  (you  will  find  this  at  the  beginning  of  the  Handbook)  would  also  be  useful  for  these  sessions  and  for  discussions  about  using  feedback  with  students.    The  main  focus  of  the  Handbook  is  on  the  practical  aspects  of  giving  feedback.  We  know  that  markers  are  often  sessional  staff,  so  we  wanted  to  provide  information  in  an  easily  accessible  form  which  we  think  can  improve  practice  with  little  or  no  extra  time  commitment.    The  Handbook  has  two  parts.  Part  A  reports  on  a  project  carried  out  with  a  large  cohort  of  Science  students  and  their  markers  at  the  University  of  Sydney.  There  are  three  sections  to  Part  A.  

• In  Section  1,  after  a  brief  overview  of  the  project,  you  can  read  short  summaries  of  students’  perspectives  on  feedback.  Since  the  University  cohort  is  multicultural  and  multilingual,  we  wanted  to  find  out  if  students’  perspectives  varied  according  to  their  facility  with  English  and  other  languages.  So  some  information  is  provided  from  students  who  are  native  English  speakers  only  (monolingual)  or  fluent  in  English  and  another  language  (bilingual),  or  whose  language  competence  in  English  is  reasonable  but  still  developing  (non-­‐English  speaking  background  or  NESB  students).  

 • In  Section  2,  you  will  find  summaries  of  markers’  experiences  and  their  

problems  with  giving  feedback.    

• Section  3  presents  markers’  suggestions  for  improving  feedback  or  solutions  to  the  problems  they  raised  in  Section  2.  

 In  Part  B,  you  will  find  a  selection  from  the  literature  on  feedback.  As  you  can  imagine,  there  has  been  a  huge  amount  published  in  this  area  so  we  have  limited  our  selection  to  those  publications  with  practical  advice,  generally  published  after  2000.      

  4  

 Part  B  is  divided  into  2  sections:  

 • The  Annotated  Bibliography  in  Section  4  provides  general  information  on  

feedback,  and  then  focuses  on  suggested  solutions  for  addressing  some  of  the  concerns  or  problems  with  giving  effective  feedback.  To  help  you  to  find  what  you  want  in  the  literature,  there  is  a  summary  table  of  the  relevant  publications  at  the  beginning  of  each  sub-­‐section.  If  you  want  more  detail,  you  can  find  it  after  each  table.  

• If  you  need  to  check  the  meaning  of  any  of  the  terms  used  in  the  area  of  feedback,  there  is  a  Glossary  in  Section  5  and  also  a  full  bibliography  at  the  end  of  the  Handbook.  

 If  you  would  like  to  give  us  feedback  on  this  Feedback  Handbook,  please  email  [email protected]        

  5  

 

Part  A:   Perspectives  on  Feedback    

1.0     The  Feedback  Project:  Student  Perspectives    Section  1  contains  a  summary  of  the  outcomes  of  the  project  entitled  :  Feedback  in  the  Sciences:  what  is  wanted,  what  is  given,  and  how  it  can  be  improved  amongst  a  diverse  student  population.    

1.1 Project Overview

The  project  involved  the  administration  of  a  questionnaire  to  second  year  University  of  Sydney  students  in  the  Faculty  of  Science  (Physiology,  Medical  Science,  Psychology  and  Molecular  Biology)  to  ascertain  the  types  of  written  feedback  they  found  useful.  The  questionnaire,  ‘Your  experience  of  receiving  feedback  on  your  writing’  was  completed  by  419  students;  their  mean  age  was  20.2  years,  and  62.6%  were  female.  Of  the  students  who  responded  to  the  question  concerning  language  spoken  at  home,  77  %  named  English,  14%  Chinese,  and  9%,  ‘Other’.    The  questionnaire  was  followed  by  six  student  focus  group  sessions,  involving  native  speakers  of  English,  both  monolingual  and  bilingual,  and  non-­‐native  speakers  (non-­‐English  speaking  background:  NESB):  these  sessions  were  designed,  in  part,  to  gain  deeper  insight  into  the  impact  of  diversity  on  student  perspectives  regarding  feedback.  Then  a  number  of  semi-­‐structured  marker  interviews  were  held  in  order  to  identify  current  feedback  practices.  After  this,  a  series  of  marker  workshops  was  held:  these  workshops  had  two  purposes.  The  first  was  to  inform  markers  of  student  perspectives  and  of  the  perspectives  reflected  in  the  current  literature  regarding  best  practice  for  providing  effective  feedback.  The  second  purpose  was  to  assist  markers  in  identifying  and  implementing  changes  to  their  own  feedback  practices  to  improve  student  learning.  The  following  sections  summarise  some  of  the  findings  from  each  of  these  aspects  of  the  project.    

1.2 What the students say/want

1.2.1   Summary  of  outcomes  from  student  questionnaire    The  questionnaire  included  four  general  questions,  each  with  11  to  15  related  statements  which  the  students  were  asked  to  rate  from  1  to  4  as  ‘strongly  disagree’  to  ‘strongly  agree’,  or  ‘not  important  at  all’  to  ‘very  important’  in  the  case  of  the  last  of  these  questions.  They  were  asked  to  circle  the  number  that  most  closely  reflected  their  ‘experience  in  second  year’  for  the  first  question,  their  ‘University  experience’  for  the  second  and  third,  and  their  ‘opinion  regarding  the  importance  of  different  types  of  written  feedback’  for  the  last.  There  were  also  five  open-­‐ended  questions  asking  students  for  their  perceptions  of  feedback.  These  included  ‘What  kind  of  feedback  is  most  helpful  to  you?’  and  ‘Do  you  use  feedback  on  written  assignments  when  writing  future  assignments?  If  yes,  how?  If  no,  why  not?’    

  6  

Responses  to  Question  1:  How  do  you  experience  feedback?  • Most  participants  indicated  that  the  written  feedback  they  received  was  

related  to  the  assessment  criteria,  course  objectives  and  the  marks  given,  but  was  too  brief.  

• Most  students  considered  negative  feedback  to  be  constructive.  • Most  students  reported  receiving  written  but  not  verbal  feedback  on  their  

work.  • The  students  who  did  receive  verbal  feedback  were  almost  equally  divided  on  

whether  they  remembered  it  or  not.    Responses  to  Question  2:  What  do  you  do  with  feedback?  

• The  majority  reported  carefully  reading  and  using  feedback  to  go  over  the  current  assignment  or  revise  work,  as  well  as  to  improve  future  assignments.  

• More  than  two-­‐thirds  of  the  students  used  the  feedback  even  if  a  high  grade  was  achieved.  

• The  students  did  not  ignore  negative  feedback  or  report  negative  reactions,  such  as  feeling  demoralised  or  angry,  in  response  to  it.  

• Positive  feedback  was  reported  to  boost  confidence.  • Students  reported  that  receiving  only  a  mark  was  unhelpful.  • The  majority  of  students  found  markers  to  be  consistent  in  their  application  

of  assessment  criteria;  however,  there  was  sizeable  minority  who  did  not  find  markers  consistent.  

 Responses  to  Question  3:  How  useful  do  you  find  feedback?  

• Students  found  feedback  helpful  for  improving  their  work,  identifying  gaps  in  their  knowledge  and  understanding  the  course  content.  

• Students  indicated  that  feedback  encouraged  reflection  on  what  they  had  learned,  prompted  revision  and  encouraged  them  to  improve.  

• As  regards  feedback  given  to  the  whole  class,  however,  opinion  was  divided,  with  nearly  equivalent  numbers  finding  this  helpful  and  unhelpful.  

 Responses  to    Question  4:  What  type  of  feedback  is  important?  

• Almost  all  types  were  useful  except  that  which  provided  general  comments.  • There  was  some  division  among  students  regarding  feedback  on  mistakes  in  

the  use  of  language.  • The  data  here  suggest  that  students  do  not  regard  general  feedback  as  highly  

as  more  detailed/specific  feedback.    The  comparison  of  item  endorsement  for  the  statements  within  each  question  showed  that,  overall,  participants  most  strongly  

• sought  to  use  feedback  to  benefit  their  learning;  • felt  that  negative  feedback  was  beneficial  and  not  detrimental;  • regarded  feedback  as  useful  for  improving  their  understanding  of  both  course  

content  and  course  requirements;  and  • regarded  feedback  that  explains  mistakes  and  specifies  how  to  improve  

written  work  as  the  most  important  of  all  feedback  types  mentioned.    As  regards  students  from  a  NESB,  analysis  of  the  data  suggests  that,  overall  

• non-­‐English  at  home  speakers  were,  on  average,  more  satisfied  with  the  quality  and  quantity  of  feedback  than  their  English  at  home  peers;  

  7  

• those  who  self-­‐rated  their  fluency  in  English  as  ‘reasonable’  or  ‘poor’  (the  categories  being  ‘native  speaker’,  ‘near  native’,  ‘reasonable’,  ‘poor’)  placed  greater  importance  on  detailed,  written  feedback  than  those  who  rated  themselves  as  more  fluent.  

 Other  results  suggest  that  

• higher  achieving,  more  confident  students  are  more  desirous  of  feedback  than  their  lower  achieving  counterparts.  

 Answers  to  the  5  open-­‐ended  questions  in  Student  Questionnaire:    1.  What  does  ‘feedback’  mean  to  you?  

• Most  students  referred  to  o information  about  weaknesses  and  strengths,  and  o suggestions  for  improvement  

• Other  common  answers  were  o constructive  criticism  o personalised  comments/advice/suggestions  o explanations  of  why  answer  is  not  correct  o examples  of  how  they  could  have  done  better  o comments  on  

§ what’s  missing/what  more  could  have  been  done  § writing  technique,  and  § student’s  work  in  relation  to  that  of  rest  of  class  

 2.  What  kind  of  feedback  is  most  helpful  to  you?  

• Answers  generally  mirror  those  for  Q1,  i.e.  related  to  o formative  feedback  

§ strong  and  weak  points  § suggestions  for  improvement  § specific,  constructive  criticism  § examples  of  good/correct  responses  

• Other  common  answers:  o written  o verbal,  one-­‐on-­‐one  o detailed  o concise  o specific  o general  o timely    

 3.  What  kind  of  feedback  is  least  helpful  to  you?  

• Answers  specified  o the  mark  only  o ticks  and  crosses  without  explanation  o short  comments,  especially  negative  feedback,  without  explanation  o ambiguous/vague  comments  

 • Other  common  answers:  

o group/generalised  o marking  rubric  with  no  comments  

  8  

o none!    4.  Do  you  use  feedback  on  written  assignments  when  writing  future  assignments?  If  yes,  how?  If  no,  why  not?  

• YES  answers  dominated:  o Most  students  referred  to  using  feedback  by  

§ trying  not  to  make  the  same  mistakes  § implementing  any  suggestions  § keeping  what  they  did  well  

o Some  students  mentioned  a  preference  for  general  comments,  others  a  preference  for  specific  comments.  

• There  was  a  small  proportion  of  NO  answers:  o Most  students  gave  the  following  reasons:  

§ forgetting  to  look  back  § feedback  on  one  assignment  not  relating  to  next  § not  enough  feedback  given  to  use  on  next  assignment  § different  marking  styles  of  tutors  § not  receiving  feedback  in  time  

 5.  What  aspect(s)  of  feedback  given  on  your  written  assignments  at  University  has/have  been  most  useful  to  you  in  helping  to  improve  your  writing?  Many  students  seem  to  have  interpreted  this  question  as  ‘What  kind  of  feedback  would  be  most  helpful  to  you?’  

• Most  students  referred  to  feedback  on  o what  they  did  wrong/badly  o why  this  was  wrong  o what  they  can  do  to  improve  o what  they  need  and  do  not  need  to  include  next  time  o what  parts  of  the  assignment  are  good/well  interpreted  o writing  style  o misunderstandings  of  subject  matter  (with  explanation)  o structural  deficiencies  o faults  in  essay  argument  o how  to  improve  referencing  o sentence  structure  and  use  of  language  

• Other  common  answers  were  o written,  specific  feedback  o precise,  relevant  criticism  o individualised,  verbal  feedback  o examples  of  better  ways  of  proceeding  o feedback  on  drafts  

 

1.2.2.   Student  Focus  groups    General  information:  

• 16  students  attended  the  voluntary  focus  group  sessions  o 7  monolinguals  o 6  bilinguals  o 3  NESBs  

• 6  sessions  were  held  o 2  with  monolinguals  only  

  9  

o 1  with  bilinguals  only  o 2  with  monolinguals  and  bilinguals  o 1  with  NESBs  only  

• The  open-­‐ended  question  answers  of  the  four  cohorts  were  reflected  in  the  individual  comments  of  the  student  focus  group  participants  

• In  some  cases,  the  comments  evidenced  disagreement  among  group  members  about  the  usefulness  of  oral  vs.  written  and  whole  class  vs.  individual  feedback  

 Native  Speakers:  Monolinguals  and  Bilinguals  1.  What  kind  of  feedback  is  most  helpful  to  you?  

• Specific,  detailed,  targeted,  personalised  o If  it’s  detailed  feedback  and  it’s  directed  at  your  problem  areas,  then  it’s  

definitely  good  feedback.  • Improvement  oriented:  

o …  I  went  really  well  but  I  still  lost  marks.  So  …  how  could  I  improve?  Because  that’s  the  main  thing  you  want  to  be  knowing.  It’s  constructive  criticism,  you  did  this  wrong  but  okay,  so  what  can  I  do  better?  

o 10%  more  –  would  have  been  nice  to  know  how  to  have  achieved  that.  • Written:  

o Oral  at  the  time  is  useful,  but  …  it’s  a  lot  more  useful  to  have  written  to  go  back  to.  

• Oral:  o Then  they’re  going  to  go  into  more  detail  and  you  can  understand  it  

better.  o If  it’s  verbal,  you  can  actually  ask  questions  if  you’re  confused.  

• Constructive  negative  feedback:  o I  really  like  it  if,  within  the  criticism,  there  might  be  some  good  points  in  

there,  some  positive  things.  • Timely:  

o Especially  when  you  can  use  [the  feedback]  for  the  next  assessment.  • Also:  

o Whole  class  oral  and  written:  § It  is  helpful  because  then  you  know  how  the  people  who  did  well,  

what  they  did  that  was  good,  and  then  you  also  found  out,  well,  I  could  do  it  better  the  next  time.  

§ how  we  went  in  relation  to  other  students  (spread  of  marks).  o Model  answers:  

§ This  is  what  a  good  paper  looks  like,  and  then  we  can  …  do  the  self-­‐learning,  actually,  we  don’t  need  to  be  told,  we  can  absorb  a  lot  of  information  just  from  seeing  what  excellence  looks  like.  

o Comprehensive  feedback  (on  all  aspects  of  the  report):  § Content,  as  well  as  …  how  we  should  approach  the  assignment  or  

report.  What  we  should  emphasise  …  all  the  different  sections  and  what  you  should  include.  What  is  the  normal  approach.  

   

  10  

 2.  What  kind  of  feedback  is  least  helpful  to  you?  

• Mark  only  and/or  ticking  of  boxes  on  a  marking  rubric/grid,  with  no  written  comments  on  assignment  itself  –  no  way  of  knowing  where  or  why  marks  were  lost:  

o No  explanation  …  you  don’t  know  what  to  think  of  it.  o Like,  where  did  I  go  wrong?  …  I  thought  I  understood  everything,  I  

thought  I’d  answered  everything  quite  well;  got  6  out  of  10,  so  I  don’t  know  …  

• Non-­‐constructive/only  negative:  o …  if  the  feedback  is  very  negative  and  doesn’t  really  give  you  a  

springboard  to  be  proactive  and  improve  where  you  are  going.  o If  all  they  say  is  where  you  went  wrong,  then  it’s  just  criticism  rather  

than  helpful  feedback.    

• Impersonal,  general  comments:  o  Directed  at  the  whole  class  o  A  general  comment  at  the  end  of  the  assignment  o Comments  without  suggestions  for  improvement,  such  as  ‘You  need  to  

elaborate  on  this’  • Too  specific:  

o So  specific  it’s  unusable  elsewhere,  in  any  other  sort  of  framework  …  You  can’t  really  transfer  it.  

• Received  after  the  next  assignment  or  after  the  exam:  o …  so  you  can’t  improve  from  one  to  the  other.  So  that  demotivates  me  to  

go  and  process  it  because  I  needed  it  earlier.  • Advice  on  writing  the  report  of  the  experiment  not  given  in  lab:  genuine  

questions  on  writing  the  report  not  answered  o They’ll  say  ‘…  you  can  completely  find  everything  you  need’,  but  then  you  

need  some  little  help,  especially  if  like  you  know  you’ve  got  to  write  a  report  on  it,  you  want  to  at  least  feel  confident  going  to  attempt  the  report,  and  if  you  …  leave  the  place  going  ‘Well,  that  was  fun  …’,  you’re  really  not  going  to  do  stuff  for  the  report.  

 3.  Do  you  use  feedback  on  written  assignments  when  writing  future  assignments?  If  yes,  how?  If  no,  why  not?  

• Yes:  When  it  can  be  used  for  another  assignment  o Return  to  feedback  when  writing  next  report  o Return  to  feedback  sometimes  even  years  later:  good  because  if  reread  

essay  then,  § you’re  coming  from  the  tutor’s  perspective  

• No:  When  it’s  so  specific  it’s  unusable  elsewhere  • Inconsistency  in  marking:  

o One  report,  you  know,  tick,  tick,  tick,  tick,  the  next  report  I  do  exactly  the  same  thing,  but  a  different  tutor  says,  ‘No,  you  shouldn’t  have  done  that,  you  shouldn’t  have  done  that’  …  

 4.  What  aspect(s)  of  feedback  given  on  your  written  assignments  at  University  has/have  been  most  useful  to  you  in  helping  to  improve  your  writing?  

• Individual,  written,  timely,  informative:  

  11  

o I’ve  had  …[demonstrators]  who  just,  you  know,  …  feedback  straight  away,  ...  write  down  exactly  where  you  lost  the  marks,  what  the  answer  is,  why,  so  ‘If  you  need  explaining,  let  me  know’.  Just  absolutely  helpful.  

• Individual/group  oral:  o Demonstrators  for  most  pracs  have  been  fantastic  in  giving  their  own  

feedback,  and  …  if  we’ve  got  any  questions,  there’s  no  problems  for  us  to  ask  them  questions,  which  is  good.  

o In  our  little  desk  with  our  tutor,  …  we  get  some  good,  like  just  non-­‐formal  feedback  from  her  …  She  often  takes  us  into  the  tute  room  to  go  over  like  the  questions  and  the  discussions  …  with  us,  just  to  make  sure  we’ve  really  understood  it  …  

• Improvement-­‐oriented:  o Things  that  help  you  improve  the  next  time,  that  you  can  utilise  in  the  

next  assessment.  o [When]  they  give  us  back  the  actual  quiz  paper  and  also  the  model  

answers  ...  I  could  actually  go  through  and  see  where  I’d  gone  wrong.  • All  feedback  is  useful  (even  a  small  amount):  

o I  did  terribly  in  first  year  reports  and  I  have  significantly  improved  this  year  …  We  weren’t  given  too  much  feedback,  but  …what  little  we  were  given  was  useful.  If  we  were  given  even  more,  then  obviously  we  would  have  done  even  better.  

• Experience  of  acting  as  a  marker  (one  student)    NESB  students  

1. What  kind  of  feedback  is  most  helpful  to  you?  • Specific:  

o  On  what  they  can  do  to  improve  next  time,  but  if  too  specific  –  won’t  be  relevant  to  next  task  

o In  whole  sentences  rather  than  phrases  o On  their  strengths:  

…  like  my  idea  is  good  but  only  problem  is  like  my  grammar  or  structure  like  that.  I  want  to  know  what  I’ve  good,  so  I  can  keep  that  and  change  the  wrong  part.  

• Improvement-­‐oriented  with  examples:  o It’s  better  with  some  examples  how  to  do  it,  that  tell  us  what’s  wrong  

with  that  and  what  we  should  do  …  in  the  next  time.  • Written:  

o Summary:  student  can  read  it  again  later,  won’t  forget  it.  If  he  doesn’t  understand  the  comments,  he’ll  make  an  appointment  to  discuss  it.  

• Oral:  o Summary:  sometimes  written  feedback  is  too  short  and  he  doesn’t  

understand  it.  He  asks  tutor  for  oral  feedback  in  pracs  and  tutorials  and  sometimes  makes  a  note  of  it:  It’s  more  interactive  in  oral  feedback.  

• Feedforward  is  the  best:  o Better  before  we  do  it!  

 2.  Feedback:  least  helpful;  future  use;  most  helpful  feedback  at  university  so  far  

• Least  helpful:  o Confusing  

§ if  the  feedback  makes  me  confused,  what  I  did  the  wrong  …  yeah,  that’s  kind  of  the  detriment.  

  12  

o In  challenging  handwriting  o Marking  rubrics:  helpful  with  structuring  the  report,  but  without  

feedback  comments  on  theory  or  content,  do  not  know  how  to  improve    • Future  use:  

o Students  read  feedback  when  marked  assignment  received  and  again  before  the  next  assignment  

• Most  helpful  feedback  so  far:  o Written  feedback  o Talking  with  tutor  after  written  feedback  

 3.  Feedback  on  language  matters:  

• Feedback  on  grammar:  o Most  students  wanted  specific  feedback  in  the  language  area,  

especially  grammar:  § I  need  to  know  specific  where  I  wrong  in  English,  because  I’m  

learning  …  I  just  need  to  know  the  biggest  problems  …  o One  student  said  tutors  should  not  have  to  correct  grammar:  he  liked  

comments  pointing  out  areas  where  he  could  improve,  e.g.  the  structure  of  an  assignment,  bringing  in  the  other  side  in  a  debated  area,  but  not  comments  where  he  felt  he  could  not  improve:  

§ hard  to  change,  like  for  example,  grammar  or  writing  skills.  • Markers’  response  to  grammar  problems:  

o Markers  draw  attention  to  grammar/expression  problems,  but  don’t  point  out  what’s  wrong  or  correct  grammar:    

§ they  write  ‘Poor  structure’,  ‘Poor  grammar’,  but  they  didn’t  point  out  where  is  the  wrong.  

• Prior  experience  of  English  language  preparation  programs/Foundation  programs  

o At  (Uni  Prep)  Taylor’s  College,  much  more  feedback  on  grammar  and  word  choice  issues  than  at  uni  

o At  uni,  most  feedback  concerned  with  content  and  structure  issues,  what’s  right  and  what’s  wrong  (e.g.  theoretical  mistakes).  

 Thus,  the  main  findings  from  the  student  questionnaire  and  the  focus  groups  showed  that  students  were  generally  positive  about  the  importance  of  good  feedback,  stating  that  feedback  was  more  helpful  if  it  was  personalised  and  timely.        

  13  

2.0   The  Feedback  Project:  Markers’  Perspectives    In  Section  2,  you  will  find  summaries  of  markers’  experiences  and  their  problems  with  giving  feedback.    

2.1 Marker Interviews

2.1.1   Problems  with  written  feedback  • Time  taken  writing  feedback  on  each  student  assignment:  no  real  payment  for  

time  spent  giving  detailed,  specific,  individualised  feedback  • Time  taken  when  giving  more  elaborated  individual  oral  feedback  • Timeliness:  feedback  irrelevant  if  received  at  the  end  of  semester/after  exam  • Feeling  that  students  are  not  going  to  read  feedback,  and  if  they  do,  are  not  

going  to  act  on  it/learn  from  it  • Possibility  of  student  misinterpretation  of  feedback  comments  • Frequent  repetition  of  same  comment  • Fatigue  leading  to  more  generic  and  less  specific  comments  • Student  misinterpretation  of  seriousness  of  flaws  if  comments  phrased  in  

relatively  positive  terms  • Students  simply  reading  comments  without  then  reconsidering  the  related  

problem  areas  in  their  work  

2.1.2   Marker  comments  re  dealing  with  diverse  language  backgrounds  of  students  Comments  from  markers  who  take  language  background  into  consideration:  

• Some  commented  on  different  feedback  given:  o If  there  are  language  problems,  and  if  a  non-­‐English  speaking  

background  (NESB)  is  suspected/known,  § they  might  phrase  their  feedback  about  these  problems  

differently  from  the  way  they  would  with  English  speaking  background  students,  i.e.,  it  might  be  more  gentle;  

§ they  make  sure  their  feedback  is  easily  understandable.  • Some  engaged  in  self-­‐questioning:  

o If  NESB  students  are  not  marked  down  for  poor  phrasing,  and/or  if  this  aspect  of  their  work  is  not  commented  on,  is  this  fair?  

 Comments  from  markers  who  try  not  to  take  language  background  into  consideration:  

• Some  said  o they  ignore  language  problems  unless  these  impede  understanding  of  

content;  o they  award  no  ‘sympathy  points’  because  of  language  background  and  

mark  on  the  merit  of  the  work  only;  o they  give  a  lower  mark  to  all  students  whose  papers  are  written  in  less  

coherent  English,  irrespective  of  their  background;  o they  have  difficulty  sometimes  distinguishing  between  students  with  

poor  language  skills  and  those  who  have  made  no  effort  (they  want  to  reward  effort);  

  14  

o if  a  student’s  written  expression  is  not  clear,  § it  is  difficult  to  know  whether  they  have  fully  mastered  the  

concept;  § and  also  difficult  to  articulate  to  them  in  detail  specifically  

where  their  writing  is  not  clear.    

2.2 Marker workshops

Three  Marker  workshop  sessions  were  held,  with  the  following  focuses  and  titles:  • ‘Current  practices  and  problems’  • ‘Feedback  practices  from  the  literature’,  and  • ‘An  opportunity  for  improvement:  Applying  new  approaches’.  

26  markers  attended  from  the  areas  of  Physiology,  Psychology,  Biological  Sciences,  Molecular  Bioscience  and  Chemistry.  Online  trigger  questions  were  completed  before  Sessions  1  and  3,  and  each  participant  read  a  journal  article  on  feedback  before  Session  2.    The  first  session  included  

• small  group  cross-­‐disciplinary  discussion  of  markers’  current  practices  and  problems  

• a  general  sharing  of  the  main  points  of  these  discussions  • a  summary  of  the  online  trigger  question  responses  • summaries  of  the  quantitative  and  qualitative  data  outcomes  

o from  the  student  questionnaires  in  the  Feedback  Project,  and  of  o the  data  from  the  student  focus  groups.  

The  second  session  included  • small  group  critical  discussion  of  the  feedback  research  articles  the  markers  

had  read,  with    o presentations  of  the  summaries  of  group  ideas  

• presentation  of  the  marker  interview  data  • the  linking  of  this  data,  the  Session  1  discussions  and  the  responses  to  Session  

1’s  first  trigger  question  to  the  wider  literature.  In  the  final  session,  

• the  related  trigger  question  responses  were  summarised  • discipline-­‐based  groups  discussed  the  feasibility  of  feedback  

strategies/changes,  and  • the  markers  committed  to  1-­‐2  ways  of  changing  practice  at  the  micro-­‐level.  

 The  workshops  were  then  evaluated  by  the  attendees,  and  they  were  asked  to  complete  a  post-­‐workshop  online  survey  summarising  and  evaluating  the  changes  they  attempted  when  providing  feedback  on  students’  written  assignments.    

2.2.1.  Current  practices  and  problems    Trigger  Question  1:  ‘What  are  the  top  3  problems  you  encounter  when  providing  feedback  on  students’  writing?’  and  the  first  workshop  discussion  identified  the  main  problems  the  markers  encountered  when  providing  feedback  on  students’  written  work.    

  15  

The  first  main  problem  area  identified  was  the  actual  feedback  provided,  i.e.  the  difficulty  of  

• expressing  their  feedback  o appropriately,  clearly  and  specifically,  and    o objectively,  concisely  and  constructively  

• dealing  with  the  poor  expression  of  ideas  • articulating  ways  of  improving  • finding  something  positive  to  say  • editing  vs.  identifying  ways  to  improve  • dealing  with  literacy  matters  vs.  discipline-­‐specific  content  • dealing  with  the  work  of  NESB  students  • deciding  whether  to  provide  specific  or  general  feedback,  or  both,  and  in  how  

much  detail  The  second  problem  area  identified  was  time,  i.e.  the  mismatch  between  the  time  taken  to  write  comprehensive  feedback  and  the  time  allocated  by  the  payment  system.  Other  important  concerns  were  the  difficulty  of  

• dealing  with  the  constraints  of  standardized  feedback  sheets  • marking  very  poor  papers  • ensuring  equity  across  all  students  • dealing  with  ‘my  students’  vs.  others,  i.e.  thinking  a  good  student  understands  

the  material,  even  if  this  hasn’t  been  articulated  • dealing  with  a  lack  of  quality  control:  no  models  provided,  little  training,  lack  

of  discussion  re  marking  criteria    Trigger  Question  2:  ‘What  do  you  think  students  want  out  of  the  feedback  they  receive?’  The  responses  included  knowing  

• what  was  done  well  • what  was  done  poorly/what  they  can  improve  on  • how  they  can  improve  • how  the  mark  was  arrived  at  • some  encouragement  

 Trigger  Question  3:  ‘Do  you  consider  student  diversity  in  your  marking  practice?  If  so,  in  what  ways?’.  

• Most  responded  ‘No’,  o stressing  that  all  students  needed  to  be  marked  in  the  same  way,  o some  adding  that  they  tried  to  focus  on  the  content  rather  than  

‘perfectly  good  English’  but  that  if  the  language  was  poor,  the  students  would  have  more  difficulty  expressing  themselves  and  so  would  be  penalised.  

• Others  responded  ‘Yes’,  o explaining  that  they  would  not  penalise  grammar  or  spelling  mistakes  

as  long  as  they  could  understand  what  the  student  was  saying;  o that  is,  they  would  focus  on  the  concepts  ‘rather  than  worrying  about  

exactly  how  they  expressed  them’.      

Thus,  marker  interviews  and  marker  workshops  raised  similar  issues:  the  time  needed  for  good  feedback  and  the  difficulty  related  to  giving  appropriate  and  good  quality  feedback,  particularly  for  students  from  diverse  language  backgrounds.        

  16  

 

3.0   The  Feedback  Project:  Suggested  Solutions  to  Feedback  Issues  and  Challenges    

3.1 Coordinator-centred issues and suggestions

• Discipline/School/Faculty  financial  support  to  coordinators  o to  develop/revise/adapt  assessments  to  align  with  the  University’s  

Assessment  Policy  o to  develop  resources  for  markers  and  students  on  how  to  provide  and  

use  feedback  o to  develop  resources  for  sessional  marker  training  

• Encourage  markers  to  provide  feedback  on  assessment  design,  marking  rubric,  marker  training,  etc.  

• Have  pre-­‐  and  post-­‐marking  meetings  for  all  markers  • Feed  back  to  lecturers  concepts  students  find  difficult  or  misunderstand  as  

evidenced  in  their  assignments  • Integrate  into  the  curriculum  how  to  use  feedback  for  students  to  build  skills  

in  interpreting  &  using  feedback  • Engage  in  coordinating  assessment  design  across  a  program  aligning  with  

feedback  to  feed  forward  in  the  development  of  discipline  specific  knowledge  and  skills  and  graduate  attributes  

 

3.2 Marker-centred issues and suggestions

3.2.1   Solutions  suggested  in  Marker  Interviews    

• The  provision  of  clear  criteria  –  some  are  too  vague,  others  too  broad  • The  holding  of  a  discussion  session  with  lecturer/coordinator  prior  to  

marking  re  what  students  might  have  written  • The  provision  of  specific  guidelines  concerning  what  to  give  marks  for  and  

what  not  to  give  marks  for  • The  provision  of  specific  guidelines  concerning  what  is  expected  of  all  

markers,  like  ‘Give  qualitative  feedback;  don’t  just  tick  the  rubric  boxes’  • The  provision  of  a  summary  of  the  components  of  a  good  response  to  each  

essay  question  or  a  detailed  rubric  for  each  question  • The  provision  of  a  sample  HD  essay  for  each  topic  • The  provision  of  feedback  statement  banks  • When  questions  are  raised  by  individual  markers,  the  emailing  of  Q  and  

answer  to  all  markers  • The  holding  of  a  post-­‐marking  meeting  re  problems  encountered  and  changes  

that  need  to  be  made  to  marking  scheme  for  following  semester  • The  provision  of  a  marker  training  course  • More  communication  between  coordinator  and  all  markers  • Encouraging  markers  to  email  coordinator  with  general  feedback  re  

assignments  which  can  be  posted  online  • Marker  ‘stars’  best  answers  and  these  are  posted  up  outside  their  office  so  

that  students  see  what  a  good  answer  entails  

  17  

• Hearing  what  students  have  to  say  about  the  marking  process  • Getting  students  to  think  like  markers:  inviting  them  to  a  voluntary  

assessment  workshop.  This  involves  o the  use  of  good  and  poor  examples  of  past  students’  work  (exemplars),  o the  discussion  of  why  these  are  good  or  poor,  o the  discussion  of  marking  criteria,  o student  group  discussion  and  the  marking  of  exemplars,  using  criteria,  o the  discussion  of  groups’  marks,  and  o the  comparison  with  marks  given  and  discussion  of  reasons  why.  

 

3.2.2   Solutions  suggested  by  Workshop  participants:    Strategies  suggested  in  discussions:  

• Better  quality  control:  institution  of  ‘blind’  marking  as  standard  • Report  writing  tutorial  for  students  as  standard  • Tutorial  on  ‘good/bad’  report  examples,  with  students  marking  these  • Student  writing  of  a  500-­‐word  final  Psychology  essay  plan:  this  has  to  be  

approved  before  work  on  essay  goes  ahead  • The  swapping  and  marking  of  draft  Biology  essays  with  peers  in  a  workshop  • Tutor  pointing  out  of  ‘the  obvious’  in  assignment  guidelines!  

 Strategies  suggested  in  the  responses  to  trigger  questions    Question  1:  ‘Identify  at  least  one  strategy  or  change  you  could  implement  that  would  address  one  or  more  of  the  problems  you  identified  when  providing  feedback  on  students’  writing’.  Responses  included  

• using  an  audio  file  instead  of  written  feedback  • focussing  feedback  on  major  points  • outlining  how  to  improve  • Starting  comments  with  ‘Next  time,  …’  • providing  more  comprehensive,  clear  and  positive  feedback  • going  over  the  marking  criteria  with  students  • having  model  reports  • student  self-­‐assessment  of  work  prior  to  submission  • focus  groups  to  discuss  tutor  feedback  effectiveness  • the  identification  of  common  problems  that  students  encounter  

   Question  2:  ‘What  help  do  you  think  could  be  given  to  markers  to  support  them  in  the  marking  and  feedback  process?’  Responses  here  included:  

• the  allocation  of  more  paid  time  per  assignment  • guidelines  from  coordinators  detailing  specific  criteria  that  should  be  met  for  

a  student  to  receive  a  particular  grade  • better/clearer  marking  criteria  • a  review  process  to  ensure  consistency  in  marking  standards  • marker  training  • listening  to  student  feedback  on  the  assessment  and  feedback  process  

 Question  3:  ‘Could  assessments  be  improved  in  any  way  to  help  you  to  give  students  better  feedback?  If  yes,  please  briefly  explain  how’.  Responses  included:  

• the  draft  submission  of  assessments  

  18  

• more  than  one  assessment  testing  the  same  skills,  allowing  students  to  apply  feedback  

• the  alignment  of  the  objectives  of  assessment  with  marking  criteria  and  course  requirements  and  outcomes  

• early  simple  essay  questions,  with  a  gradual  increase  in  difficulty  • clear,  detailed  and  specific  marking  criteria  to  reduce  inconsistency  among  

markers  • outlining  that  English  expression  will  form  part  of  the  assessment  

 Strategies  markers  implemented  after  the  workshop  and  found  useful:  

• Drawing  up  a  list  of  the  most  common  comments  made  on  assessments  and  using  these  to  identify  the  most  common  problem  areas  in  written  assignments;  

• Rather  than  thinking  of  the  positive  and  negative  aspects  of  what  the  student  wrote,  focusing  on  what  the  student  should  change  in  order  to  produce  a  better  outcome  in  the  next  assessment  (i.e.  implying  strengths  and  weaknesses  this  way);  

• Providing  audio  files  for  each  student,  in  addition  to  brief,  written  comments;  • Providing  specific,  direct  feedback;  • Giving  students  a  clear  marking  scheme  before  submission  and  using  this  as  

the  feedback  sheet;  • Providing  positive  feedback;  • Being  more  concise  in  feedback  comments,  and  focussing  on  a  few  major  

points:  at  the  end  of  student  report,  brief  written  feedback  about  what  was  done  well  and  what  could  be  improved  on;  

• Being  more  positive  in  comments  made,  starting  with  a  positive  statement  before  moving  to  comment  on  mistakes;  being  constructive:  noting  the  mistake,  but  also  explaining  how  to  improve;  

• When  giving  essay  feedback,  drawing  on  a  pool  of  generic  comments  that  can  be  copied  from  a  Word  document;  additional  individualised  feedback  also  given.  

 Other  marker-­‐suggested  but  unimplemented  strategies:  

• as  well  as  annotating  individual  assignments,  providing  students  with  a  photocopied  sheet  of  longer  explanations  for  the  most  frequently  encountered  problems;  

• getting  students  to  write  a  draft  report  and  giving  feedback  on  this;  students  rewrite  report,  integrating  the  feedback  given.  

 

3.3 Suggestions from the literature to bring students into ‘the community of assessment practice’

• Before/at  the  time  of  final  submission:  o Peer  feedback  using  assessment  criteria  before  final  draft  written:  this  

can  be  organised  via  webCT  o Student  written  self-­‐assessment  according  to  assessment  criteria  when  

handing  in  a  piece  of  work  o Draft,  then  feedback  from  marker,  then  final  submission  with  mark  

only  given  

  19  

o Student  required  to  explain  (in  writing)  how  they’ve  used  prior  feedback  in  current  assignment  

• After  submission:  o Digital  audio  feedback:  using  voicefiles  

§ see  the  ‘Sounds  Good’  website  at:  http://sites.google.com/site/soundsgooduk/  

o Marker  feedback  given  on  work  submitted,  then  student  comments  on  this:  no  mark  given  until  student  comments  received  by  marker  

o Work  returned,  with  feedback  and  criteria:  students  asked  to  rate  themselves  with  reasons,  before  mark  given  

o Early  generic  feedback  to  whole  class  rather  than  specific  individual  feedback  which  is  available  too  late  

o Limiting  the  number  of  feedback  comments      

  20  

Part  B:   What  the  literature  says:  ‘Suggested  Solutions’  Themed  Annotated  Bibliography  (non-­‐exhaustive):    NOTE: This Bibliography is limited to documents published, in almost all cases, from 2000 on.  Part  B  is  divided  into  two  main  sections,  each  providing  summary  tables  and  an  annotated  bibliography.  Section  4.1  presents  an  overview  of  recent  literature  on  feedback,  giving  first  some  relevant  background  information,  then  focusing  on  the  purposes  and  effectiveness  of  feedback,  and  students’  understanding  and  use  of  it.  Section  4.2  summarises  the  suggested  solutions  found  in  the  literature,  focusing  on  good  feedback  practice,  the  wording  of  feedback  comments,  student  engagement,  dialogic  approaches,  using  technology,  dealing  with  large  classes,  and  feedforward.  These  solutions  are  then  developed  in  the  annotated  bibliography  which  follows.  Section  5  provides  a  glossary  of  key  terms.      

  21  

Table  1:  From  the  literature:  Overview  of  Feedback  –  Guide  Sources G

eneral concerns Purposes of feedback

Problems w

ith transm

ission view of

feedback

Student understanding of feedback com

ments

Student engagement

with feedback

Student use of feedback

Student perceptions of tutors’ w

ritten feedback

Effectiveness of feedback

Importance of ‘tacit

knowledge’

Black, P., & Wiliam, D. (1998). Assessment and classroom learning. Assessment in Education, 5(1), 7-74. doi: 10.1080/0969595980050102 ✓ Carless, D. (2006). Differing perceptions in the feedback process. Studies in Higher Education, 31(2), 219-233. doi: 10.1080/03075070600572132 ✓ Chanock, K. (2000). Comments on Essays : Do students understand what tutors write? Teaching in Higher Education, 5(1), 95-105. ✓ Gibbs, G., & Simpson, C. (2004). Conditions Under Which Assessment Supports Students ’ Learning. Learning and Teaching in Higher Education, 1(1), 3-31. ✓ ✓ Glover, C., & Brown, E. (2006). Written Feedback for Students : too much , too detailed or too incomprehensible to be effective ? Bioscience Education eJournal, 7(May).

Hattie, J., & Timperley, H. (2007). The power of feedback. Review of Educational Research, 77(1), 81-112. doi: 10.3102/003465430298487 ✓ Higgins, R., Hartley, P., & Skelton, A. (2001). Getting the message across: The problem of communicating assessment feedback. Teaching in Higher Education, 6(2), 269-274. doi: 10.1080/13562510120045230

Higgins, R., Hartley, P., & Skelton, A. (2002). The Conscientious Consumer : Reconsidering the role of assessment feedback in student learning. Studies in Higher Education, 27(1), 53-64. doi: 10.1080/03075070120099368

✓ ✓ ✓

Holmes, L. E., & Smith, L. J. (2003). Student evaluations of faculty grading methods. Journal of Education for Business, 78(6), 318-323. ✓ Hounsell, D., McCune, V., Hounsell, J., & Litjens, J. (2008). The quality of guidance and feedback to students. Higher Education Research & Development, 27(1), 55-67. doi: 10.1080/07294360701658765

Hyland, F., & Hyland, K. (2001). Sugaring the pill: Praise and criticism in written feedback. Journal of Second Language Writing, 10, 185-212. ✓ Nicol, D. J., & Macfarlane-Dick, D. (2006). Formative assessment and self‐regulated learning: A model and seven principles of good feedback practice. Studies in Higher Education, 31(2), 199-218. doi: 10.1080/03075070600572090

✓ ✓ ✓ ✓ ✓

  22  

Sources General concerns

Purposes of feedback

Problems w

ith transm

ission view of

feedback

Student understanding of feedback com

ments

Student engagement w

ith feedback

Student use of feedback

Student perceptions of tutors’ w

ritten feedback

Effectiveness of feedback

Importance of ‘tacit

knowledge’

Orsmond, P., & Merry, S. (2011). Feedback alignment: Effective and ineffective links between tutors’ and students’ understanding of coursework feedback. Assessment & Evaluation in Higher Education, 36(2), 125-136. doi: 10.1080/02602930903201651

Orsmond, P., Merry, S., & Reiling, K. (2005). Biology students' utilization of tutors' formative feedback: A qualitative interview study. Assessment & Evaluation in Higher Education, 30(4), 369-386. doi: 10.1080/02602930500099177

✓ ✓

Price, M., Carroll, J., O'Donovan, B., & Rust, C. (2010). If I was going there I wouldn't start from here: a critical commentary on current assessment practice. Assessment & Evaluation in Higher Education. Retrieved from doi:10.1080/02602930903512883

✓ ✓ ✓

Price, M., Handley, K., Millar, J., & O'Donovan, B. (2010). Feedback : all that effort, but what is the effect? Assessment & Evaluation in Higher Education, 35(3), 277-289. doi: 10.1080/02602930903541007

✓ ✓

Rust, C. (2007). Towards a scholarship of assessment. Assessment & Evaluation in Higher Education, 32(2), 229-237. doi: 10.1080/02602930600805192 ✓ Rust, C., Price, M., & O'Donovan, B. (2003). Improving students' learning by developing their understanding of assessment criteria and processes. Assessment & Evaluation in Higher Education, 28(2), 147-164. doi: 10.1080/02602930301671

Sadler, D. R. (1989). Formative assessment and the design of instructional systems. Instructional Science, 18, 119-144. ✓ ✓ University of Sydney feedback webpages ✓ ✓ University websites: Open University & Sheffield Hallam, Oxford Brookes, Strathclyde, Edinburgh; bioassess website ✓ ✓ Vardi, I. (2012). Effective feedback for student learning in higher education:

HERDSA. ✓ Weaver, M. R. (2006). Do students value feedback? Student perceptions of tutors’ written responses. Assessment & Evaluation in Higher Education, 31(3), 379-394. doi: 10.1080/02602930500353061

  23  

4.1: From the literature: Overview of Feedback

4.1.1   General  concerns  For  Hattie  and  Timperley  (2007),  feedback  ‘is  one  of  the  most  powerful  influences  on  learning  and  achievement’;  however  ‘this  impact  can  be  either  positive  or  negative’  (p.  81).  The  main  purpose  of  feedback  is  ‘to  reduce  discrepancies  between  current  understandings  and  performance  and  a  goal’  (p.  86).  They  summarise  the  results  of  their  previous  (1999)  synthesis  of  meta-­‐analyses  as  showing  that  ‘the  most  effective  forms  of  feedback  provide  cues  or  reinforcement  to  learners;  are  in  the  form  of  video-­‐,  audio-­‐,  or  computer-­‐assisted  instructional  feedback;  and/or  relate  to  goals’  (p.  84).  The  authors  then  identify  the  conditions  that  ‘maximize  the  positive  effects  on  learning’,  using  a  model  of  feedback  as  a  framework.  They  argue  that  effective  feedback  needs  to  answer  the  questions:  ‘Where  am  I  going?  (What  are  the  goals?),  How  am  I  going?  (What  progress  is  being  made  toward  the  goal?),  and  Where  to  next?  (What  activities  need  to  be  undertaken  to  make  better  progress?)’,  these  questions  corresponding  to  the  concepts  of  ‘feed  up,  feed  back  and  feed  forward’  (p.  86).  Their  model  also  distinguishes  between  four  levels  of  feedback:  ‘the  task,  the  processing,  the  regulatory,  and  the  self  levels’  (p.  102).  The  authors  then  address  a  number  of  ‘commonly  debated  issues’,  namely  ‘the  timing  of  feedback,  the  effects  of  positive  and  negative  feedback,  the  optimal  classroom  use  of  feedback,  and  the  role  of  assessment  in  feedback’  (p.  98).  They  conclude  that  ‘a  feedback  intervention  provided  for  a  familiar  task,  containing  cues  that  support  learning,  attracting  attention  to  feedback-­‐standard  discrepancies  at  the  task  level,  and  void  of  cues  that  direct  attention  to  the  self  is  likely  to  yield  impressive  gains  in  students’  performance’  (p.  104).    Gibbs  and  Simpson’s  (2004)  paper  centres  on  ‘how  to  design  assessment  that  supports  worthwhile  learning’,  arguing  that  ‘assessment  works  best  …  when  a  series  of  conditions  are  met’  (p.  4).  These  conditions  refer  to  two  categories:  ‘the  influence  of  the  design  of  assessment  systems  and  assignments  on  how  much  students  study,  what  they  study  and  on  the  quality  of  their  engagement’  and  ‘the  influence  of  feedback  on  learning’  (p.  12).  The  latter  conditions  (seven  of  the  ten  given)  are  that:  ‘Sufficient  feedback  is  provided,  both  often  enough  and  in  enough  detail’;  ‘The  feedback  focuses  on  students’  performance,  on  their  learning  and  on  actions  under  the  students’  control,  rather  than  on  the  students  themselves  and  on  their  characteristics’;  ‘The  feedback  is  timely  in  that  it  is  received  by  students  while  it  still  matters  to  them  and  in  time  for  them  to  pay  attention  to  further  learning  or  receive  further  assistance’;  ‘Feedback  is  appropriate  to  the  purpose  of  the  assignment  and  to  its  criteria  for  success’;  ‘Feedback  is  appropriate,  in  relation  to  students’  understanding  of  what  they  are  supposed  to  be  doing’;  ‘Feedback  is  received  and  attended  to’;  and  ‘Feedback  is  acted  upon  by  the  student’  (pp.  16-­‐25).    Nicol  and  Macfarlane-­‐Dick  (2006),  based  on  a  synthesis  of  the  research  literature,  identify  seven  principles  of  good  feedback  practice  that  support  student  self-­‐regulation  of  their  learning.  Self-­‐regulation  refers  to  ‘the  degree  to  which  students  can  regulate  aspects  of  their  thinking,  motivation  and  behaviour  during  learning’,  and  is  manifested  in,  e.g.,  ‘the  setting  of,  and  orientation  towards,  learning  goals’,  and  ‘reactions  to  external  feedback’  (p.  199).  The  principles  are  that  good  feedback  practice:  ‘1)  helps  clarify  what  good  performance  is  (goals,  criteria,  expected  standards);  2)  facilitates  the  development  of  self-­‐assessment  (reflection)  in  learning;  3)  delivers  high  quality  information  to  students  about  their  learning;  4)  encourages  

  24  

teacher  and  peer  dialogue  around  learning;  5)  encourages  positive  motivational  beliefs  and  self-­‐esteem;  6)  provides  opportunities  to  close  the  gap  between  current  and  desired  performance;  7)  provides  information  to  teachers  that  can  be  used  to  help  shape  teaching’  (p.  205).  The  authors  go  on  to  specify  teacher  strategies  to  assist  self-­‐regulation  for  each  of  these  principles,  including,  e.g.,  for  the  first  principle,  ‘involving  students  in  assessment  exercises  where  they  mark  or  comment  on  other  students’  work  in  relation  to  defined  criteria  and  standards’  (p.  207),  and  for  the  third  principle,  ‘providing  corrective  advice,  not  just  information  on  strengths/  weaknesses’  (p.  210).    Black  and  Wiliam’s  (1998)  literature  review  on  classroom  formative  assessment,  although  largely  concerned  with  school  situations,  makes  the  point  that  ‘the  quality  of  the  feedback  provided  is  a  key  feature  in  any  procedure  for  formative  assessment’  (p.  36).  They  add  that  ‘feedback  interventions  that  cue  individuals  to  direct  attention  to  the  self  rather  than  the  task  appear  to  be  likely  to  have  negative  effects  on  performance.  Thus  praise,  like  other  cues  which  draw  attention  to  self-­‐esteem  and  away  from  the  task,  generally  has  a  negative  effect’  (p.  49).  Gibbs  and  Simpson  (2004,  p.  9)  comment  that  ‘Black  and  Wiliam’s  (1998)  comprehensive  review  of  formative  assessment  emphasizes  the  extraordinarily  large  and  consistent  positive  effects  that  feedback  has  on  learning  compared  with  other  aspects  of  teaching.’    Hounsell,  McCune,  Hounsell  &  Litjens  (2008)  report  on  the  findings  of  a  large  survey  (questionnaires  and  group  interviews)  of  students’  experiences  of  the  provision  of  guidance  and  feedback.  They  present  a  ‘guidance  and  feedback  loop’  with  six  interrelated  steps:  ‘Students’  prior  experiences  of  assessments’,  ‘Preliminary  guidance’,  ‘Ongoing  clarification’,  ‘Feedback  on  performance/  achievement’,  ‘Supplementary  support’  and  ‘Feed-­‐forward’  (p.  60).  They  also  identify  potential  troublespots,  e.g.  ‘feedback  dwells  on  shortcomings  &  demotivates’,  and  ‘feedback  lacks  transparency’  (p.  66),  and  explain  how  a  shortcoming  in  one  step  could  affect  the  following  steps  in  the  loop,  either  positively  or  negatively.    Rust  (2007)  argues  for  the  articulation  and  establishment  of  a  ‘scholarship  of  assessment’  as  a  vital  part  of  the  scholarship  of  teaching  and  learning,  as  assessment  ‘seems  to  be  the  part  of  the  learning  and  teaching  process  that  has  been  especially  criticised’  (p.  229),  and  the  area  of  feedback  the  ‘worst  of  all’.  So  the  scholarship  of  assessment  should  include  ‘explicit  guidelines  on  giving  effective  and  prompt  feedback’  (p.  230).    

4.1.2   Purposes  of  feedback  Price,  Handley,  Millar  and  O’Donovan  (2010)  identify  five  roles  attributed  to  feedback:  ‘correction,  reinforcement,  forensic  diagnosis,  benchmarking  and  longitudinal  development  (feed-­‐forward)’.  The  authors  see  these  roles  as  acting  as  a  ‘nested  hierarchy’,  i.e.,  ‘each  building  on  information  provided  by  the  previous  category’  (p.  278).  They  argue  that  the  first  two  roles  are  not  really  relevant  for  much  feedback  in  higher  education,  which  ‘has  limited  scope  to  “correct”  complex  work’,  but,  citing  Sadler  (1989),  that  the  forensic  role,  ‘diagnosing  problems  with  the  work’,  links  with  the  benchmarking  role  ‘where  feedback  identifies  a  gap  between  what  is  understood/has  been  demonstrated  and  the  standard  of  performance  expected’.  Whether  feedback  can  help  to  fill  that  gap  depends  on  the  nature  of  the  gap:  

Where  the  gap  relates  to  the  curriculum  content,  the  feedback  may  be  able  to  specify  the  knowledge  that  needs  to  be  understood.  However,  where  the  gap  

  25  

identifies  the  need  for  development  of,  for  example,  academic  or  cognitive  skills,  feedback  may  not  be  able  to  be  specific  in  its  remedy  for  filling  the  gap  due  to  the  slowly  learnt  nature  of  knowledge  needed  to  address  the  gap.  (p.  278)  

(For  feedforward,  see  below.)    Feedback  is  also  used  by  markers  to  justify  the  grade  given.  Price,  Carroll,  O’Donovan  and  Rust  (2010),  discussing  current  assessment  practice,  argue  that  this,  rather  than  to  develop  learning,  tends  to  be  the  purpose  when  feedback  ‘is  concentrated  on  summative  work’,  because  of  the  ‘emphasis  on  reliability  of  marking’  (p.  3)  in  the  higher  education  sector.    

4.1.3   Problems  with  the  transmission  view  of  feedback  Nicol  and  Macfarlane-­‐Dick  (2006)  argue  that  in  higher  education,  ‘formative  assessment  and  feedback  are  still  largely  …  seen  as  the  responsibility  of  teachers;  and  feedback  is  still  generally  conceptualised  as  a  transmission  process,  even  though  some  influential  researchers  have  recently  challenged  this  viewpoint  (Sadler,  1998;  Boud,  2000;  Yorke,  2003)’  (p.  200).  They  identify  a  number  of  problems  with  this  transmission  view,  including  that  of  how  students  ‘can  become  empowered  and  develop  the  self-­‐regulation  skills  needed  to  prepare  them  for  learning  outside  university  and  throughout  life’,  and  the  problem  of  the  assumption  that  teachers’  feedback  information  is  ‘easily  decoded  and  translated  into  action’  (p.  201).    

4.1.4   Student  understanding  of  feedback  comments  Higgins,  Hartley  and  Skelton  (2001)  state  that  feedback  comments  ‘convey  a  message  based  on  an  implicit  understanding  of  particular  academic  terms,  which  in  turn  reflect  a  much  more  complex  academic  discourse,  which  in  turn  may  be  only  partially  understood  by  students’  (p.  272).  According  to  the  authors,  Hounsell  (1997)  and  McCune  (1999)  suggest  that  ‘HE  students  may  struggle  to  access  the  particular  discourses  underpinning  tutors’  comments’.    Chanock  (2000)  carried  out  a  study  on  the  common  marker  comment  on  essays  in  the  Humanities  of  ‘Too  much  description;  not  enough  analysis’,  asking  tutors  and  students  what  they  thought  this  comment  meant.  She  found  that  ‘almost  half  of  the  students  who  responded  did  not  interpret  this  comment  in  the  way  their  tutors  intended  it’  (p.  95).  Possible  reasons  for  this  include  the  ambiguous  nature  of  the  term  ‘analysis’  and  its  discipline-­‐specific  meaning.  She  concludes  that  ‘marking  comments  need  to  be  carefully  explained  with  examples  from  the  discourse  of  lectures,  tutorials  and  readings  in  the  disciplines’  (p.  95).    Hyland  and  Hyland  (2001)  offer  a  detailed  analysis  of  ‘the  written  feedback  given  by  two  teachers  to  ESL  students  over  a  complete  proficiency  course’,  considering  especially  its  interpersonal  aspects,  namely  ‘its  functions  as  praise,  criticism,  and  suggestions’.  Criticisms  and  suggestions  were  often  softened  by  praise,  ‘hedging  devices,  question  forms  and  personal  attribution’;  however,  this  sometimes  resulted  in  lack  of  student  understanding  of  these  indirect  comments.  So  such  indirectness  ‘carries  the  very  real  potential  for  incomprehension  and  miscommunication’  (2001,  p.  185).    

  26  

Holmes  and  Smith  (2003)  report  on  a  study  of  student  problems  with  grading  methods,  including  that  of  too  little  feedback  from  markers.  They  refer  to  the  research  of  Speck  and  Jones  (1998)  on  the  use  of  codes,  symbols  and  checkmarks  within  papers  ‘as  “minimalist  grading”  that  saves  faculty  members  time  but  may  not  enhance  learning’  (p.  319).  Putting  question  and  exclamation  marks  and  ‘Yes’  or  ‘No’  in  margins  ‘assumes  that  students  will  understand  what  the  teacher  is  referencing’,  while  using  codes  and  symbols  ‘assume[s]  that  students  …  will  look  up  what  the  symbols  mean  and  improve  their  work  in  following  assignments’,  assumptions  that  ‘may  be  erroneous’.  They  add  that  marginal  comments  ‘should  be  accompanied  by  closing  comments  that  give  a  summative  picture  of  the  student’s  work’,  addressing  strengths  and  weaknesses.  They  add  that  ‘giving  positive  comments  along  with  suggestions  for  improvement,  using  language  that  is  respectful,  and  giving  evidence  of  the  fairness  of  the  grades  are  good  practices  supported  by  the  literature’  (p.  319).    Higgins,  Hartley  and  Skelton  (2002,  p.  56),  referring  to  Hounsell  (1987),  discuss  the  failure  of  some  students  to  understand  ‘the  taken-­‐for-­‐granted  academic  discourses  which  underpin  assessment  criteria  and  the  language  of  feedback’.  They  point  out  that  teachers  ‘may  recognise  a  good  performance,  yet  struggle  to  articulate  exactly  what  they  are  looking  for  because  conceptions  of  quality  usually  take  the  form  of  tacit  knowledge’.  They  refer  to  a  body  of  literature  (from  2000  and  before)  which  echoes  the  view  ‘that  students  often  experience  problems  interpreting  the  academic  language  underpinning  assessment.’    See  also  Carless  (2006).      

4.1.5   Student  engagement  with  feedback  According  to  Price,  Carroll,  O’Donovan  and  Rust  (2010),  student  engagement  with  feedback  is  most  effective  ‘where:  the  purpose  of  the  feedback  is  clear;  it  has  utility  and  can  be  used  in  future  work;  it  accounts  for  the  transfer  of  tacit  as  well  as  explicit  knowledge  about  standards  and  disciplinary  knowledge;  and  the  student  perceives  some  sort  of  relationship  with  the  marker’  (p.  4).    Nicol  and  Macfarlane-­‐Dick  (2006)  report  that  ‘if  students  do  not  share  (at  least  in  part)  their  teachers’  conceptions  of  goals  (and  criteria  and  standards),  then  the  feedback  information  they  receive  is  unlikely  to  “connect”  (Hounsell,  1997)’  (p.  206).  This  makes  it  difficult  for  them  to  ‘evaluate  discrepancies  between  required  and  actual  performance’.    

4.1.6   Student  use  of  feedback  Orsmond,  Merry  and  Reiling’s  (2005)  study  of  Biology  students’  use  of  formative  feedback  showed  that  feedback  was  used  to  ‘enhance  motivation,  enhance  learning,  encourage  reflection  and  clarify  …  progress’  (p.  373).  Some  students  also  used    feedback  to  ‘enrich  their  learning  environment’  (p.  374).  For  the  students  interviewed,  motivation  resulted  from  either  positive  or  negative  comments,  though  some  comments  were  ‘upsetting’;  and  they  saw  learning  as  ‘developmental  and  progressive’  (p.  375).  Feedback  also  allowed  ‘clarification  of  tutors’  general  expectations  of  students’  performance  in  assignments’  (p.  376).  Among  the  authors’  recommendations  for  both  biology  and  other  discipline  areas  are  that  ‘previous  student  feedback  could  be  linked  to  “feedforward”  discussions  and  so  could  provide  a  focal  point  for  assignment  preparation’,  and  that  ‘heightened  awareness  by  tutors  

  27  

of  how  students  implement  feedback  could  positively  influence  how  tutors  write  feedback’  (p.  382).    

4.1.7   Student  perceptions  of  tutors’  written  feedback  The  research  project  of  Higgins,  Hartley  and  Skelton  (2002)  focused  on  (Business  and  Humanities)  students’  understandings  of  feedback.  The  research  suggests  that  the  students  were  ‘eager  to  read  feedback  comments’,  and  believed  they  deserved  feedback.  This  links  to  ‘a  perception  that  higher  education  is  a  service  and,  as  such,  it  is  also  the  tutor’s  “duty”  to  offer  feedback’.  The  authors  conclude  that  ‘it  is  …  likely  that  many  of  today’s  students  have  a  “consumerist  awareness”  reflected  in  a  focus  on  achieving  a  grade  alongside  intrinsic  motivations’  (p.  61).    Weaver’s  (2006)  small-­‐scale  study  of  Business  and  Art  &  Design  students’  perceptions  of  feedback  showed  that  ‘feedback  is  valued’,  but  that  ‘tutor  comments  could  be  more  helpful’  (p.  379).  Four  main  themes  of  feedback  considered  unhelpful  for  improving  learning  were  uncovered:  ‘comments  which  were  too  general  or  vague,  lacked  guidance,  focused  on  the  negative,  or  were  unrelated  to  assessment  criteria’.  The  results  suggest  that  ‘providing  a  better  balance  between  positive  and  critical  feedback’  is  ‘an  important  step  towards  making  feedback  more  effective’  (p.  392).    Orsmond  and  Merry  (2011)  discusses  the  misalignment  between  student  conceptions  of  the  role  of  feedback  and  the  focus  of  tutors  when  giving  feedback.  The  authors  make  a  number  of  recommendations  to  tutors,  including  that  they  should  ‘write  feedback  with  variation  in  it  and  draw  particular  attention  to  students’  development  as  professional  biologists  and  the  acquisition  of  lifelong  learning  skills’,  ‘make  greater  use  of  exemplars  and  whole-­‐class  feedback  to  show  variation’,  and  ‘encourage  a  greater  awareness  in  students  concerning  the  value  of  self-­‐assessment’  (p.  134).    

4.1.8   Effectiveness  of  feedback  Sadler  (1989)  outlines  a  theory  of  formative  assessment  (see  Glossary)  which  applies  ‘wherever  multiple  criteria  are  used  in  making  judgments  about  the  quality  of  student  responses’  (p.  119).  In  these  ‘complex  learning  settings’,  even  where  teachers  provide  ‘regular,  accurate  feedback’,  students  often  show  ‘little  or  no  growth  or  development’.  Sadler  argues  that  ‘students  …  need  more  than  summary  grades  if  they  are  to  develop  expertise  intelligently’  (p.  121):  

…  the  learner  has  to  (a)  possess  a  concept  of  the  standard  (or  goal,  or  reference  level)  being  aimed  for,  (b)  compare  the  actual  (or  current)  level  of  performance  with  the  standard,  and  (c)  engage  in  appropriate  action  which  leads  to  some  closure  of  the  gap.  

He  distinguishes  between  feedback  and  self-­‐monitoring,  explaining  that  formative  assessment  includes  both,  and  it  is  the  transition  from  feedback  to  self-­‐monitoring  which  is  at  the  heart  of  closure  of  the  gap.  This  transition  can  only  be  made  when  the  above  three  conditions  are  satisfied.  Sadler  contends  that  criteria  sheets  do  not  necessarily  deal  adequately  with  the  qualities  of  a  piece  of  work  because  there  are  both  manifest  and  latent  criteria  involved  in  evaluation,  and  that  students  need  to  develop  their  knowledge  of  both  in  order  to  be  able  to  monitor  their  own  performance,  i.e.  they  need  to  ‘be  given  adequate  evaluative  experience  themselves’  (p.  134)  to  develop  ‘an  appropriate  body  of  tacit  knowledge’  (see  Glossary).  This  can  

  28  

be  done  by  engaging  them  in  ‘evaluative  and  corrective  activity  on  other  students’  work’  (p.  140).    Price,  Handley,  Millar  and  O’Donovan  (2010)  point  out  that  feedback  can  only  be  effective  ‘when  the  learner  understands  the  feedback  and  is  willing  and  able  to  act  on  it’  (p.  279).  However,  in  higher  education,  most  feedback  requires  interpretation  by  the  learner,  and  students  can  choose  whether  to  act  on  it  or  not.  This  choice  may  depend  on  the  student’s  positive  or  negative  response  to  the  feedback,  on  its  timing,  and  on  the  student’s  level  of  engagement  with  it,  the  latter  depending  in  part  on  the  student-­‐tutor  relationship.  The  authors  report  on  a  three-­‐year  study  of  student  engagement  with  feedback,  revealing  the  difficulty  of  measuring  the  effectiveness  of  feedback  because  of  its  multiple  purposes  (see  above),  ‘its  temporal  nature  and  the  capabilities  of  evaluators’  (p.  277),  i.e.  staff,  students,  and/or  external  observers.  They  conclude  that  ‘the  pedagogic  literacy  of  students  is  key  to  evaluation  of  feedback  and  feedback  processes’  (p.  277).    Gibbs  and  Simpson  (2004)  argue  that  writing  comments  on  assignments  ‘remains  a  major  component  of  teachers’  workload  in  higher  education’,  and,  as  a  result  of  increased  class  sizes,  teachers  ‘can  find  themselves  spending  much  of  their  time  marking’.  They  ask  ‘Is  all  this  effort  worthwhile?’  and  ‘How  well  does  feedback  actually  work?’  (p.  10).    Nicol  and  Macfarlane-­‐Dick  (2006,  p.  213)  refer  to  the  second  of  Yorke’s  (2003)  questions  regarding  external  feedback,  namely  whether  feedback  leads  to  changes  in  student  behaviour,  citing  Boud:  

The  only  way  to  tell  if  learning  results  from  feedback  is  for  students  to  make  some  kind  of  response  to  complete  the  feedback  loop  (Sadler,  1989).  This  is  one  of  the  most  often  forgotten  aspects  of  formative  assessment.  Unless  students  are  able  to  use  the  feedback  to  produce  improved  work,  through,  for  example,  re-­‐doing  the  same  assignment,  neither  they  nor  those  giving  the  feedback  will  know  that  it  has  been  effective.  (Boud,  2000,  p.  158)    

 Orsmond,  Merry  and  Reiling  (2005)  report  on  Orsmond  et  al.  (2002)  which  showed  that  effective  feedback  ‘resulted  from  the  use  of  exemplars  as  a  focus  for  discussion  between  tutor  and  student  allowing  a  common  understanding  of  terms  to  be  developed’,  and  from  the  student  and  tutor’s  ‘common  understanding  of  how  this  feedback  might  be  implemented’  (p.  370).    Glover  and  Brown  (2006)  present  a  ‘detailed  analysis’  of  the  effectiveness  of  written  feedback  in  Biosciences  and  Physical  Sciences  at  two  UK  universities.  They  found  that  the  conditions  leading  to  effective  feedback  (Gibbs  &  Simpson,  2004)  ‘are  not  met  as  frequently  as  originally  believed’,  and  that  although  there  may  be  ‘plenty’  of  feedback,  ‘it  is  not  always  received  in  time  to  be  of  use  to  students’,  and  ‘may  be  misunderstood  in  relation  to  assessment  criteria’.  Other  findings  are  that  the  prime  function  of  the  feedback  given  is  ‘to  inform  the  students  about  their  past  achievement  rather  than  looking  forward  to  future  work’,  and  most  serves  primarily  ‘to  justify  the  grade’.  The  authors  suggest  that  ‘far  less  emphasis’  needs  to  be  placed  on  marks,  and  ‘far  more  on  achievement’;  that  ‘teaching,  learning  outcomes  and  assessment’  be  ‘constructively  aligned’;  and  that  ‘main  weaknesses’  be  focused  on  and  explained  in  depth.    

  29  

Vardi  (2012)  is  an  excellent  (HERDSA)  Guide  to  ‘Effective  feedback  for  student  learning  in  Higher  Education’.  The  author  makes  an  overview  of  the  literature  on  feedback,  and  provides  practical  lists  of  what  improves  and  what  impedes  student  performance  and  confidence.  She  then  discusses  practical  ways  of  putting  theory  into  action  in  a  university  context.  These  ways  include  planning  a  learning  program  which  embeds  the  provision  of  opportunities  for  teacher  feedback,  both  ‘quick  and  immediate’  in-­‐class  and  ‘more  detailed  delayed’,  and  for  ‘students  to  act  and  build  on  that  feedback’  (p.  vii).  She  describes,  and  gives  examples  of,  ‘how  to  give  powerful  feedback  immediately  in  class  situations  such  as  lectures,  laboratories  and  tutorials  and  their  online  equivalents’,  as  well  as  ‘how  to  give  good  quality  written  teacher  feedback  on  assignments’,  this  feedback  feeding  forward  to  future  student  work.  She  then  focuses  on  students  as  recipients  and  providers  of  feedback,  detailing  practical  ways  ‘to  improve  student  awareness  and  receptiveness  to  feedback’,  providing  ‘useful  tips’  on  helping  them  identify,  seek  and  use  feedback  effectively,  and  explaining  ‘how  to  support  peer  feedback  for  maximum  success’  (p.  viii).    

4.1.9   Importance  of  ‘tacit  knowledge’  (see  Glossary)  Rust,  Price  and  O’Donovan  (2003)  argue  that  this  kind  of  knowledge  is  experience-­‐based  and  ‘can  only  be  revealed  through  the  sharing  of  experience-­‐socialisation  processes  involving  observation,  imitation,  dialogue  and  practice  (Nonaka,  1991;  Baumard,  1999)’:  

So  over  time  discussion  and  shared  experiences  of  marking  and  moderation  among  staff  enable  the  sharing  of  tacit  knowledge,  resulting  in  more  standardised  marking  (Saunders  and  Davis,  1998).  It  follows  that  inviting  students  into  this  shared  experience  should  also  enable  more  effective  knowledge  transfer  of  assessment  processes  and  standards  to  them.  (p.  152)  

 See  also:  Higgins,  Hartley  and  Skelton  (2002),  Price,  Carroll,  O’Donovan  and  Rust  (2010),  Sadler  (1989),  and  Nicol  and  Macfarlane-­‐Dick  (2006),  above.    

4.1.10     University  Feedback  Websites  (see  others  under  ‘Using  Technology  to  provide  peer  and  marker  feedback’,  below)    The  Open  University  and  Sheffield  Hallam  University:  FAST  project  The  FAST  project,  a  collaboration  between  The  Open  University  and  Sheffield  Hallam  University,  ‘was  designed  to  change  assessment  so  as  to  support  student  learning  more  effectively’.  The  project  methodology  ‘involved  a  two  year  cycle  of  diagnostic  evaluation,  innovation,  and  evaluation  of  change  followed  by  one  year  of  wider  dissemination  to  science  teachers  in  other  institutions’.  The  project  aimed  to  address  Gibbs  and  Simpson’s  (2004)  eleven  conditions  under  which  assessment  supports  student  learning,  which  ‘relate  to  the  quantity,  distribution  and  quality  and  level  of  student  effort;  the  quantity,  timing  and  quality  of  feedback;  and  the  student  response  to  that  feedback.’  Two  major  outcomes  of  the  project  were  the  generation  of  Case  Studies  (the  website  provides  the  outcomes  of  this  activity)  and  ‘the  establishment  of  a  single  “Special  Interest  Group”,  looking  at  formative  assessment  in  science.  See  also  Brown  (n.d.)  below.  http://www.open.ac.uk/fast/FASTProject/Project.htm  Oxford  Brookes  University’s  ASKe  (Assessment  Standards  Knowledge  exchange)  website:  

  30  

This  website  makes  available  series  of  '1,2,3'  leaflets  ‘which  highlight  some  practical  ways  in  which  teaching  staff  can  improve  their  students'  learning’;  there  are  also  leaflets  for  students.  Each  leaflet  ‘focuses  on  a  piece  of  assessment-­‐related  research  and  clearly  states  how  that  research  can  be  applied  to  teaching  practice  in  three  easy  steps’.  Relevant  leaflets  for  tutors  are:  ‘How  to  make  your  feedback  work  in  three  easy  steps’,  ‘Making  Peer  Feedback  work  in  three  easy  steps’,  ‘Using  Generic  Feedback  Effectively’,  ‘Improve  your  students’  performance  in  90  minutes!’,  and  ‘Face-­‐to-­‐face  feedback’;  and  for  students,  ‘Advice  for  students:  Feedback  –  Make  it  work  for  you!’  http://www.brookes.ac.uk/aske/resources/    University  of  Strathclyde:  ‘Feedback  is  a  dialogue’  The  resources  on  this  site  include  the  ‘Feedback  is  a  dialogue  Staff  Leaflet’  and  the  ‘Feedback  is  a  dialogue  Student  Leaflet’.  http://www.strath.ac.uk/learnteach/teaching/staff/assessfeedback/    University  of  Edinburgh:  ‘Enhancing  Feedback’  site  This  site  has  information  for  staff  and  students,  featuring,  for  staff,  ‘Time-­‐friendly  ways  to  boost  feedback’,  and  ‘Ideas,  strategies  and  case  examples’,  among  other  topics.  The  links  in  the  first,  e.g.,  include  those  to  ‘assessment  dialogues’,  ‘engaging  with  criteria  and  standards’,  ‘elective  feedback’,  ‘self-­‐generated  feedback’,  and  ‘the  use  of  exemplars’.  http://www.enhancingfeedback.ed.ac.uk/  See  especially  the  ‘Feedback  in  your  subject’  pages’.  The  links  take  you  to  an  annotated  bibliography  of  relevant  documents  in  various  subjects,  including,  from  Science  and  Engineering,  Biological  Sciences  and  Chemistry;  from  Medicine  and  Veterinary  Medicine,  Medicine;  and  from  Humanities  and  Social  Science,  Psychology.  http://www.enhancingfeedback.ed.ac.uk/staff/subjectfeedback.html    There  is  also  a  useful  feedback  resource  available  at  the  bioassess  website  (Harris,  Krause,  Gleeson,  Peat,  Taylor,  &  Garnett,  2007).    Harris,  K-­‐L.,  Krause,  K.,  Gleeson,  D.,  Peat,  M.,  Taylor,  C.  &  Garnett,  R.  (2007).  Enhancing  Assessment  in  the  Biological  Sciences:  Ideas  and  resources  for  university  educators.  Available  at:  www.bioassess.edu.au    

4.1.11     University  of  Sydney  Feedback  webpages  The  Institute  for  Teaching  and  Learning  website  features,  in  ‘Projects  and  Planning’  –  ‘Assessment  Resources’,  a  webpage  on  Feedback  which  includes  links  to  further  resources  such  as  their  own  ‘Teaching  Insights  –  4’  on  ‘Giving  Feedback’,  consisting  of  a  set  of  guidelines  (e.g.  ‘Clarify  your  assessment  standards’,  ‘Use  exemplars’).  http://www.itl.usyd.edu.au/    The  First  Year  Experience  (FYE)  website  is  a  resource  providing  support  for  coordinators  of  first  year  programs  and  units  of  study.  In  its  ‘Teaching  and  Learning’  pages,  the  site  features  a  ‘Giving  quality  feedback’  section,  the  latter  giving  an  overview  of  the  factors  involved  in  making  feedback  effective  for  learning,  of  methods  for  giving  feedback,  including  innovative  approaches,  and  using  online  feedback.  There  are  also  links  to  related  journal  articles  and  websites,  as  well  as  

  31  

related  University  of  Sydney  links  (e.g.  The  Business  School’s  page  on  feedback  for  students  in  large  classes).  sydney.edu.au/staff/fye    

4.2 Suggested Solutions from the Literature

 In  Section  3  of  this  Handbook,  a  division  was  made  between  ‘University/  Coordinator-­‐centred  Suggestions’  and  ‘Solutions  suggested  by  Markers’.  In  the  following  ‘Suggested  Solutions  –  Guide’,  for  practical  reasons  concerning  overlap,  the  suggestions  relate  to  markers  more  generally,  and  include  those  which  will  depend  on  decisions  and  action  by  Coordinators.    Vardi’s  (2012)  HERDSA  Guide  (Effective  feedback  for  student  learning  in  Higher  Education),  sections  of  which  are  referred  to  in  the  ‘Guide’  below,  gives  much  good  advice  to  Coordinators.  Two  examples  of  this  advice  follow.    Preparing  markers  to  give  feedback  Vardi  argues  that  tutors  ‘often  stand  at  the  heart  of  a  university  teaching  program’,  and  that  ‘students  need  to  be  able  to  trust  their  judgement’  (p.  32).  To  ensure  that  this  trust  is  present,  Vardi  lists  some  useful  ways  of  supporting  staff  ‘to  provide  powerful  feedback’.  These  include:  ‘Induct  your  staff  thoroughly  into  all  aspects  of  the  unit.’  ‘Develop  or  improve  your  performance  expectations  in  collaboration  with  your  teaching  staff.’  ‘Provide  your  tutors  with  a  feedback  tip  sheet  outlining  what  they  need  to  do.’    Evaluating  and  improving  the  effectiveness  of  feedback  Vardi  provides  a  model  for  evaluation  which  includes,  firstly,  ‘whether  or  not  you  put  into  place  the  right  mechanisms  for  a  powerful  context,  with  powerful  feedback  and  responsive  students’  (p.  41),  and  gives  a  table  of  questions  (‘Is  there  …’)  and  a  corresponding  ‘Action  list’.  The  second  part  consists  of  ‘check[ing]  how  well  each  component  is  working’,  and  she  again  provides  a  table  with  ‘Check  …:  and  ‘Collect  data  …’  headings.  The  third  part  of  the  model  concerns  the  relationship  of  the  context,  feedback  and  student  responsiveness  to  institutional  data,  and  the  last,  to  the  improvements  that  can  be  made  and  how  these  could  be  addressed.      

  32  

Table  2:  From  the  literature:  Suggested  Solutions  –  Guide    Sources

Feed

back

pr

actic

e

Exam

ples

of

actu

al

feed

back

re

spon

ses

Stud

ent

enga

gem

ent

with

fe

edba

ck

Dia

logi

c ap

proa

ches

to

feed

back

Usi

ng

tech

nolo

gy

for p

eer a

nd

mar

ker

feed

back

Dea

ling

with

la

rge

clas

ses

Feed

forw

ard

Ballantyne, R., Hugher, K., & Mylonas, A. (2002). Developing procedures for implementing peer assessment in large classes using an action research process. Assessment & Evaluation in Higher Education, 27(5), 427-441. doi: 10.1080/0260293022000009302

Basnet, B., Brodie, L., & Worden, J. (2010). Enhanced Feedback - Does Peer Assessment achieve this goal? Paper presented at the AaeE 2010: Past, Present, Future - the 'Keys' to Engineering Education Research and Practice.

Bridgeman, A., & Rutledge, P. (2010). Getting personal: Feedback for the masses. Synergy (Institute for Teaching and Learning, University of Sydney)(30), 61-68. ✓ ✓ Brown, E. (n.d.). Effective feedback. Formative Assessment in Science Teaching (FAST) Retrieved 22.03.12 ✓ Brown, E., & Glover, C. (2006). Evaluating written feedback. In C. Bryan & K. Clegg (Eds.), Innivative assessment in Higher Education (pp. 81-91). London: Routledge. ✓ ✓ Carless, D. (2006). Differing perceptions in the feedback process. Studies in Higher Education, 31(2), 219-233. doi: 10.1080/03075070600572132 ✓ Carless, D., Salter, D., Yang, M., & Lam, J. (2011). Developing sustainable feedback practices. Studies in Higher Education, 36(4), 395-407. doi: 10.1080/03075071003642449 ✓ ✓ Cartney, P. (2010). Exploring the use of peer assessment as a vehicle for closing the gap between feedback given and feedback used. Assessment & Evaluation in Higher Education, 35(5), 551-564. doi: 10.1080/02602931003632381

Case, S. (2007). Reconfiguring and realigning the assessment feedback processes for an undergraduate criminology degree. Assessment & Evaluation in Higher Education, 32(3), 285-299. doi: 10.1080/02602930600896548

Covic, T., & Jones, M. K. (2008). Is the essay resubmission option a formative or a summative assessment and does it matter as long as the grades improve? Assessment & Evaluation in Higher Education, 33(1), 75-85. doi: 10.1080/02602930601122928

Denton, P., Madden, J., Roberts, M., & Rowe, P. (2008). Students' response to traditional and computer-assisted formative feedback: A comparative case study. British Journal of Educational Technology, 39(3), 486-500. doi: 10.1111/j.1467-8535.2007.00745.x

Duncan, N. (2007). 'Feed-forward': improving students' use of tutors' comments. Assessment & Evaluation in Higher Education, 32(3), 271-283. doi: 10.1080/02602930600896498

  33  

Sources

Feed

back

pr

actic

e

Exam

ples

of

actu

al

feed

back

re

spon

ses

Stud

ent

enga

gem

ent

with

fe

edba

ck

Dia

logi

c ap

proa

ches

to

feed

back

Usi

ng

tech

nolo

gy

for p

eer a

nd

mar

ker

feed

back

Dea

ling

with

la

rge

clas

ses

Feed

forw

ard

Gibbs, G., & Simpson, C. (2004). Conditions Under Which Assessment Supports Students Learning. Learning and Teaching in Higher Education, 1(1), 3-31. ✓ Hattie, J., & Timperley, H. (2007). The power of feedback. Review of Educational Research, 77(1), 81-112. doi: 10.3102/003465430298487 ✓ Hendry, G. D., Bromberger, N., & Armstrong, S. (2011). Constructive guidance and feedback for learning: The usefulness of exemplars, marking sheets and different types of feedback in a first year Law subject. Assessment & Evaluation in Higher Education, 36(1), 1-11. doi: 10.1080/02602930903128904

✓ ✓

Hepplestone, S., Holden, G., Irwin, B., Parkin, H. J., & Thorpe, L. (2011). Using technology to encourage student engagement with feedback: A literature review. Research in Learning Technology, 19(2), 117-127. doi: 10.1080/21567069.2011.586677

Higgins, R., Hartley, P., & Skelton, A. (2001). Getting the message across: The problem of communicating assessment feedback. Teaching in Higher Education, 6(2), 269-274. doi: 10.1080/13562510120045230

✓ ✓

Higgins, R., Hartley, P., & Skelton, A. (2002). The Conscientious Consumer : Reconsidering the role of assessment feedback in student learning. Studies in Higher Education, 27(1), 53-64. doi: 10.1080/03075070120099368

Hounsell, D., McCune, V., Hounsell, J., & Litjens, J. (2008). The quality of guidance and feedback to students. Higher Education Research & Development, 27(1), 55-67. doi: 10.1080/07294360701658765

Lunsford, R. F. (1997). When less is more: Principles for responding in the disciplines. In P. Elbow & M. D. Sorcinelli (Eds.), Writing to learn: Strategies for assigning and responding to writing across the disciplines (pp. 91-104). San Francisco: Jossey-Bass.

Lunt, T., & Curran, J. (2009). ‘Are you listening please?’ The advantages of electronic audio feedback compared to written feedback. Assessment & Evaluation in Higher Education, 35(7), 759-769. doi: 10.1080/02602930902977772

Merry, S., & Orsmond, P. (2008). Students' attitudes to and usage of academic feedback provided via audio files. Bioscience Education, 11. Retrieved from http://www.bioscience.heacademy.ac.uk/journal/vol11/beej-11-3.pdf

  34  

Sources

Feed

back

pr

actic

e

Exam

ples

of

actu

al

feed

back

re

spon

ses

Stud

ent

enga

gem

ent

with

fe

edba

ck

Dia

logi

c ap

proa

ches

to

feed

back

Usi

ng

tech

nolo

gy

for p

eer a

nd

mar

ker

feed

back

Dea

ling

with

la

rge

clas

ses

Feed

forw

ard

Nicol, D. J. (2009). Assessment for learner self-regulation: Enhancing achievement in the first year using learning technologies. Assessment & Evaluation in Higher Education, 34(3), 335-352. doi: 10.1080/02602930802255139 Nicol, D. J. (2010). From monologue to dialogue: improving written feedback processes in mass higher education. Assessment & Evaluation in Higher Education, 35(5), 501-517. doi: 10.1080/02602931003786559

Nicol, D. J. (2010). From monologue to dialogue: improving written feedback processes in mass higher education. Assessment & Evaluation in Higher Education, 35(5), 501-517. doi: 10.1080/02602931003786559

Nicol, D. J. (2011). Developing students' ability to construct feedback. Paper presented at the QAA Enhancement Themes Conference, Heriot-Watt University. ✓ Nicol, D. J., & Macfarlane-Dick, D. (2006). Formative assessment and self‐regulated learning: A model and seven principles of good feedback practice. Studies in Higher Education, 31(2), 199-218. doi: 10.1080/03075070600572090

✓ ✓ ✓

Orsmond, P., Merry, S., & Reiling, K. (2005). Biology students' utilization of tutors' formative feedback: A qualitative interview study. Assessment & Evaluation in Higher Education, 30(4), 369-386. doi: 10.1080/02602930500099177

Pearce, J., Mulder, R., & Baik, C. (2009). Involving students in peer review: Case studies and practical strategies for university teaching. Retrieved from http://www.cshe.unimelb.edu.au/

✓ ✓

Price, M., Handley, K., Millar, J., & O'Donovan, B. (2010). Feedback : all that effort, but what is the effect? Assessment & Evaluation in Higher Education, 35(3), 277-289. doi: 10.1080/02602930903541007

Price, M., O'Donovan, B., & Rust, C. (2007). Putting a social-constructivist assessment process model into practice: Building the feedback loop into the assessment process through peer review. Innovations in Education and Teaching International, 44(2), 143-152.

Providing Helpful Feedback, 2012, e-learning, University of Sydney. ✓ Rotheram, B. (2007). Using an MP3 recorder to give feedback on student assignments. Educational Developments, 8(2), 7-10 ✓ Rotheram, B. (2009). Practice tips on using digital audio for assessment feedback Retrieved 31 January, 2012, from http://www.engsc.ac.uk/events/using-technology-to-enhance-learning-teaching-assessment-and-feedback

Rust, C. (2001). A briefing on assessment of large groups. Generic Centre Assessment Series No. 12, from http://www.bioscience.heacademy.ac.uk/resources/themes/assessment.aspx

  35  

Sources

Feed

back

pra

ctic

e

Exam

ples

of a

ctua

l fe

edba

ck re

spon

ses

Stud

ent e

ngag

emen

t with

fe

edba

ck

Dia

logi

c ap

proa

ches

to

feed

back

Usi

ng te

chno

logy

for p

eer

and

mar

ker f

eedb

ack

Dea

ling

with

larg

e cl

asse

s

Feed

forw

ard

Rust, C., Price, M., & O'Donovan, B. (2003). Improving students' learning by developing their understanding of assessment criteria and processes. Assessment & Evaluation in Higher Education, 28(2), 147-164. doi: 10.1080/02602930301671

Sadler, D. R. (1989). Formative assessment and the design of instructional systems. Instructional Science, 18, 119-144. ✓ Sadler, D. R. (2008). Transforming holistic assessment and grading into a vehicle for complex learning. In G. Joughin (Ed.), Assessment, learning and judgement in higher education (pp. 49-64). Dordrecht: Springer.

Sadler, D. R. (2010). Beyond feedback: Developing student capability in complex appraisal. Assessment & Evaluation in Higher Education, 35(5), 535-550. doi: 10.1080/02602930903541015

Sandhu, S. (2004). The use of statement banks as a means of assessing large groups in Accounting. Investigations in University Teaching and Learning, 2(1), 49-55. ✓ University websites: Sydney University: ITL Assessment Resources: Feedback Leeds Metropolitan University: ‘Sounds Good’ project Nottingham University: ‘Enhancing student feedback through the use of technology’ project Sheffield Hallam University: ‘Technology, Feedback, Action!’ project

Vardi, I. (2012). Effective feedback for student learning in higher education: HERDSA. ✓ ✓ Yorke, M. (2003). Formative assessment in higher education: Moves towards theory and the enhancement of pedagogic practice. Higher Education, 45, 477-501. ✓        

  36  

Table  3:  Suggested  Solutions  –  Feedback  practice    

Sources

Tuto

rs’ a

ctiv

e en

gage

men

t with

as

sess

men

t crit

eria

Goo

d qu

ality

fe

edba

ck fr

om

mar

kers

Impr

oved

crit

eria

sh

eets

Lim

iting

the

num

ber

of fe

edba

ck

com

men

ts

Giv

ing

gene

ric

feed

back

on

the

wor

k of

the

clas

s as

a

who

le

Mul

ti-st

age

assi

gnm

ents

: re

draf

ting

afte

r tu

tor/p

eer f

eedb

ack

Res

ubm

issi

on o

f w

ork

Brown, E., & Glover, C. (2006). Evaluating written feedback. In C. Bryan & K. Clegg (Eds.), Innivative assessment in Higher Education (pp. 81-91). London: Routledge. ✓ Carless, D., Salter, D., Yang, M., & Lam, J. (2011). Developing sustainable feedback practices. Studies in Higher Education, 36(4), 395-407. doi: 10.1080/03075071003642449 ✓ Covic, T., & Jones, M. K. (2008). Is the essay resubmission option a formative or a summative assessment and does it matter as long as the grades improve? Assessment & Evaluation in Higher Education, 33(1), 75-85. doi: 10.1080/02602930601122928

Hendry, G. D., Bromberger, N., & Armstrong, S. (2011). Constructive guidance and feedback for learning: The usefulness of exemplars, marking sheets and different types of feedback in a first year Law subject. Assessment & Evaluation in Higher Education, 36(1), 1-11. doi: 10.1080/02602930903128904

Higgins, R., Hartley, P., & Skelton, A. (2001). Getting the message across: The problem of communicating assessment feedback. Teaching in Higher Education, 6(2), 269-274. doi: 10.1080/13562510120045230 ✓ Lunsford, R. F. (1997). When less is more: Principles for responding in the disciplines. In P. Elbow & M. D. Sorcinelli (Eds.), Writing to learn: Strategies for assigning and responding to writing across the disciplines (pp. 91-104). San Francisco: Jossey-Bass.

Nicol, D. J., & Macfarlane-Dick, D. (2006). Formative assessment and self‐regulated learning: A model and seven principles of good feedback practice. Studies in Higher Education, 31(2), 199-218. doi: 10.1080/03075070600572090

✓ ✓ ✓ ✓

Orsmond, P., Merry, S., & Reiling, K. (2005). Biology students' utilization of tutors' formative feedback: A qualitative interview study. Assessment & Evaluation in Higher Education, 30(4), 369-386. doi: 10.1080/02602930500099177

Rust, C., O'Donovan, B., & Price, M. (2005). A social constructivist assessment process model: how the research literature shows us this could be best practice. Assessment & Evaluation in Higher Education, 30(3), 231-240. doi: 10.1080/02602930500063819

✓ ✓ ✓

     

  37  

Suggested  Solutions  –  Examples  of  feedback  responses  markers  can  use    Providing Helpful Feedback, 2012, e-learning, University of Sydney (Dr Meloni Muir) Examples of helpful feedback comments on Psychology essays (Dr Fiona Hibberd) Vardi, I. (2012). Effective feedback for student learning in higher education: HERDSA.    Table  4:  Suggested  Solutions  –    Student  engagement  with  feedback    

Sources

Stud

ent r

eque

sts

for k

inds

of

feed

back

Stud

ent i

dent

ifica

tion

of a

reas

of

diffi

culty

Stud

ent s

elf-a

sses

smen

t bef

ore

wor

k su

bmitt

ed

Giv

ing

feed

back

but

no

grad

e w

hen

wor

k in

itial

ly re

turn

ed

Stud

ents

as

mar

kers

: exe

mpl

ars

and/

or p

eer a

sses

smen

t/fee

dbac

k

Dev

elop

ing

stud

ents

’ abi

lity

to s

elf-

asse

ss

(Impr

ovem

ent i

n pe

rfor

man

ce a

fter

peer

mar

king

exe

rcis

e)

Hel

ping

cha

nges

in fe

edba

ck

prac

tice

to w

ork

Basnet, B., Brodie, L., & Worden, J. (2010). Enhanced Feedback - Does Peer Assessment achieve this goal? Paper presented at the AaeE 2010: Past, Present, Future - the 'Keys' to Engineering Education Research and Practice.

Cartney, P. (2010). Exploring the use of peer assessment as a vehicle for closing the gap between feedback given and feedback used. Assessment & Evaluation in Higher Education, 35(5), 551-564. doi: 10.1080/02602931003632381

Gibbs, G., & Simpson, C. (2004). Conditions Under Which Assessment Supports Students ’ Learning. Learning and Teaching in Higher Education, 1(1), 3-31.

Hendry, G. D., Bromberger, N., & Armstrong, S. (2011). Constructive guidance and feedback for learning: The usefulness of exemplars, marking sheets and different types of feedback in a first year Law subject. Assessment & Evaluation in Higher Education, 36(1), 1-11. doi: 10.1080/02602930903128904

Nicol, D. J., & Macfarlane-Dick, D. (2006). Formative assessment and self‐regulated learning: A model and seven principles of good feedback practice. Studies in Higher Education, 31(2), 199-218. doi: 10.1080/03075070600572090

✓ ✓ ✓ ✓

Nicol, D. J. (2011). Developing students' ability to construct feedback. Paper presented at the QAA Enhancement Themes Conference, Heriot-Watt University.

  38  

Pearce, J., Mulder, R., & Baik, C. (2009). Involving students in peer review: Case studies and practical strategies for university teaching. Retrieved from http://www.cshe.unimelb.edu.au/

Price, M., O'Donovan, B., & Rust, C. (2007). Putting a social-constructivist assessment process model into practice: Building the feedback loop into the assessment process through peer review. Innovations in Education and Teaching International, 44(2), 143-152.

Rust, C., O'Donovan, B., & Price, M. (2005). A social constructivist assessment process model: how the research literature shows us this could be best practice. Assessment & Evaluation in Higher Education, 30(3), 231-240. doi: 10.1080/02602930500063819

✓ ✓ ✓

Rust, C., Price, M., & O'Donovan, B. (2003). Improving students' learning by developing their understanding of assessment criteria and processes. Assessment & Evaluation in Higher Education, 28(2), 147-164. doi: 10.1080/02602930301671

Sadler, D. R. (1989). Formative assessment and the design of instructional systems. Instructional Science, 18, 119-144. ✓ Sadler, D. R. (2008). Transforming holistic assessment and grading into a vehicle for complex learning. In G. Joughin (Ed.), Assessment, learning and judgement in higher education (pp. 49-64). Dordrecht: Springer. ✓ Sadler, D. R. (2010). Beyond feedback: Developing student capability in complex appraisal. Assessment & Evaluation in Higher Education, 35(5), 535-550. doi: 10.1080/02602930903541015 ✓ Vardi, I. (2012). Effective feedback for student learning in higher education: HERDSA. ✓ ✓

Yorke, M. (2003). Formative assessment in higher education: Moves towards theory and the enhancement of pedagogic practice. Higher Education, 45, 477-501.

         Suggested  Solutions  –  Dialogic  approaches  to  feedback   Carless, D. (2006). Differing perceptions in the feedback process. Studies in Higher Education, 31(2), 219-233. doi: 10.1080/03075070600572132 Carless, D., Salter, D., Yang, M., & Lam, J. (2011). Developing sustainable feedback practices. Studies in Higher Education, 36(4), 395-407. doi:

10.1080/03075071003642449 Nicol, D. J. (2010). From monologue to dialogue: improving written feedback processes in mass higher education. Assessment & Evaluation in Higher Education,

35(5), 501-517. doi: 10.1080/02602931003786559 Nicol, D. J., & Macfarlane-Dick, D. (2006). Formative assessment and self‐regulated learning: A model and seven principles of good feedback practice. Studies in

Higher Education, 31(2), 199-218. doi: 10.1080/03075070600572090

  39  

Table  5:  Suggested  Solutions  –  Using  technology  to  provide  peer  and  marker  feedback   Sources

Onl

ine

lear

ning

task

s

Use

of a

udio

file

s

Peer

revi

ew u

sing

onl

ine

tool

s

Com

pute

r ass

iste

d fe

edba

ck,

incl

udin

g st

atem

ent b

anks

Rel

ated

Uni

vers

ity w

ebsi

tes

Bridgeman, A., & Rutledge, P. (2010). Getting personal: Feedback for the masses. Synergy (Institute for Teaching and Learning, University of Sydney)(30), 61-68. ✓ Case, S. (2007). Reconfiguring and realigning the assessment feedback processes for an undergraduate criminology degree. Assessment & Evaluation in Higher Education, 32(3), 285-299. doi: 10.1080/02602930600896548 ✓ Denton, P., Madden, J., Roberts, M., & Rowe, P. (2008). Students' response to traditional and computer-assisted formative feedback: A comparative case study. British Journal of Educational Technology, 39(3), 486-500. doi: 10.1111/j.1467-8535.2007.00745.x

Hepplestone, S., Holden, G., Irwin, B., Parkin, H. J., & Thorpe, L. (2011). Using technology to encourage student engagement with feedback: A literature review. Research in Learning Technology, 19(2), 117-127. doi: 10.1080/21567069.2011.586677 ✓ Lunt, T., & Curren, J. C. (2009). ‘Are you listening please?’ The advantages of electronic audio feedback compared to written feedback. Assessment & Evaluation in Higher Education, 35(7), 759-769. doi: 10.1080/02602930902977772 ✓ Merry, S., & Orsmond, P. (2008). Students' attitudes to and usage of academic feedback provided via audio files. Bioscience Education, 11. Retrieved from http://www.bioscience.heacademy.ac.uk/journal/vol11/beej-11-3.pdf ✓ Nicol, D. J. (2009). Assessment for learner self-regulation: Enhancing achievement in the first year using learning technologies. Assessment & Evaluation in Higher Education, 34(3), 335-352. doi: 10.1080/02602930802255139 ✓ Pearce, J., Mulder, R., & Baik, C. (2009). Involving students in peer review: Case studies and practical strategies for university teaching. Retrieved from http://www.cshe.unimelb.edu.au/ ✓ Rotheram, B. (2007). Using an MP3 recorder to give feedback on student assignments. Educational Developments, 8(2), 7-10. ✓ Rotheram, B. (2009). Practice tips on using digital audio for assessment feedback Retrieved 31 January, 2012, from http://www.engsc.ac.uk/events/using-technology-to-enhance-learning-teaching-assessment-and-feedback ✓ University websites: Leeds Metropolitan University: ‘Sounds Good’ project Nottingham University: ‘Enhancing student feedback through the use of technology’ project Sheffield Hallam University: ‘Technology, Feedback, Action!’ project

  40  

 Table  6:  Suggested  Solutions  –  Dealing  with  large  classes    Sources Feedback with

large classes

Peer assessment in large classes

Use of statement banks with large classes

Ballantyne, R., Hugher, K., & Mylonas, A. (2002). Developing procedures for implementing peer assessment in large classes using an action research process. Assessment & Evaluation in Higher Education, 27(5), 427-441. doi: 10.1080/0260293022000009302

Bridgeman, A., & Rutledge, P. (2010). Getting personal: Feedback for the masses. Synergy (Institute for Teaching and Learning, University of Sydney)(30), 61-68. ✓ Rust, C. (2001). A briefing on assessment of large groups. Generic Centre Assessment Series No. 12, from http://www.bioscience.heacademy.ac.uk/resources/themes/assessment.aspx

Sandhu, S. (2004). The use of statement banks as a means of assessing large groups in Accounting. Investigations in University Teaching and Learning, 2(1), 49-55.

   Suggested  Solutions  –  Feedforward   Brown, E. (n.d.). Effective feedback. Formative Assessment in Science Teaching (FAST) Retrieved 22.03.12, 2012 Brown, E., & Glover, C. (2006). Evaluating written feedback. In C. Bryan & K. Clegg (Eds.), Innivative assessment in Higher Education (pp. 81-91). London: Routledge Duncan, N. (2007). 'Feed-forward': improving students' use of tutors' comments. Assessment & Evaluation in Higher Education, 32(3), 271-283. doi:

10.1080/02602930600896498 Hattie, J., & Timperley, H. (2007). The power of feedback. Review of Educational Research, 77(1), 81-112. doi: 10.3102/003465430298487 Higgins, R., Hartley, P., & Skelton, A. (2001). Getting the message across: The problem of communicating assessment feedback. Teaching in Higher Education, 6(2),

269-274. doi: 10.1080/13562510120045230 Hounsell, D., McCune, V., Hounsell, J., & Litjens, J. (2008). The quality of guidance and feedback to students. Higher Education Research & Development, 27(1), 55-

67. doi: 10.1080/07294360701658765 Price, M., Handley, K., Millar, J., & O'Donovan, B. (2010). Feedback : all that effort, but what is the effect? Assessment & Evaluation in Higher Education, 35(3), 277-

289. doi: 10.1080/02602930903541007  

  41  

4.2.1   Feedback  practice    Key  Points  

• Tutors’  active  engagement  with  assessment  criteria  • Good  quality  feedback  from  markers  • Improved  criteria  sheets  • Limiting  the  number  of  feedback  comments  • Giving  generic  feedback  on  the  work  of  the  class  as  a  whole  • Two-­‐stage  (or  more)  assignments  involving  redrafting  after  tutor  or  peer  

feedback  • Resubmission  of  work  

 Tutors’  active  engagement  with  assessment  criteria  For  Rust,  O’Donovan  and  Price  (2005,  p.  233),  active  engagement  with  assessment  criteria  by  both  markers  and  students  is  the  key  to  ‘achieving  meaningful  understanding  of  assessment’.  For  tutors,  the  process  takes  the  form  of  discussion  of  the  criteria,  followed  by  some  marking  using  the  criteria,  then  further  discussion,  and  then  discussion  in  moderation  sessions  after  the  completion  of  marking.  (See  the  two  (staff  and  student)  parallel,  interrelated  and  dynamic  cycles  of  desirable  practice.)    Good  quality  feedback  from  markers  According  to  Nicol  and  Macfarlane-­‐Dick  (2006),  there  is  little  consensus  in  the  research  literature  about  what  good  quality  (external,  as  opposed  to  students’  own  internal)  feedback  is.  They  define  it  as  ‘information  that  helps  students  troubleshoot  their  own  performance  and  self-­‐correct:  that  is,  it  helps  students  take  action  to  reduce  the  discrepancy  between  their  intentions  and  the  resulting  effects’  (p.  208).    Higgins,  Hartley  and  Skelton  (2002,  p.  62)  offer  the  following  advice  to  tutors/  coordinators:  ‘Firstly,  …  timely  feedback  is  vital;  …  interim  feedback  on  a  first  draft  or  an  essay  plan  might  also  be  productive’.  Secondly,  …  misconceptions  need  to  be  explained  and  improvements  for  future  work  suggested.’  The  authors  explain  that  students  ‘may  not  view  comments  on  “surface”  aspects  of  their  work  as  particularly  relevant  or  useful’,  and  that  ‘fostering  “higher  order”  critical  skills  may  have  more  long-­‐term  educational  value’.    See  also  Brown  and  Glover  (2006)  below.    Improved  criteria  sheets  One  of  the  strategies  suggested  by  Nicol  and  Macfarlane-­‐Dick  (2006)  which  have  ‘proved  effective  in  clarifying  criteria,  standards  and  goals’  for  students  is  the  provision  of  ‘better  definitions  of  requirements  using  carefully  constructed  criteria  sheets  and  performance-­‐level  definitions’  (p.  207).    Limiting  the  number  of  feedback  comments  Lunsford  (1997)  examines  the  written  feedback  comments  on  student  writing  of  12  experienced  composition  theory  markers,  and  presents  guidelines  for  responding  to  this  writing.  He  refers  to  his  previously  published  ‘seven  principles  for  response’,  which  include  the  writing  of  ‘well-­‐developed  and  text-­‐specific  comments’,  a  focus  on  ‘global,  not  local,  concerns’,  and  the  framing  of  comments  in  ‘nonauthoritative  modes’,  and  categorises  these  principles  in  terms  of  ‘Development  and  Specificity’,  ‘Purposeful  Commenting’,  ‘Correctness’,  and  ‘Extra-­‐textual  Response’  (pp.  91-­‐92).  

  42  

Under  ‘Purposeful  Commenting’,  Lunsford  explains  that  ‘less  is  more’  if  students  are  going  to  improve  their  writing.  The  markers  studied  averaged  ‘just  over  three  issues  per  paper  marked’,  i.e.  they  ‘arrive[d]  at  a  set  of  themes  for  their  responses’.  They  marked  two  or  three  examples  of  the  focus  problem,  and  provided  ‘one  reasonably  fully  developed  comment  [which]  might  be  two  or  three  sentences  long,  allowing  the  teacher  to  give  the  student  a  mini-­‐lesson  on  the  issue’  (1997,  p.  93).  He  also  develops  the  idea  of  the  making  of  nonauthoritative  comments  and  of  comments  which  respond  to  what  the  student  writers  have  to  say,  rather  than  to  errors  or  form  more  generally.    The  results  of  the  small-­‐scale  study  by  Hendry,  Bromberger  and  Armstrong  (2011)  (see  below)  show  that  students  perceive  feedback  in  the  form  of  both  ‘a  limited  number  of  personalised,  concrete  comments’  focussing  on  performance  improvement  and  a  ‘standards-­‐based,  “checklist-­‐like”’  type,  as  helping  them  ‘target  their  learning  efforts  effectively’  (p.  9).    Orsmond,  Merry  and  Reiling  (2005,  p.  370)  report,  from  a  study  by  Brockbank  and  McGill  (1998),  that  ‘the  quantity  of  both  positive  and  negative  feedback  may  overwhelm  the  student  so  that  they  are  unable  to  take  it  in  overall’.        Giving  generic  feedback  on  the  work  of  the  class  as  a  whole  Rust,  O’Donovan  and  Price  (2005,  p.  235)  suggest  that  giving  only  generic  feedback  which  ‘highlight[s]  common  mistakes,  things  that  were  generally  done  well,  etc.’  means  a  short  time  lag  between  submission  and  feedback.  Students  can  then  be  encouraged  to  ‘try  and  identify  which  of  the  generic  feedback  applies  to  them,  and  from  that  identify  what  they  need  to  [do]  differently  next  time’.  They  add  that  this  approach  could  be  linked  to  initial  withholding  of  the  grade.    Two-­‐stage  (or  more)  assignments  involving  redrafting  after  tutor  or  peer  feedback  Rust,  O’Donovan  and  Price  (2005)  presents  the  authors’  ‘social,  constructivist  approach’  to  assessment,  in  which  active  engagement  with  assessment  criteria  by  both  markers  and  students  is  the  key  to  ‘achieving  meaningful  understanding  of  assessment’  (p.  233).  The  authors  argue  that  one  way  of  achieving  active  engagement  with  feedback  on  the  part  of  students  is  two-­‐stage  (or  more)  assignments,  i.e.  those  with  a  first  draft  and  then  a  redraft  after  tutor  or  peer  feedback.  The  example  they  mention  involves:  peer  discussion  of  feedback  received  using  a  checklist;  then  writing  about  their  feedback  using  a  pro-­‐forma,  this  writing  including  ‘deriv[ing]  conclusions  about  what  to  do  differently  in  future’;  and  benchmarking  their  work  against  an  ‘A’  grade  assignment  (p.  235).    The  assessment  design  strategy  which  Carless  et  al.  (2011,  p.  398)  see  as  encouraging  dialogic  feedback  involves  two-­‐stage  or  multi-­‐stage  assignments:  ‘Such  assignments  facilitate  sustainable  feedback  when  required  standards  are  becoming  increasingly  transparent  and  dialogic  feedback  processes  support  students  in  self-­‐monitoring  their  work  while  it  is  being  developed’.    Nicol  and  Macfarlane-­‐Dick  (2006,  p.  213)  view  closing  the  ‘gap  between  current  performance  and  the  performance  expected  by  the  teacher’  as  being,  in  part,  about  supporting  the  act  of  production  of  a  piece  of  work:  this  requires  ‘the  generation  of  concurrent  or  intrinsic  feedback  that  students  can  interact  with  while  engaged  in  an  

  43  

assessment  task’.  This  feedback  could  come  from  peers  in  a  group  task,  or  the  task  might  have  a  series  of  components  which  each  led  to  feedback,  or  the  sub-­‐task  feedback  might  be  ‘automatically  generated’  (i.e.  electronically)  to  support  student  engagement.    Resubmission  of  work  Nicol  and  Macfarlane-­‐Dick  (2006,  p.  213)  argue  that  closing  the  learning  gap  (Sadler,  1989)  is,  in  part,  about  ‘providing  opportunities  to  repeat  the  same  “task-­‐performance-­‐external  feedback  cycle”  by,  for  example,  allowing  resubmission  …  (Boud,  2000)’.      Covic  and  Jones’  (2008)  study  provided  a  group  of  54  students  with  the  opportunity  to  submit  an  essay  and  receive  ‘detailed  written  feedback’  as  well  as  a  grade,  and  to  have  the  option  of  modifying  the  essay  according  to  this  feedback  and  to  resubmit  it  at  the  end  of  two  weeks.  26  students  resubmitted  their  essays,  and  25  of  these  essays  received  higher  marks,  one  receiving  the  same  mark.  12  of  those  who  did  not  resubmit  cited  ‘workload  and  time  constraints’  as  the  reason,  while  9  were  ‘happy  with  their  mark’  on  the  first  submission.  Nine  students  provided  comments  on  the  feedback  received,  and  all  comments  were  positive.  However,  the  findings  suggest  that  ‘for  a  number  of  students,  the  resubmission  option  was  viewed  as  a  ‘safety  net’  rather  than  an  opportunity  to  learn  and  apply  that  knowledge  to  improving  the  essay  through  resubmission’  (p.  82),  while  the  assessment  strategy  was  a  time-­‐consuming  exercise  for  the  marker.  The  authors  suggest  that  

meaningful  formative  feedback  can  be  integrated  into  assessment  without  too  much  additional  time  investment,  for  example,  breaking  up  the  essay  into  separately  marked  components,  such  as  …  an  introduction  section  that  with  feedback  will  inform,  if  utilized,  the  rest  of  the  essay.  (p.  84)  

 

4.2.2   Examples  of  feedback  responses  markers  can  use    In  Physiology,  Dr  Meloni  Muir  was  the  Project  Leader  in  the  production  of  the  Providing  Helpful  Feedback  website  (available  via  Blackboard).  The  project  involved  collaboration  between  the  University  of  Sydney’s  discipline  of  Physiology  and  e-­‐learning.  The  site  contains  advice,  examples  and  exercises  for  markers  on  giving  feedback  on  authentic  written  reports.  Currently  access  is  limited  to  markers  in  Physiology;  however,  it  is  possible  that  this  resource  could  be  made  available  to  markers  who  need  to  give  feedback  on  reports  in  the  Faculty  of  Science.  An  example  of  the  way  the  site  works  follows.  In  the  ‘Introduction’  section  (other  sections  include  ‘Method’,  ‘Results’,  and  ‘Discussion’),  Sample  1  features  a  student  text  with  a  button  stating  ‘unhelpful  feedback’,  and  an  instruction  ‘To  reveal  feedback  comments  click  on  each  arrow  in  turn’.  The  unhelpful  comments  revealed,  directly  related  to  highlighted  sections  of  the  Introduction  text  are  ✓  Watch  your  spacing  Incorrect  grammar  This  section  well  done  but  justify  predictions!  By  clicking  on  a  ‘show  info’  button,  the  marker  receives  an  explanation  as  to  why  the  feedback  is  unhelpful,  and  also  examples  of  helpful  feedback.  For  example,  the  unhelpfulness  of  ✓  is  explained  thus:  Tick  marks  are  ambiguous;  need  to  specify  what  the  tick  means/refers  to.  and  the  helpful  feedback  comment  is    

  44  

Overall  Intro  is  well  written  &  paragraphs  are  well  organised.  The  unhelpfulness  of  Incorrect  grammar  is  explained  as  Punctuation  errors  need  rule  explained  and  correction  suggested.  and  the  helpful  feedback  comment  is  A  semi-­‐colon  is  required  before  “however”  in  the  middle  of  a  sentence.    The  unhelpfulness  of  This  section  well  done  but  justify  predictions!  is  explained  as  Need  to  explain/identify  what  has  been  done  well  and  to  spell  out  what  is  missing  and  the  helpful  feedback  comment  is  Good:  aim  has  been  clearly  distinguished  from  hypotheses.  However  insufficient  background  in  the  Intro  to  support  predictions  made  in  hypotheses,  e.g.  relationship  between  PCO2  and  drive  to  breathe.    In  Psychology,  Dr  Fiona  Hibberd  (interviewed  for  this  project),  as  Coordinator,  provides  her  markers  with  a  list  of  sample  comments  to  help  them  give  their  students  constructive  and  tactful/sensitive  feedback.  She  specifies  that  the  aim  ‘is  to  provide  appropriate,  legible  and  careful  brief  comments  that  make  clear  the  strengths  and  weaknesses  of  each  piece  of  work  in  sufficient  detail  to  allow  the  student  to  understand  the  grounds  for  the  mark  they  received  and  improve  their  performance  on  subsequent  assignments’  (from  her  document  ‘Providing  feedback  comments  on  written  work’).  Examples  of  these  comments  are:  The  Introduction  to  this  essay  is  excellent  (or  could  be  improved).  It  provides  (fails  to  adequately  provide)  some  background  to  the  topic,  and  explains  to  the  reader  the  scope  of  the  topics  to  be  covered  in  the  essay.  This  is  an  excellent  point  in  that  …  The  point  you  make  is  flawed  in  that  …  The  essay  is  well  structured  and  flows  logically  and  smoothly.  The  wording  of  this  paragraph  could  be  improved  by  …  This  section  is  vague  and  general  –  aim  to  be  precise  and  to  be  more  specific.    Vardi’s  (2012)  HERDSA  Guide  (Effective  feedback  for  student  learning  in  Higher  Education)  provides  helpful  examples  of  immediate  and  delayed  (written)  feedback.  An  example  of  immediate  –  evaluative,  instructive,  and  encouraging,  affirming  and  motivating  –  feedback  responses  (in  lecture,  tutorial,  laboratory  or  fieldwork  situations)  is:  That  summarises  the  facts  of  the  case  well.  Now,  how  could  you  analyse  that?  (p.  26).  The  delayed  feedback  example  (written)  responses  are  divided  by  Vardi  into  ‘Direct,  prescriptive  and  specific’  and  ‘Direct,  explanatory  and  specific’.  An  example  of  the  first  is:  This  called  for  the  use  of  the  X  method  of  analysis  and  of  the  second:  That  presupposes  …    However,  if  one  takes  into  account  …  (p.  28).    Vardi  also  provides  examples  of  ‘global’  and  ‘local’  feedback,  describing  the  first  as  ‘a  reflection  on  the  assignment  as  a  whole’,  as  including  the  holistic  mark/grade,  and  as  justifying  the  mark/grade  given  ‘by  highlighting  the  performance  levels  achieved’  (p.  29).  Vardi  comments  that  ‘Simply  augmenting  your  global  evaluation  with  the  types  of  global  feedback  listed  has  the  potential  to  significantly  impact  on  future  student  learning  and  achievement  in  subsequent  related  assessment  tasks.’  An  example  is:  ‘2.  Explain  what  they  specifically  needed  to  have  done  in  order  to  achieve  the  next  higher  level  of  performance.  

  45  

In  order  to  achieve  a  Distinction,  you  needed  to  address  and  research  …  in  the  question.  This  will  help  you  deepen  your  analysis  by  …  and  interpret  this  in  relation  to  …  You  also  need  to  address  …’  Local  feedback  is  that  provided  ‘at  specific  points  in  the  assignment’.  Vardi  recommends  that  this  be  of  a  ‘limited  amount’.  (p.  31).  She  comments  that  ‘students  like  to  have  some  indication  that  you  have  read  their  work,  and  would  like  to  know  where  they  were  on  track  and  where  they  faltered’.  She  then  provides  a  list  of  ‘powerful  local  feedback’.  An  example  is:  ‘2.  Focus  on  the  conclusions  drawn  in  each  section.  This  can  significantly  impact  on  the  level  of  critical  thought  applied.  You  would  benefit  from  a  conclusion  to  this  section  that  compares  …  This  conclusion  neatly  draws  together  …  However,  it  omits  …’    

4.2.3   Student  engagement  with  feedback    Key  points  

• Student  requests  for  the  kinds  of  feedback  they  would  like  when  they  hand  in  work  

• Student  identification  of  where  they  are  having  difficulties  when  they  hand  in  assessed  work  

• Student  self-­‐assessment  before  work  submitted  • Developing  students’  ability  to  self-­‐assess  • Giving  students  feedback  but  no  grade  when  the  work  is  initially  returned  • Students  as  markers,  using  exemplars  and/or  peer  assessment/feedback  • (Demonstrating  improvement  in  performance  as  a  result  of  a  peer  marking  

exercise)  • Helping  changes  in  feedback  practice  (to  a  more  student-­‐centred  model)  to  

work    Student  requests  for  the  kinds  of  feedback  they  would  like  when  they  hand  in  work  This  is  one  of  the  strategies  to  improve  student  self-­‐regulation  of  learning  suggested  by  Nicol  and  Macfarlane-­‐Dick  (2006).    Student  identification  of  where  they  are  having  difficulties  when  they  hand  in  assessed  work  This  is  one  of  the  strategies  to  improve  student  self-­‐regulation  of  learning  suggested  by  Nicol  and  Macfarlane-­‐Dick  (2006).    Student  self-­‐assessment  before  work  submitted  Nicol  and  Macfarlane-­‐Dick  (2006,  p.  208)  suggest  student  identification  of  the  strengths  and  weaknesses  in  their  own  work  against  criteria  or  standards  before  submission  as  an  example  of  ‘structured  reflection  and  self-­‐assessment’.  They  also  refer  to  research  showing  that  ‘self-­‐assessment  can  lead  to  significant  enhancements  in  learning  and  achievement’  (p.  207).  One  study  by  Taras  (2001,  2002,  2003),  e.g.,  revealed  that  ‘self-­‐assessment  with  integrated  tutor  feedback  helped  students  identify  and  correct  more  errors  …  than  self-­‐assessment  prior  to  peer  or  tutor  feedback’.      Developing  students’  ability  to  self-­‐assess  

  46  

Vardi’s  (2012)  (HERDSA)  Guide  (Effective  feedback  for  student  learning  in  Higher  Education)  builds  on  the  peer  feedback  section  to  focus  on  ‘Developing  students’  ability  to  self-­‐assess’.  She  argues  for  the  incorporation  of  self-­‐assessment  into  learning  programs  by  asking  students  to,  e.g.  ‘1.  Identify  what  they  have  learnt  or  understood.  What  did  you  learn  from  …?  2.  Identify  the  gaps  in  learning.  What  are  you  unsure  about?  What  haven’t  you  been  able  to  do?    …  4.  Compare  what  they  have  done  to  what  experts  do  or  successful  models.  Compare  your  work  with  the  exemplar.  5.  Set  goal  or  plan  for  how  they  will  improve.  Where  do  you  need  to  improve  …?  What  are  you  going  to  do  about  this?    Giving  students  feedback  but  no  grade  when  the  work  is  initially  returned  For  Rust,  O’Donovan  and  Price  (2005),  active  engagement  with  assessment  criteria  by  both  markers  and  students  is  the  key  to  ‘achieving  meaningful  understanding  of  assessment’  (p.  233).  The  authors  argue  that  one  way  of  achieving  active  engagement  with  feedback  on  the  part  of  students  is  that  they  be  given  comments  only  on  their  returned  work,  and  be  asked  to  ‘identify  from  reading  the  comments  what  grade  they  think  …  has  been  awarded’,  the  actual  grade  being  communicated  only  after  they  do  this  (p.  235).    Students  as  markers,  using  exemplars  and/or  peer  assessment/feedback  Rust,  Price  and  O’Donovan  (2003)  report  on  a  structured  intervention  with  a  group  of  300+  first  year  undergraduates  in  a  Business  module,  ‘an  intervention  that  placed  increased  emphasis  on  the  tacit  knowledge  transfer  processes  of  practice  and  imitation  to  achieve  shared  understanding’  (p.  152).  The  students,  whose  attendance  at  the  workshop  was  voluntary,  engaged  in  activities  involving  ‘discussion  of  exemplars  and  marking  criteria,  marking  exercises  and  self-­‐assessment’.  The  study  found  that  workshop  participants  achieved  ‘significantly  better  results’  in  their  assessed  coursework  than  their  non-­‐participant  peers,  and  one  year  later,  were  still  showing  significantly  better  results  (p.  156).    Rust,  O’Donovan  and  Price  (2005)  argue  that  active  engagement  with  assessment  criteria  by  both  markers  and  students  is  the  key  to  ‘achieving  meaningful  understanding  of  assessment’  (p.  233).  For  students,  this  takes  the  form  of  involvement  in  a  marking  exercise  in  which  they  use  the  criteria  to  mark  sample  texts  and  then  discuss  their  evaluations  with  peers  and  tutors,  and/or  of  peer  marking  of  student  work.  See  below  for  discussion  of  two-­‐stage  assignments.    Sadler  (1989)  (see  above)  contends  that  criteria  sheets  do  not  necessarily  deal  adequately  with  the  qualities  of  a  piece  of  work  because  there  are  both  manifest  and  latent  criteria  involved  in  evaluation,  and  that  students  need  to  develop  their  knowledge  of  both  in  order  to  be  able  to  monitor  their  own  performance,  i.e.  they  need  to  ‘be  given  adequate  evaluative  experience  themselves’  (p.  134)  to  develop  ‘an  appropriate  body  of  tacit  knowledge’  (see  Glossary).  This  can  be  done  by  engaging  them  in  ‘evaluative  and  corrective  activity  on  other  students’  work’  (p.  140).    Sadler  (2008)  argues  for  a  student  induction  process  in  the  appraisal  of  ‘open-­‐response’  assignments  in  order  to  develop  students’  evaluative  expertise  and  enable  them  to  self-­‐monitor  the  development  of  their  own  assignments.  He  specifies  that  they  need  to  develop  a  concept  of  what  constitutes  quality,  and  that  this  can  be  done  

  47  

by  making  evaluative  activity  ‘the  primary  pedagogical  vehicle  for  teaching  a  considerable  proportion  of  the  substantive  content  of  a  course’  (p.  59).    Sadler  (2010)  points  to  a  gap  between  marker  feedback  and  student  ‘appreciation  of  its  practical  import’  (p.  541),  and  suggests  that  learners  be  provided  with  teacher-­‐like  assignment  appraisal  experience.  He  advocates  the  use  of  ‘purposeful  peer  assessment’  as  part  of  the  teaching  design,  and  identifies  ‘task  compliance,  quality  and  criteria’  as  making  up  ‘the  explicit  agenda  to  be  addressed’  (p.  542).      Nicol  and  Macfarlane-­‐Dick  (2006),  referring  to  Rust  et  al.  (2003)  and  other  studies,  report  on  the  difficulty  of  making  assessment  standards  and  criteria  explicit  to  students  using  verbal  descriptions:  ‘Most  criteria  for  academic  tasks  are  complex,  multidimensional  (Sadler,  1989)  and  difficult  to  articulate:  they  are  often  “tacit”  and  unarticulated  in  the  mind  of  the  teacher’  (p.  206).  Referring  to  Orsmond,  Merry  and  Reiling  (2002),  they  argue  that  providing  students  with  exemplars  of  performance  is  an  approach  which  ‘has  proved  particularly  powerful  in  clarifying  goals  and  standards’  (p.  206):  ‘exemplars  are  effective  because  they  make  explicit  what  is  required,  and  they  define  a  valid  standard  against  which  students  can  compare  their  work’.  The  authors  also  suggest  the  involvement  of  students  in  peer-­‐assessment  exercises  in  relation  to  ‘defined  criteria  and  standards’.  Providing  students  with  opportunities  to  evaluate  and  provide  feedback  on  the  work  of  peers  ‘help[s]  develop  the  skills  needed  to  make  objective  judgements  against  standards,  skills  which  are  transferred  when  students  turn  to  producing  and  regulating  their  own  work  (Boud  et  al,.  1999;  Gibbs,  1999)’  (p.  208).    Hendry,  Bromberger,  &  Armstrong  (2011)  report  on  a  small  study  exploring  first  year  Law  students’  ‘perceptions  of  the  usefulness  of  exemplars  and  different  types  of  feedback  for  guiding  them  in  completing  assessments’  (p.  1).  Their  results  show  ‘clearly’  the  greater  usefulness  to  students  of  the  marking  and  discussion  of  exemplars  in  class  compared  to  the  usefulness  of  a  marking  sheet.  For  Hendry  et  al.,  exemplars  provide  ‘concrete  illustrations  of  the  style  of  language  and  structure  expected  and  not  expected  in  a  good  assignment’,  while  the  marking  sheet  is  a  more  ‘abstract  representation  of  teacher  expectations’  (p.  8).  Their  results  also  show  that  students  perceive  feedback  in  the  form  of  both  ‘a  limited  number  of  personalised,  concrete  comments’  focussing  on  performance  improvement  and  a  ‘standards-­‐based,  “checklist-­‐like”’  type,  as  helping  them  ‘target  their  learning  efforts  effectively’  (p.  9).    Pearce,  Mulder  and  Baik  (2009),  in  their  study  of  peer  review,  address  three  questions:  ‘What  is  student  peer  review?  Why  incorporate  student  peer  review  in  subjects/courses?  How  can  student  peer  review  be  implemented  effectively  in  subjects/courses?’  (p.  2).  They  provide  an  overview  of  the  literature,  focussing  on  the  benefits  and  potential  impediments  to  implementation  of  peer  review,  and  present  four  case  studies  of  how  peer  review  was  implemented  in  teaching,  including  in  Departments  of  Zoology  and  Information  Systems.  In  each  case,  student  perceptions  of  the  peer  review  process  are  summarised,  and  there  is  a  ‘Lessons  learned’  concluding  section.  A  ‘reflective  discussion’  of  ‘two  academics’  experiences  in  facilitating  and  implementing  peer  review’  and  an  examination  of  students’  perceptions  of  the  experience  follow.  The  student  evaluations  suggest  that  ‘peer  review  ‘is  generally  perceived  to  be  effective  in  improving  learning’  (p.  17).  The  authors  then  provide  some  ‘key  considerations  in  implementing  student  peer  review’,  such  as  ‘Dealing  with  variation  in  review  quality’,  ‘Disparities  between  reviewer  feedback  and  assessment  outcomes’,  and  ‘Plagiarism’.  

  48  

 Nicol  (2011,  p.  2)  makes  a  ‘case’  for  peer  feedback,  arguing  that  it  ‘can  add  significantly  to  the  amount  and  variety  of  feedback  students  receive,  without  a  corresponding  increase  in  teacher  workload’,  and  can  also  ‘address  the  timeliness  issue’.  He  explains  that  students  ‘cannot  avoid  engagement  [with  feedback]  if  they  are  required  to  produce  commentaries  on  the  work  of  others’,  that  is,  they  have  to  ‘actively  engage  with  assessment  criteria  and  standards’.  They  also  ‘learn  about  different  approaches  and  tactics  to  the  same  assignment’,  ‘realise  that  quality  is  not  a  fixed  attribute  but  can  be  produced  in  different  ways’,  and  are  encouraged  to  ‘take  more  responsibility  for  learning,  both  their  own  and  that  of  others.’  He  then  details  ways  of  implementing  peer  feedback,  including  that  ‘peer  review  should  begin  in  the  first  year  with  some  simple  review  tasks,  with  complexity  and  depth  being  enhanced  in  later  years.’  He  explains  how  to  ‘address  academic  integrity  and  circumvent  plagiarism,  i.e.  by  ‘having  students  review  assignments  produced  by  peers  and  then  comment  on  their  own  assignment,  but  without  having  the  opportunity  to  rewrite.’  (p.  3).    Vardi  (2012,  pp.  39-­‐40)  gives  advice  on  ensuring  that  students  are  ‘informed  providers  of  feedback’,  i.e.  ‘informed  evaluators’,  not  just  giving  their  opinion.  She  lists  a  number  of  strategies,  examples  of  which  follow:  ‘1.  Place  peer  feedback  activities  appropriately  in  the  learning  program.  Ensure  that  students  have  already  undertaken  the  necessary  work  (e.g.  readings,  activities)  prior  to  an  activity  incorporating  peer  feedback.  2.  Provide  a  clear  focus  for  the  feedback.  Focus  the  students’  attention:  evaluation  of  the  use  of  an  analysis  technique,  evaluation  of  a  draft  assignment.  …  5.  Provide  the  students  with  the  necessary  training.  Before  students  provide  feedback,  ensure  that  they  thoroughly  understand  a.  the  criteria,  frameworks  or  measures  they  are  to  use  b.  how  to  give  feedback.  Provide  students  with  practice  in  giving  feedback  using  exemplars.’    See  also:  Cartney  (2010).      Demonstrating  improvement  in  performance  as  a  result  of  a  peer  marking  exercise  Yorke  (2003)  cites  Gibbs’  (1999,  p.  43ff.)  example  of  Engineering  students  who,  six  times  a  year,  were  required  to  grade  their  peers’  work  anonymously  one  hour  after  submission:  ‘the  improvement  in  students’  end-­‐of-­‐course  outcomes  was  very  marked,  and  attributable  to  aspects  of  good  learning’.  Among  these  aspects  were  the  promptness  of  the  feedback,  and  ‘the  stimulus  to  self-­‐regulation  regarding  the  standard  of  submitted  work’  (p.  479).    Gibbs  and  Simpson  (2004)  summarise  a  (Forbes  and  Spence  (1991))  study  of  assessment  on  an  engineering  course  which  demonstrates  that  periodic  peer-­‐assessment  of  problem  sheets  ‘as  a  course  requirement  but  without  the  marks  contributing’  was  followed  by  a  ‘dramatic[]  increase’  in  the  students’  exam  marks,  i.e.  ‘to  a  level  well  above  that  achieved  previously  when  the  lecturers  did  the  marking’.  They  comment  that  ‘what  achieved  the  learning  was  the  quality  of  student  engagement  in  learning  tasks,  not  teachers  doing  lots  of  marking’  (p.  8).    (Cf.:  

  49  

Price,  O’Donovan  &  Rust  (2007):  failure  to  demonstrate  any  tangible  improvement  as  a  result  of  the  intervention,  but  ‘both  students  and  assessors  evaluate[d  the  latter]  positively  with  both  believing  it  to  be  beneficial’  (p.  150).  Orsmond,  Merry  and  Reiling  (2005,  p.  371)  reporting  on  Orsmond  et  al.  (2004):  ‘actively  engaging  students  in  a  formative  assessment  exercise  centred  on  marking  criteria’  resulted  in  their  being  able  to  ‘both  construct  and  interpret  detailed  marking  criteria  and  to  mark  assignments  …  accurately  compared  to  tutors’.  However,  the  students  were  unable  to  write  ‘academically  meaningful  comments  …  addressing  specific  criteria’,  i.e.  ‘the  formative  feedback  provided  had  failed  to  promote  demonstrable  learning’.  Basnet,  Brodie  and  Worden  (2010,  p.  205)  reports  on  students’  evaluation  of  peer  assessment  of  assignments,  finding  that  while  a  narrow  majority  of  respondents  found  the  peer  feedback  to  be  useful,  about  26%  ‘considered  peers’  feedback  of  little  use’.  A  number  commented  that  their  peers  ‘lacked  sufficient  knowledge  to  be  able  to  provide  useful  feedback’,  while  other  suggested  ‘providing  the  choice  of  a  set  of  specific  comments’  so  that  peers  would  not  write  ‘“offending”  comments’  as  feedback.)    Helping  changes  in  feedback  practice  (to  a  more  student-­‐centred  model)  to  work  For  Rust,  O’Donovan  and  Price  (2005),  active  engagement  with  assessment  criteria  by  both  markers  and  students  is  the  key  to  ‘achieving  meaningful  understanding  of  assessment’  (p.  233).  The  authors  make  a  number  of  suggestions  to  tutors  to  try  and  prevent  a  negative  reaction  from  students  who  might  think  that  what  they  are  being  asked  to  do  is  ‘the  tutor’s  job’.  These  suggestions  include  the  following  (p.  235):  

• Always  explain  clearly  why  the  approach  is  being  implemented,  i.e.  what  are  ‘the  pedagogic  reasons  behind  the  process’;  

• Where  this  is  possible,  ‘share  [the]  research  …  evidence  of  its  effectiveness’;  and  

• ‘Consider  attaching  assessment  incentives  to  the  activities,  if  appropriate’.    

4.2.4   Dialogic  approaches  to  feedback    Nicol  and  Macfarlane-­‐Dick  (2006)  argue  for  the  conceptualisation  of  feedback  as  ‘dialogue  rather  than  as  information  transmission’.  Conceptualising  feedback  as  ‘a  transmission  process  involving  “telling”’  disregards  ‘the  active  role  the  student  must  play  in  constructing  meaning  from  feedback  messages,  and  of  using  this  to  regulate  performance’  (p.  210).  If  teacher  dialogue  is  difficult  because  of  large  class  sizes,  feedback  can  be  increased  by  peer  dialogue,  e.g.  by  asking  students  in  tutorials  to  discuss  the  feedback  comments  they  have  received  and  suggest  strategies  for  improved  performance  in  the  next  assignment,  or  asking  students  to  discuss  examples  of  feedback  comments  they  have  found  useful  and  how  they  helped.    Carless  (2006,  pp.  230-­‐231)  argues  for  ‘assessment  dialogues’  between  tutors  and  students,  that  is,  ‘discussions  related  to  the  assessment  process  as  a  general  concept,  but  not  related  to  the  specifics  of  subject  matter  or  what  students  need  to  do  for  a  particular  assignment’.  He  identifies  some  of  the  areas  in  which  assessment  dialogues  might  be  useful,  including  ‘unpacking  assessment  criteria  or  involving  students  in  generating  or  applying  criteria’.    Carless,  Salter,  Yang  and  Lam  (2011)  argue  that  the  feedback  process  needs  to  be  reconceptualised,  and  that  part  of  this  reconceptualisation  relates  to  enhancing  the  

  50  

student  role  in  the  feedback  process.  Thus  the  authors  place  the  development  of  student  self-­‐regulation  at  the  heart  of  feedback  processes,  arguing  that  feedback  needs  to  be  viewed  ‘more  as  dialogue  than  information  transmission’  (p.  396).  Student  involvement  in  assessment  is  related  to  the  concept  of  ‘sustainable  feedback’,  which  Carless  et  al.  define  as  ‘dialogic  processes  and  activities  which  can  support  and  inform  the  student  on  the  current  task,  whilst  also  developing  the  ability  to  self-­‐regulate  performance  on  future  tasks’  (p.  397).  This  concept  is  at  the  opposite  end  of  the  feedback  practices  continuum  from  ‘conventional’  practices,  which  involve  strategies  such  as  written  feedback  on  drafts  or  final  versions  of  assignments.  So  student  self-­‐regulation  is  ‘a  central  aspect  of  sustainable  feedback’  (p.  398).    The  assessment  design  strategy  which  Carless  et  al.  see  as  encouraging  dialogic  feedback  involves  two-­‐stage  or  multi-­‐stage  assignments  (see  below).  Peer-­‐review,  blogs  engaging  students  ‘in  dialogues  around  learning’  as  part  of  the  class  participation  grade,  and  exploratory  rather  than  directive  feedback  in  PBL  in  class  to  promote  self-­‐directed  learning,  are  all  examples  of  involving  students  in  activities  promoting  self-­‐evaluation.    Nicol  (2010)  ‘suggests  ways  in  which  the  nature  and  quality  of  feedback  dialogue  can  be  enhanced  when  student  numbers  are  large  without  necessarily  increasing  demands  on  academic  staff’.  It  ‘concludes  with  a  conceptual  discussion  of  the  merits  of  taking  a  dialogical  approach  when  designing  feedback’  (p.  501).    

4.2.5   Using  technology  to  provide  peer  and  marker  feedback    Key  points  

• Online  learning  tasks  • Use  of  audio  files  • Peer  review  using  online  tools  • Computer  assisted  feedback,  including  the  use  of  statement  banks  • Related  University  websites  

 Online  learning  tasks  Nicol  (2009)  explores  ‘how  formative  assessment  and  feedback  might  be  used  to  enhance  the  first-­‐year  experience  and  enable  students  to  develop  the  skills  needed  for  self-­‐regulated  learning’  (p.  335).  The  paper  also  demonstrates  ways  in  which  technology  can  be  used  to  provide  feedback  in  large  classes.  The  author  understands  formative  assessment  [see  Glossary]  as  including  the  involvement  of  students  in  ‘evaluative  judgements  about  their  own  work  and  the  work  of  their  peers’  (p.  338).  In  the  first  case  study  of  a  redesigned  course,  the  online  learning  tasks  involved  collaborative  writing,  peer  dialogue  and  feedback,  the  use  of  model  answers,  and  general,  usually  motivational  feedback  from  the  course  leader  via  the  class  discussion  board.  Nicol  comments  that  the  redesign  did  not  increase  staff  workloads,  and  also  that  there  was  a  significant  increase  in  (self  and  peer)  feedback  opportunities.    Use  of  audio  files  Merry  and  Orsmond  (2008)  report  on  a  small-­‐scale  study  of  biology  students’  and  tutors’  attitudes  to  receiving  and  giving  formative  feedback  on  written  work  using  audio  files,  and  on  the  students’  use  of  this  feedback.  The  student  response  was  very  positive,  the  audio  feedback  being  judged  as  of  good  quality  because  ‘it  was  easier  to  understand,  had  more  depth  and  was  more  personal’.  A  large  majority  of  the  

  51  

students  listened  to  the  audio  feedback  more  than  once,  and  used  it  ‘to  annotate  or  alter  their  original  work  as  they  listened’.  Tutors  found  that  while  ‘providing  audio  feedback  did  not  save  them  time,  …  it  might  do  so  with  more  practice’,  and  that  its  main  benefit  was  its  ‘higher  quality’.  Merry  and  Orsmond  compared  the  audio  and  written  comments  for  assignments  marked  in  both  ways,  and  found  that  there  was  a  ‘statistically  significant  …  preponderance’  of  written  feedback  for  ‘identifying  errors’  and  one  of  audio  feedback  for  ‘demonstrating  best  practice’.  Guidelines  are  provided  for  tutors  wishing  to  provide  audio-­‐file  feedback.    Lunt  and  Curran  (2009)  report  on  a  small-­‐scale  study  of  the  use  of  virtual  learning  environment  (VLE)-­‐  and  email-­‐delivered  audio  feedback  from  the  perspectives  of  two  markers  and  26  students.  Both  tutors  ‘found  the  software  easy  to  download,  understand  and  use’  (p.  761).  The  use  of  MP3  (as  opposed  to  WAV)  files  solved  the  problem  of  emailing  large  files.  The  average  time  taken  to  record  the  audio  feedback  was  five  minutes  for  a  2000-­‐word  assignment,  and  was  shown  to  be  significantly  quicker  than  for  handwriting  or  typing  the  same  comments.  The  students  who  responded  to  the  survey  after  receiving  their  audio  feedback  were  very  positive  about  it,  the  feedback  being  judged  as  ‘more  detailed  than  written  comments’,  and  65%  agreeing  or  strongly  agreeing  that  they  ‘would  like  to  receive  [their]  feedback  using  the  audio  method’  (pp.  763-­‐764).  Lunt  and  Curran  provide  a  number  of  suggestions  for  tutors  interested  in  using  the  audio  method.    Rotheram  (2009)  provides  some  ‘practice  tips  on  using  digital  audio  for  assessment  feedback’  under  four  headings:  saving  time,  technical  matters,  administration,  and  feedback  structure.  These  tips  include  the  following:  ‘after  some  practice,  audio  may  allow  you  to  give  noticeably  more  feedback  without  spending  much  longer  on  the  task’;  ‘MP3  is  a  very  useful  audio  format:  compact  and  widely  playable.  …  You  can  make  MP3  recordings  with  a  …  handheld  digital  audio  recorder  [which  can  record  directly  to  MP3]’;  ‘Make  sure  key  administrative  and  quality-­‐assurance  staff  accept  that  you  are  giving  audio  rather  than  written  feedback.  Letting  people  listen  to  a  few  examples  usually  helps  to  convince  them  of  the  educational  value’,  and  ‘Keep  the  files  short  –  don’t  “overdo  it”’.  Rotheram  helpfully  includes  ‘an  outline  of  [his]  most  recent  practice  when  commenting  on  a  typical  essay  or  report  via  an  MP3  recording’.  The  link  to  his  complete  project  report  can  be  accessed  via  the  Sounds  Good  website  (www.soundsgood.org.uk).  See  also  Rotheram  (2007)  for  some  background  on  using  MP3  recordings  to  give  feedback  on  student  assignments.    Peer  review  using  online  tools  Pearce,  Mulder  and  Baik  (2009),  in  part,  present  ‘an  overview  of  online  tools  that  are  currently  available  to  academics  for  managing  student  peer  reviews’  (p.  2).  They  discuss  ‘well-­‐established  and  accessible  platforms  (for  which  good  documentation  exists),  and  …  highlight  their  capabilities,  strengths  and  weaknesses’  (p.  23).  These  software  packages  are:  CPR,  PRAZE  (the  University  of  Melbourne’s  in-­‐house  program),  SPA  (Blackboard),  SWoRD,  and  Turnitin,  which  are  compared  in  terms,  e.g.,  of  ‘Primary  use/niche’,  ‘Main  advantages’,  ‘Main  drawbacks’,  ‘Features’  and  ‘Availability/Cost’.  Pearce  et  al.  argue  that  ‘the  best  choice  of  software  for  a  particular  subject  or  assignment  will  depend  on  the  level  of  individual  customisation  that  is  required  …,  and  the  administrator’s  level  of  experience  and/or  willingness  to  experiment  …’  (p.  24).    Computer  assisted  feedback,  including  the  use  of  statement  banks  

  52  

Denton,  Madden,  Roberts  and  Rowe  (2008)  report  on  a  study  which  compared  students’  responses  to  ‘traditional’  handwritten  feedback  on  an  extended  lab  report  with  their  responses  to  computer-­‐assisted  feedback  (an  MS  Office  application,  Electronic  Feedback  freeware).  Tutors  selected  11  statements  from  a  bank  of  64  ‘criterion-­‐style’  comments  and  up  to  14  comments  which  had  been  inputted,  in  addition  to  making  their  general  comments,  and  then  generated  a  report  for  each  student  which  was  emailed  out  automatically.  The  findings  were  that  students  generally  showed  a  preference  for  the  electronic  feedback,  which  Denton  et  al.  explain  as  being  due  to  the  ‘clear  structure  of  the  feedback  and  its  integration  with  the  marking  scheme’,  and  the  greater  amount  of  feedback  given  as  a  result  of  the  use  of  a  statement  bank  (p.  495).  The  markers  reported  that  they  took  2-­‐3  minutes’  less  time  to  mark  each  script,  but  ‘this  time  has  to  be  offset  against  the  effort  required  to  write  the  initial  bank  of  feedback  statements’  (p.  497).      Case  (2007)  reports  on  the  implementation  of  a  reconfigured  system  of  assessment  feedback  with  undergraduate  students,  a  system  which  he  describes  as  addressing  the  crucial  aspects  of  explicit  engagement  with  assessment  criteria  and  constructive  feedback  (p.  287).  The  reformulation  process  involved  the  devising  of  an  electronic  template  feedback  form  which  incorporated  ‘the  assessment  criteria  relevant  to  each  student’s  grade  as  a  table’.  Students  were  provided  with  this  table,  as  well  as  ‘recommendations  for  how  to  use  it’.  Markers  were  provided  with  an  electronically-­‐stored  pre-­‐formulated  statement  bank,  ‘chosen  to  save  time  and  effort  reiterating  comments  on  scripts  relating  to  common  student  weaknesses  …,  mistakes  …,  recommended  improvements  …,  and  strengths’  (p.  289),  and  they  were  also  able  to  add  their  own  comments.  Students  were  offered  a  follow-­‐up,  face-­‐to-­‐face  meeting  with  their  marker.  The  reconfigured  system  resulted  in  ‘significant  improvements  in  assessment  grades’,  and  student  feedback  suggested  that  it  promoted  ‘a  strategic  approach  to  learning’.  It  also  streamlined  the  feedback  procedure  for  lecturers,  resulting  in  ‘reduced  marking  time  …,  fewer  student  queries  …,  and  an  efficient  student  feedback  system’  (p.  292).  And  it  seems  that  students  are  ‘engaging  positively’  with  the  feedback  provided,  and  are  ‘beginning  to  understand  the  reasons  behind  assessment  decisions’  (p.  294).    Hepplestone,  Holden,  Irwin,  Parkin  and  Thorpe  (2011)  present  a  literature  review  on  ‘Using  technology  to  encourage  student  engagement  with  feedback’.  In  this  review,  the  authors  focus  on  research  exploring  ‘the  publication  of  feedback  and  grades  online,  the  automated  adaptive  release  of  grades  (or  how  technology  can  support  encouraging  student  engagement  through  disengaging  the  grade  from  feedback),  and  the  use  of  technology  to  automatically  generate  feedback  aligned  to  assessment  criteria  through  pre-­‐populated  statement  banks’  (p.  119).    See  also  Bridgeman  and  Rutledge  (2010)  in  3.2.6  –  Dealing  with  large  classes  –  below.    Related  University  websites  Sheffield  Hallam  University:  ‘Technology,  Feedback,  Action!’  project  The  project  ‘Technology,  Feedback,  Action!:  The  impact  of  learning  technology  upon  students’  engagement  with  their  feedback’  was  undertaken  in  2008/2009.  In  the  final  short  report  on  this  project  (2010),  Hepplestone,  Parkin,  Irwin,  Holden  and  Thorpe  explain  that  it  ‘evaluated  how  a  range  of  technical  interventions  might  encourage  students  to  engage  with  their  feedback  and  formulate  actions  to  improve  future  learning’.  These  interventions  included  the  online  publication  of  feedback  and  

  53  

grades,  disengaging  the  grade  from  the  feedback,  and  linking  the  feedback  given  to  assessment  criteria  (p.  4).  The  first  two  of  these  were  found  to  ‘significantly  enhance  students'  engagement  with  their  feedback’.  The  third,  however,  ‘while  effective  in  enabling  students  to  identify  strengths  and  weaknesses  at  a  glance  and  helping  to  identify  learning  targets,  was  less  effective  in  terms  of  enhancing  engagement  with  feedback’.  The  materials  also  include  the  full  final  report  and  an  ‘Annotated  Bibliography’,  where  the  literature  is  set  out  on  a  grid  specifying  which  documents  relate  to  eight  topic  areas.  The  latter  include  ‘Key  source/research  on  feedback’,  ‘Technology-­‐enabled  feedback’,  ‘Linking  feedback  to  learning  outcomes/comment  banks’,  and  ‘Feedback  as  dialogue’.  They  also  include  ‘A  best  practice  guide  for  academic  staff’,  ‘A  student  guide  to  using  feedback’,  and  ‘A  ten  minute  guide  for  senior  managers’.  http://evidencenet.pbworks.com/w/page/19383525/Technology,%20Feedback,%20Action!%3A%20Impact%20of%20Learning%20Technology%20on%20Students%27%20Engagement%20with%20Feedback    Leeds  Metropolitan  University:  ‘Sounds  Good’  project  Final  Report  (Rotheram,  2009,  http://sites.google.com/site/soundsgooduk/)    This  project  (Jan  2008  –  March  2009)  explored  the  use  of  digital  audio  to  give  assessment  feedback.  Its  main  aim  was  ‘to  test  the  hypothesis  that  using  digital  audio  for  feedback  can  benefit  staff  and  students  by  saving  assessors’  time  (speaking  the  feedback  rather  than  writing  it),  and  providing  richer  feedback  to  students  (speech  is  a  richer  medium  than  written  text)’.  Five  key  questions  were  explored,  including  ‘(Without  reducing  the  amount  of  feedback)  in  what  circumstances  can  using  digital  audio  save  assessors’  time?’  and  ‘What  do  assessors  think  of  digital  audio  as  a  medium  for  providing  feedback  to  students?’  (p.  2).  The  answers  revealed  that  the  most  favourable  circumstances  included  the  assessor  being  ‘comfortable  with  the  technology’,  and  the  availability  of  ‘a  quick  and  easy  method  of  delivering  the  audio  file  to  the  student’.  They  also  revealed  that  most  staff  said  they  intended  to  continue  using  digital  audio  feedback,  several  commenting  that  ‘they  were  able  to  give  more,  and  higher-­‐quality,  feedback  using  audio,  which  they  felt  was  worthwhile’  (p.  2).    Brown  (n.d.)  presents  an  overview  of  best  practice  in  feedback,  including  the  use  of  oral  feedback,  statement  banks,  computer-­‐assisted  assessment,  electronic  feedback,  and  self  and  peer  assessment.  (See  also:  Brown,  Sally  (2006)  Powerpoint  ‘Using  Formative  Assessment  to  promote  student  learning’  www.ldu.leeds.ac.uk/news/events/documents/BrownPowerPoint.pdf)    Nottingham  University:  ‘Enhancing  student  feedback  through  the  use  of  technology’  project  This  project  (‘Enhancing  student  feedback  through  the  use  of  technology:  trialling  audio  and  video  e-­‐feedback  on  students’  work  in  the  School  of  English’),  which  began  in  September  2011  with  a  one-­‐day  conference  at  Nottingham,  ‘aims  to  improve  feedback  to  students  and  the  student  academic  experience  by  trialling  the  use  of  audio  and  video  feedback  in  the  School  [of  English]’.  It  ‘will  test  out  the  effectiveness  of  these  technologies  with  both  undergraduate  and  postgraduate  students  and  in  the  “live”  and  distance  learning  environments’.    http://www.nottingham.ac.uk/english/teachinglearning/enhancing-­‐feedback-­‐report.aspx    

  54  

4.2.6   Dealing  with  large  classes    Key  points  

• Peer  assessment  and  feedback  in  large  classes  • Use  of  statement  banks  for  feedback  with  large  classes  

 Feedback  with  large  classes  Rust  (2001)  argues  that,  as  class  sizes  go  up,  the  amount  and  quality  of  feedback  students  receive  is  likely  to  become  ‘a  major  casualty’  (p.  4).  Strategies  for  addressing  this  problem  include  giving  general  rather  than  individual  class  feedback,  peer  assessment,  using  feedback  sheets,  and  using  statement  banks.    Bridgeman  and  Rutledge  (2010)  report  on  a  their  ‘open-­‐source  software’  ‘Fast  &  Personal  e-­‐Feedback  system  (FPF)’,  a  new  approach  which  allows  ‘the  rapid,  efficient  delivery  of  personalised  feedback  to  large  student  cohorts’  (p.  61),  and  which  is  now  used  in  First  and  Second  Year  Chemistry  units  at  the  University  of  Sydney.  The  program  ‘processes  marks  files  to  produce  individual  reports  and  emails  for  every  student  in  a  class,  whatever  the  class  size’  (p.  62).  All  email  correspondence  is  personalised,  and  the  messages  are  ‘short  and  highly  pertinent’.  The  system  has  proved  to  be  ‘enormously  popular  with  students’  as  the  FPF  software  writes  a  report  for  each  student  on  the  regular  quizzes  they  complete,  and  this  report  is  attached  to  their  email  as  a  pdf  file.  The  report  provides  a  full  solution  or  explanation  for  each  question  answered  incorrectly,  and  full  solutions  to  all  questions  are  made  available  online.  Where  a  series  of  incorrectly  answered  linked  questions  indicates  a  weakness  in  a  certain  area,  the  student’s  personal  feedback  report  lists  resources  that  can  be  used  to  ‘remedy  their  shortcoming’  (p.  64),  the  feedback  thus  feeding  forward  to  future  assessments.  Rapidity  is  another  ‘key  feature’  of  the  FPF,  as  computer  marking  enables  emails  to  be  sent  out  ‘almost  immediately’  after  the  task  has  been  completed,  thus  maintaining  student  interest.  The  system  also  provides  feedback  for  teachers,  and  exam  feedback  for  students.    Peer  assessment  in  large  classes:  Ballantyne,  Hughes  and  Mylonas  (2002)  reports  on  the  development  of  peer  assessment  procedures  for  use  in  large  classes.  The  development  of  procedures  ‘is  discussed  in  relation  to  assessment  tasks,  assessment  criteria,  anonymity,  procedural  guidelines,  distribution  systems,  marking  procedures  and  tutor  remarking’.  The  study  suggest  that  ‘the  specific  difficulties  associated  with  the  use  of  peer  assessment  in  large  classes  …  are  outweighed  by  the  learning  benefits  for  students’,  and  recommendations  are  made  ‘for  ways  in  which  peer  assessment  might  be  successfully  applied  in  large  classes’  (p.  427).    Use  of  statement  banks  with  large  classes:  Sandhu  (2004)  presents  a  case  study  which  evaluates  the  effectiveness  of  statement  banks  in  the  assessment  of  large  groups  of  students  in  Accounting.  The  author  found  that  tutors  were  able  to  give  more  formative  feedback  (the  markers  could  also  add  in  further  comments),  that,  since  the  process  of  marking  was  ‘speed[ed]  up’,  the  students  received  ‘more  timely  feedback’,  and  that  the  use  of  the  statement  banks  ‘enhance[d]  marker  consistency’.  Students  were  able  to  ‘recognise  their  strengths  and  weaknesses  and  thus  understand  how  their  final  grade  was  established’  (p.  51).  Sandhu  makes  a  number  of  recommendations  concerning  the  feedback  form  used,  e.g.  it  had  to  ‘show  both  positive  and  negative  comments  [and  ]  in  detail  where  the  student  may  have  mis-­‐understood  the  question  being  asked’,  and  to  ‘be  flexible  and  

  55  

adaptable  if  it  was  to  be  developed  over  time’,  and  includes  a  copy  of  the  feedback  form  used  in  this  study.    

4.2.7   Feedforward    (See  Glossary)  Higgins,  Hartley  and  Skelton  (2001,  p.  274),  in  an  article  dealing  with  feedback  as  a  communication  process,  suggest  in  their  last  lines  that  the  emphasis  may  need  to  be  moved  from  feeding  back  to  ‘feeding  forward’  into  a  piece  of  student  work,  i.e.  ‘equip[ping]  students  with  a  better  appreciation  of  what  is  expected  of  them’,  and  ‘develop[ing]  their  understandings  of  academic  terms  and  appropriate  practices  before  or  as  they  begin  to  write’.    In  Hattie  and  Timperley’s  (2007)  feedback  model  (a  ‘framework  to  understand  why  particular  kinds  of  feedback  promote  learning  effectively  and  why  some  others  do  not’  (p.  86)),  the  feedforward  question  is  ‘Where  to  next?’:  

The  power  of  feedback  …  can  be  used  to  specifically  address  this  question  by  providing  information  that  leads  to  greater  possibilities  for  learning.  These  may  include  enhanced  challenges,  more  self-­‐regulation  over  the  learning  process,  greater  fluency  and  automaticity,  more  strategies  and  processes  to  work  on  the  tasks,  deeper  understanding,  and  more  information  about  what  is  and  what  is  not  understood.  This  feedforward  question  can  have  some  of  the  most  powerful  impacts  on  learning.  (p.  90)  

 Duncan  (2007)  reports  on  a  small-­‐scale  research  project  developed  in  response  to  students’  non-­‐use  of  feedback  due  to  ‘the  lack  of  appreciation  that  comments  on  one  essay  could  help  achievement  in  a  later  assignment’  (p.  272).  The  project  involved  16  student  volunteers  in  one  module  of  study  submitting  eight  or  more  feedback  sheets  from  previous  assignments,  and  tutor  analysis  of  the  key  issues  ‘that  appeared  as  repeated  difficulties  for  each  participant’.  Then  module  tutors  ‘interpreted  the  criteria  for  the  next  assignment  [of  which  they  were  to  be  the  markers]  in  the  light  of  the  comments  on  students’  previous  work’  in  order  to  provide  feedforward  to  the  individual  students,  who  were  also  offered  a  one-­‐on-­‐one  session  for  needs  analysis  (D.  Hounsell  et  al.,  2008)  and  the  formulation  of  ‘an  individual  learning  plan  for  the  next,  summative,  assignment’,  and  a  draft-­‐reading  service  (p.  273).  The  statistical  outcomes  ‘showed  a  small  gain  in  the  grades  achieved  against  those  who  did  not  participate’  (p.  271).    For  Hounsell,  McCune,  Hounsell  &  Litjens  (2008),  feedforward  is  the  last  of  the  six  steps  of  their  guidance  and  feedback  loop,  and  its  potential  can  only  be  realised  if  feedback  is  timely.    Price,  Handley,  Millar  and  O’Donovan  (2010)  identify  five  roles  attributed  to  feedback  (see  ‘Purposes  of  feedback’  above):  ‘correction,  reinforcement,  forensic  diagnosis,  benchmarking  and  longitudinal  development  (feedforward)’.  As  regards  the  latter,  the  authors  argue  that  the  increasing  emphasis  on  feedforward  means  ‘not  only  including  feedback  directed  at  supporting  improvements  in  the  next  assignment  but  also  providing  advice  and  guidance  that  supports  slowly  learnt  literacies  (Knight  and  York  2004)  and  coming  to  understand  threshold  concepts  (Meyer  and  Land  2006)’  (p.  279).    

  56  

Brown  (n.d.)  argues  that  formative  feedback  can  be  defined  as  ‘feedback  that  feeds  forward’,  in  other  words,  that  feedback  ‘must  be  relevant  to  future  work’.  For  Brown  and  the  other  members  of  the  Formative  Assessment  in  Science  Teaching  (FAST)  project  team,  most  written  feedback  received  by  students  ‘serves  to  justify  their  grades  for  the  current  work  by  exposing  (mainly)  their  weaknesses,  without  necessarily  explaining  clearly  the  bases  of  the  weaknesses  or  enabling  students  to  improve  their  future  performance’.  The  author  refers  to  a  number  of  ‘strategies  for  change’,  one  such  strategy  involving  the  avoidance  of  any  reference  to  marks  in  the  feedback  provided  in  order  to  encourage  students  to  focus  on  the  areas  for  improvement  identified  by  the  marker.    Brown  and  Glover  (2006)  report  on  a  study  which  first  ‘classif[ied]  systematically  the  different  types  of  teacher  [written]  comments  that  constitute  feedback  so  that  the  quality  of  feedback  [could]  be  analysed’  (p.  82).  This  classification  system  has  five  main  categories:  comments  ‘about  the  content  of  a  student’s  response’,  comments  ‘that  help  a  student  to  develop  appropriate  skills’,  comments  ‘that  actively  encourage  further  learning’,  comments  ‘that  are  motivational’,  and  comments  ‘that  may  de-­‐motivate’  (p.  83).  The  authors  identify  three  levels  of  feedback,  namely  that  which  ‘acknowledge[s]  a  weakness:  i.e.,  ‘acknowledge[s]  a  performance  gap  exists  (level  1)’,  that  which  ‘provide[s]  correction:  i.e.,  gives  the  student  the  information  needed  to  close  the  gap  (level  2)’,  and  that  which  ‘explain[s]  why  the  student’s  response  is  inappropriate/why  the  correction  is  a  preferred  response:  i.e.,  enable[s]  the  student  to  use  the  information  to  close  the  gap  (level  3)’  (p.  85).  This  classification  system  was  then  used  to  analyse  tutor  feedback.  The  changes  made  as  a  result  of  the  identification  of  weaknesses  in  the  type  and  quality  of  feedback  include  ‘encouraging  tutors  to  highlight  aspects  of  the  student’s  strengths  and  weaknesses  that  have  relevance  for  future  work’,  and  ‘providing  exemplars  (specimen  answers  )  for  students  with  explanatory  notes  that  stress  skills  development  and  the  relevance  to  future  work’  (p.  89).            

  57  

5.0:   Glossary  of  key  terms  –  definitions  from  the  literature:    Explicit  knowledge  

• that   part   of   knowledge   which   ‘can   be   expressed   in   formal   and   systematic  language’  (Nonaka  et  al.,  2002,  p.  43,  cited  in  Rust  et  al.,  2003).  

 Feedback  defined  in  terms  of  its  informational  content:  

• ‘is  …  information  about  how  successfully  something  has  been  or  is  being  done  …  Most  [skills]  require  practice  in  a  supportive  environment  which  incorporates  feedback  loops.  This  usually  includes  a  teacher  who  knows  which  skills  are  to  be  learned,  and  who  can  recognize  and  describe  a  fine  performance,  and  indicate  how  a  poor  performance  can  be  improved’  (Sadler,  1989,  p.  120).  

defined  in  terms  of  its  effect:  • ‘is  information  about  the  gap  between  the  actual  level  and  the  reference  level  

of  a  system  parameter  which  is  used  to  alter  the  gap  in  some  way’  (Ramaprasad,  1983,  p.  4,  cited  in  Sadler,  1989,  p.  120).  

Main  audiences:  the  teacher  and  the  student  • ‘Teachers  use  feedback  to  make  programmatic  decision  with  respect  to  

readiness,  diagnosis,  and  remediation.  Students  use  it  to  monitor  the  strengths  and  weaknesses  of  their  performances,  so  that  aspects  associated  with  success  or  high  quality  can  be  recognized  and  reinforced,  and  unsatisfactory  aspects  modified  or  improved’  (Sadler,  1989,  pp.  120-­‐121).  

 Feedforward  

• refers  to  feedback  that  has  the  potential  to  feed  forward  into  a  subsequent  assignment  or  assessment  (Hounsell  et  al.,  2008)  .  For  Hounsell  et  al.,  it  represents  the  last  of  the  six  steps  in  their  guidance  and  feedback  loop.  

 Formative  assessment  

• ‘refers  to  assessment  that  is  specifically  intended  to  generate  feedback  on  performance  to  improve  and  accelerate  learning’  (Nicol  &  Macfarlane-­‐Dick,  2006,  p.  199).  

• ‘is  concerned  with  how  judgments  about  the  quality  of  student  responses  (performances,  pieces,  or  works)  can  be  used  to  shape  and  improve  the  student’s  competence  by  short-­‐circuiting  the  randomness  and  inefficiency  of  trial-­‐and-­‐error  learning’  (Sadler,  1989,  p.  120).  

 Qualitative  judgement(s)  

• ‘one  made  directly  by  a  person,  the  person’s  brain  being  both  the  source  and  the  instrument  for  the  appraisal’  (Sadler,  1989,  p.  124).  

• are  those  made  in  a  context  where  ‘student  development  is  multidimensional  rather  than  sequential,  and  prerequisite  learnings  cannot  be  conceptualised  as  neatly  packaged  units  of  skills  or  knowledge’,  and  where  ‘it  is  more  appropriate  to  think  in  terms  of  the  quality  of  a  student’s  response  or  the  degree  of  expertise  than  in  terms  of  facts  memorized,  concepts  acquired  or  content  mastered’  (Sadler,  1989,  p.  123).  

• have  five  characteristics:  o ‘multiple  [interlocking]  criteria  are  used  in  appraising  the  quality  of  

performances’;  

  58  

o ‘at  least  some  of  the  criteria  used  in  appraisal  are  fuzzy  [with  ‘continuous  gradation  from  one  state  to  another]  rather  than  sharp’  [e.g.  correct  or  incorrect];  

o ‘the  teacher  may  find  it  impossible  to  specify  all  the  relevant  criteria  in  advance,  or  may  find  that  a  fixed  set  of  criteria  is  not  uniformly  applicable  to  different  student  responses  …  to  the  same  task’;  

o ‘the  final  court  of  appeal  is  to  another  qualitative  judgment’;  o ‘the  final  decision  is  never  arrived  at  by  counting  things’  –  if  …  marks  …  

are  used,  they  are  assigned  after  the  judgment  has  been  made’  (Sadler,  1989,  pp.  124-­‐125).  

 Summative  assessment  

• ‘is  concerned  with  summing  up  or  summarizing  the  achievement  status  of  a  student,  and  is  geared  towards  reporting  at  the  end  of  a  course  of  study  especially  for  purposes  of  certification  (Sadler,  1989,  p.  120).  

 Tacit  knowledge  

• that  part  of  [a  teacher’s]  knowledge  which  is  difficult  to  articulate  and  can  go  ‘missing’.  It  is  ‘highly  personal  and  hard  to  formalise’  (Rust  et  al.,  2003,  pp.  151-­‐152).  

• Deeply  rooted  in  action  and  often  in  an  individual’s  commitment  to  a  profession,  tacit  knowledge  consists  partly  of  technical  skills  based  on  professional  experience,  and  in  a  more  cognitive  dimension,  in  our  ingrained  mental  models,  beliefs  and  perspectives  (Nonaka  et  al.,  1991,  cited  in  Rust  et  al.,  2003).  

 

  59  

References    Ballantyne,  R.,  Hugher,  K.,  &  Mylonas,  A.  (2002).  Developing  procedures  for  

implementing  peer  assessment  in  large  classes  using  an  action  research  process.  Assessment  &  Evaluation  in  Higher  Education,  27(5),  427-­‐441.  doi:  10.1080/0260293022000009302  

Basnet,  B.,  Brodie,  L.,  &  Worden,  J.  (2010).  Enhanced  Feedback  -­‐  Does  Peer  Assessment  achieve  this  goal?  Paper  presented  at  the  AaeE  2010:  Past,  Present,  Future  -­‐  the  'Keys'  to  Engineering  Education  Research  and  Practice.  

Black,  P.,  &  Wiliam,  D.  (1998).  Assessment  and  classroom  learning.  Assessment  in  Education,  5(1),  7-­‐74.  doi:  10.1080/0969595980050102  

Bridgeman,  A.,  &  Rutledge,  P.  (2010).  Getting  personal:  Feedback  for  the  masses.  Synergy  (Institute  for  Teaching  and  Learning,  University  of  Sydney)(30),  61-­‐68.    

Brown,  E.  (n.d.).  Effective  feedback.  Formative  Assessment  in  Science  Teaching  (FAST)    Retrieved  22.03.12,  2012  

Brown,  E.,  &  Glover,  C.  (2006).  Evaluating  written  feedback.  In  C.  Bryan  &  K.  Clegg  (Eds.),  Innivative  assessment  in  Higher  Education  (pp.  81-­‐91).  London:  Routledge.  

Carless,  D.  (2006).  Differing  perceptions  in  the  feedback  process.  Studies  in  Higher  Education,  31(2),  219-­‐233.  doi:  10.1080/03075070600572132  

Carless,  D.,  Salter,  D.,  Yang,  M.,  &  Lam,  J.  (2011).  Developing  sustainable  feedback  practices.  Studies  in  Higher  Education,  36(4),  395-­‐407.  doi:  10.1080/03075071003642449  

Cartney,  P.  (2010).  Exploring  the  use  of  peer  assessment  as  a  vehicle  for  closing  the  gap  between  feedback  given  and  feedback  used.  Assessment  &  Evaluation  in  Higher  Education,  35(5),  551-­‐564.  doi:  10.1080/02602931003632381  

Case,  S.  (2007).  Reconfiguring  and  realigning  the  assessment  feedback  processes  for  an  undergraduate  criminology  degree.  Assessment  &  Evaluation  in  Higher  Education,  32(3),  285-­‐299.  doi:  10.1080/02602930600896548  

Chanock,  K.  (2000).  Comments  on  Essays  :  Do  students  understand  what  tutors  write?  Teaching  in  Higher  Education,  5(1),  95-­‐105.    

Covic,  T.,  &  Jones,  M.  K.  (2008).  Is  the  essay  resubmission  option  a  formative  or  a  summative  assessment  and  does  it  matter  as  long  as  the  grades  improve?  Assessment  &  Evaluation  in  Higher  Education,  33(1),  75-­‐85.  doi:  10.1080/02602930601122928  

Denton,  P.,  Madden,  J.,  Roberts,  M.,  &  Rowe,  P.  (2008).  Students'  response  to  traditional  and  computer-­‐assisted  formative  feedback:  A  comparative  case  study.  British  Journal  of  Educational  Technology,  39(3),  486-­‐500.  doi:  10.1111/j.1467-­‐8535.2007.00745.x  

Duncan,  N.  (2007).  'Feed-­‐forward':  improving  students'  use  of  tutors'  comments.  Assessment  &  Evaluation  in  Higher  Education,  32(3),  271-­‐283.  doi:  10.1080/02602930600896498  

Gibbs,  G.,  &  Simpson,  C.  (2004).  Conditions  Under  Which  Assessment  Supports  Students  ’  Learning.  Learning  and  Teaching  in  Higher  Education,  1(1),  3-­‐31.    

Glover,  C.,  &  Brown,  E.  (2006).  Written  Feedback  for  Students  :  too  much  ,  too  detailed  or  too  incomprehensible  to  be  effective  ?  Bioscience  Education  eJournal,  7(May).    

Hattie,  J.,  &  Timperley,  H.  (2007).  The  power  of  feedback.  Review  of  Educational  Research,  77(1),  81-­‐112.  doi:  10.3102/003465430298487  

Hendry,  G.  D.,  Bromberger,  N.,  &  Armstrong,  S.  (2011).  Constructive  guidance  and  feedback  for  learning:  The  usefulness  of  exemplars,  marking  sheets  and  

  60  

different  types  of  feedback  in  a  first  year  Law  subject.  Assessment  &  Evaluation  in  Higher  Education,  36(1),  1-­‐11.  doi:  10.1080/02602930903128904  

Hepplestone,  S.,  Holden,  G.,  Irwin,  B.,  Parkin,  H.  J.,  &  Thorpe,  L.  (2011).  Using  technology  to  encourage  student  engagement  with  feedback:  A  literature  review.  Research  in  Learning  Technology,  19(2),  117-­‐127.  doi:  10.1080/21567069.2011.586677  

Higgins,  R.,  Hartley,  P.,  &  Skelton,  A.  (2001).  Getting  the  message  across:  The  problem  of  communicating  assessment  feedback.  Teaching  in  Higher  Education,  6(2),  269-­‐274.  doi:  10.1080/13562510120045230  

Higgins,  R.,  Hartley,  P.,  &  Skelton,  A.  (2002).  The  Conscientious  Consumer  :  Reconsidering  the  role  of  assessment  feedback  in  student  learning.  Studies  in  Higher  Education,  27(1),  53-­‐64.  doi:  10.1080/03075070120099368  

Holmes,  L.  E.,  &  Smith,  L.  J.  (2003).  Student  evaluations  of  faculty  grading  methods.  Journal  of  Education  for  Business,  78(6),  318-­‐323.    

Hounsell,  D.,  McCune,  V.,  Hounsell,  J.,  &  Litjens,  J.  (2008).  The  quality  of  guidance  and  feedback  to  students.  Higher  Education  Research  &  Development,  27(1),  55-­‐67.  doi:  10.1080/07294360701658765  

Hyland,  F.,  &  Hyland,  K.  (2001).  Sugaring  the  pill:  Praise  and  criticism  in  written  feedback.  Journal  of  Second  Language  Writing,  10,  185-­‐212.    

Lunsford,  R.  F.  (1997).  When  less  is  more:  Principles  for  responding  in  the  disciplines.  In  P.  Elbow  &  M.  D.  Sorcinelli  (Eds.),  Writing  to  learn:  Strategies  for  assigning  and  responding  to  writing  across  the  disciplines  (pp.  91-­‐104).  San  Francisco:  Jossey-­‐Bass.  

Lunt,  T.,  &  Curran,  J.  (2009).  ‘Are  you  listening  please?’  The  advantages  of  electronic  audio  feedback  compared  to  written  feedback.  Assessment  &  Evaluation  in  Higher  Education,  35(7),  759-­‐769.  doi:  10.1080/02602930902977772  

Merry,  S.,  &  Orsmond,  P.  (2008).  Students'  attitudes  to  and  usage  of  academic  feedback  provided  via  audio  files.  Bioscience  Education,  11.  Retrieved  from  http://www.bioscience.heacademy.ac.uk/journal/vol11/beej-­‐11-­‐3.pdf  website:    

Nicol,  D.  J.  (2009).  Assessment  for  learner  self-­‐regulation:  Enhancing  achievement  in  the  first  year  using  learning  technologies.  Assessment  &  Evaluation  in  Higher  Education,  34(3),  335-­‐352.  doi:  10.1080/02602930802255139  

Nicol,  D.  J.  (2010).  From  monologue  to  dialogue:  improving  written  feedback  processes  in  mass  higher  education.  Assessment  &  Evaluation  in  Higher  Education,  35(5),  501-­‐517.  doi:  10.1080/02602931003786559  

Nicol,  D.  J.  (2011).  Developing  students'  ability  to  construct  feedback.  Paper  presented  at  the  QAA  Enhancement  Themes  Conference,  Heriot-­‐Watt  University.    

Nicol,  D.  J.,  &  Macfarlane-­‐Dick,  D.  (2006).  Formative  assessment  and  self-­‐regulated  learning:  A  model  and  seven  principles  of  good  feedback  practice.  Studies  in  Higher  Education,  31(2),  199-­‐218.  doi:  10.1080/03075070600572090  

Orsmond,  P.,  &  Merry,  S.  (2011).  Feedback  alignment:  Effective  and  ineffective  links  between  tutors’  and  students’  understanding  of  coursework  feedback.  Assessment  &  Evaluation  in  Higher  Education,  36(2),  125-­‐136.  doi:  10.1080/02602930903201651  

Orsmond,  P.,  Merry,  S.,  &  Reiling,  K.  (2005).  Biology  students'  utilization  of  tutors'  formative  feedback:  A  qualitative  interview  study.  Assessment  &  Evaluation  in  Higher  Education,  30(4),  369-­‐386.  doi:  10.1080/02602930500099177  

Pearce,  J.,  Mulder,  R.,  &  Baik,  C.  (2009).  Involving  students  in  peer  review:  Case  studies  and  practical  strategies  for  university  teaching.  Retrieved  from  http://www.cshe.unimelb.edu.au/  

  61  

Price,  M.,  Carroll,  J.,  O'Donovan,  B.,  &  Rust,  C.  (2010).  If  I  was  going  there  I  wouldn't  start  from  here:  a  critical  commentary  on  current  assessment  practice.  Assessment  &  Evaluation  in  Higher  Education.  Retrieved  from    doi:10.1080/02602930903512883  

Price,  M.,  Handley,  K.,  Millar,  J.,  &  O'Donovan,  B.  (2010).  Feedback  :  all  that  effort,  but  what  is  the  effect?  Assessment  &  Evaluation  in  Higher  Education,  35(3),  277-­‐289.  doi:  10.1080/02602930903541007  

Price,  M.,  O'Donovan,  B.,  &  Rust,  C.  (2007).  Putting  a  social-­‐constructivist  assessment  process  model  into  practice:  Building  the  feedback  loop  into  the  assessment  process  through  peer  review.  Innovations  in  Education  and  Teaching  International,  44(2),  143-­‐152.    

Rotheram,  B.  (2007).  Using  an  MP3  recorder  to  give  feedback  on  student  assignments.  Educational  Developments,  8(2),  7-­‐10.    

Rotheram,  B.  (2009).  Practice  tips  on  using  digital  audio  for  assessment  feedback    Retrieved  31  January,  2012,  from  http://www.engsc.ac.uk/events/using-­‐technology-­‐to-­‐enhance-­‐learning-­‐teaching-­‐assessment-­‐and-­‐feedback  

Rust,  C.  (2001).  A  briefing  on  assessment  of  large  groups.  Generic  Centre  Assessment  Series  No.  12,  from  http://www.bioscience.heacademy.ac.uk/resources/themes/assessment.aspx  

Rust,  C.  (2007).  Towards  a  scholarship  of  assessment.  Assessment  &  Evaluation  in  Higher  Education,  32(2),  229-­‐237.  doi:  10.1080/02602930600805192  

Rust,  C.,  O'Donovan,  B.,  &  Price,  M.  (2005).  A  social  constructivist  assessment  process  model:  how  the  research  literature  shows  us  this  could  be  best  practice.  Assessment  &  Evaluation  in  Higher  Education,  30(3),  231-­‐240.  doi:  10.1080/02602930500063819  

Rust,  C.,  Price,  M.,  &  O'Donovan,  B.  (2003).  Improving  students'  learning  by  developing  their  understanding  of  assessment  criteria  and  processes.  Assessment  &  Evaluation  in  Higher  Education,  28(2),  147-­‐164.  doi:  10.1080/02602930301671  

Sadler,  D.  R.  (1989).  Formative  assessment  and  the  design  of  instructional  systems.  Instructional  Science,  18,  119-­‐144.    

Sadler,  D.  R.  (2008).  Transforming  holistic  assessment  and  grading  into  a  vehicle  for  complex  learning.  In  G.  Joughin  (Ed.),  Assessment,  learning  and  judgement  in  higher  education  (pp.  49-­‐64).  Dordrecht:  Springer.  

Sadler,  D.  R.  (2010).  Beyond  feedback:  Developing  student  capability  in  complex  appraisal.  Assessment  &  Evaluation  in  Higher  Education,  35(5),  535-­‐550.  doi:  10.1080/02602930903541015  

Sandhu,  S.  (2004).  The  use  of  statement  banks  as  a  means  of  assessing  large  groups  in  Accounting.  Investigations  in  University  Teaching  and  Learning,  2(1),  49-­‐55.    

Vardi,  I.  (2012).  Effective  feedback  for  student  learning  in  higher  education:  HERDSA.  Weaver,  M.  R.  (2006).  Do  students  value  feedback?  Student  perceptions  of  tutors’  

written  responses.  Assessment  &  Evaluation  in  Higher  Education,  31(3),  379-­‐394.  doi:  10.1080/02602930500353061  

Yorke,  M.  (2003).  Formative  assessment  in  higher  education:  Moves  towards  theory  and  the  enhancement  of  pedagogic  practice.  Higher  Education,  45,  477-­‐501.    

   

   

Addendum:  Update  of  Feedback  Bibliography    4.1.5   Student  engagement  with  feedback    Burns,  C.  A.,  &  Foo,  M.  (2012).  Evaluating  a  formative  feedback  intervention  for  international  students.  Practitioner  Research  in  Higher  Education,  6(1),  40-­‐49.    Burns,  C.,  &  Foo,  M.  (2013).  How  is  feedback  used?–The  international  student  response  to  a  Formative  Feedback  Intervention.  The  International  Journal  of  Management  Education,  11(3),  174-­‐183.    Handley,  M.,  &  Williams,  L.,  (2009).  From  copying  to  learning:  using  exemplars  to  engage  students  with  assessment  criteria  and  feedback.  Assessment  &  Evaluation  in  Higher  Education,  36(1),  95-­‐108.    4.1.6   Student  use  of  feedback    Tian,  M.,  &  Lowe,  J.  (2013).  The  role  of  feedback  in  cross-­‐cultural  learning:  a  case  study  of  Chinese  taught  postgraduate  students  in  a  UK  university.  Assessment  &  Evaluation  in  Higher  Education,  38(5),  580-­‐598.    4.1.7   Student  perceptions  of  tutors’  written  feedback    Agius,  N.  M.,  &  Wilkinson,  A.  (2014).  Students'  and  teachers'  views  of  written  feedback  at  undergraduate  level:  A  literature  review.  Nurse  education  today,  34(4),  552-­‐559.    Ali,  N.,  Rose,  S.,  &  Ahmed,  L.  (2014).  Psychology  students'  perception  of  and  engagement  with  feedback  as  a  function  of  year  of  study.  Assessment  &  Evaluation  in  Higher  Education.  DOI:  10.1080/02602938.2014.936355    Dowden,  T.,  Pittaway,  S.,  Yost,  H.,  &  McCarthy,  R.  (2013).  Students’  perceptions  of  written  feedback  in  teacher  education:  Ideally  feedback  is  a  continuing  two-­‐way  communication  that  encourages  progress.  Assessment  &  Evaluation  in  Higher  Education,  38(3),  349-­‐362.    Murdoch-­‐Eaton,  D.,  &  Sargeant,  J.  (2012).  Maturational  differences  in  undergraduate  medical  students’  perceptions  about  feedback.  Medical  education,  46(7),  711-­‐721.    Pears,  A.,  Harland,  J.,  Hamilton,  M.,  &  Hadgraft,  R.  (2013,  March).  What  is  Feedback?  Connecting  Student  Perceptions  to  Assessment  Practices.  In  Learning  and  Teaching  in  Computing  and  Engineering  (LaTiCE),  2013  (pp.  106-­‐113).  IEEE.    Strong,  J.,  Hughes,  C.,  Wilson,  W.,  Arnott,  W.,  Isles,  R.,  &  Bennison,  A.  (2012).  Perceptions  of  feedback  among  undergraduate  and  postgraduate  students  of  four  health  science  disciplines.  The  Internet  Journal  of  Allied  Health  Sciences  and  Practice,  10(4).          

   

4.1.8   Effectiveness  of  Feedback    Colthorpe,  K.,  Liang,  S.,  &  Zimbardi,  K.  (2013).  Facilitating  timely  feedback  in  the  Biomedical  Sciences.  International  Journal  of  Innovation  in  Science  and  Mathematics  Education  (formerly  CAL-­‐laborate  International),  21(3).    DiBattista,  D.  (2013).  The  Immediate  Feedback  Assessment  Technique:  A  learner-­‐centered  multiple-­‐choice  response  form.  Canadian  Journal  of  Higher  Education,  35(4),  111-­‐131.    Duijnhouwer,  H.,  Prins,  F.  J.,  &  Stokking,  K.  M.  (2012).  Feedback  providing  improvement  strategies  and  reflection  on  feedback  use:  Effects  on  students’  writing  motivation,  process,  and  performance.  Learning  and  Instruction,  22(3),  171-­‐184.    Fernández-­‐Toro,  M.,  Truman,  M.,  &  Walker,  M.  (2013).  Are  the  principles  of  effective  feedback  transferable  across  disciplines?  A  comparative  study  of  written  assignment  feedback  in  Languages  and  Technology.  Assessment  &  Evaluation  in  Higher  Education,  38(7),  816-­‐830.    Field,  D.,  Pulman,  S.,  &  Whitelock,  D.  (2014).  The  pragmatics  of  margin  comments:  An  empirical  study.  ACL  2014,  43.    Gul,  R.,  Lakhani,  A.,  &  Rizvi,  N.  F.  (2014).  Teachers’  Accounts  of  Their  Perceptions  and  Practices  of  Providing  Written  Feedback  to  Nursing  Students  on  Their  Assignments.  International  Journal  of  Higher  Education,  3(3),  p70.    Kishbaugh,  T.  L.,  Cessna,  S.,  Horst,  S.  J.,  Leaman,  L.,  Flanagan,  T.,  Neufeld,  D.  G.,  &  Siderhurst,  M.  (2012).  Measuring  beyond  content:  a  rubric  bank  for  assessing  skills  in  authentic  research  assignments  in  the  sciences.  Chemistry  Education  Research  and  Practice,  13(3),  268-­‐276.    Jones,  O.,  &  Gorra,  A.  (2013).  Assessment  feedback  only  on  demand:  Supporting  the  few  not  supplying  the  many.  Active  Learning  in  Higher  Education,  1469787413481131.    Kerrigan,  M.  J.,  Walker,  S.,  Gamble,  M.,  Lindsay,  S.,  Reader,  K.,  Papaefthimiou,  M.  C.,  ...  &  Saunders,  G.  (2012).  The  Making  Assessment  Count  (MAC)  consortium  maximising  assessment  and  feedback  design  by  working  together.  Research  in  Learning  Technology,  19.    Li,  J.,  &  De  Luca,  R.  (2014).  Review  of  assessment  feedback.  Studies  in  Higher  Education,  39(2),  378-­‐393.    MacWilliam,  T.,  &  Malan,  D.  J.  (2013,  July).  Streamlining  grading  toward  better  feedback.  In  Proceedings  of  the  18th  ACM  conference  on  Innovation  and  technology  in  computer  science  education  (pp.  147-­‐152).  ACM.    Maggs,  L.  A.  (2014).  A  case  study  of  staff  and  student  satisfaction  with  assessment  feedback  at  a  small  specialised  higher  education  institution.  Journal  of  Further  and  Higher  Education,  38(1),  1-­‐18.    

   

Molloy,  E.,  &  Boud,  D.  (2013).  Changing  conceptions  of  feedback.  Feedback  in  higher  and  professional  education:  Understanding  it  and  doing  it  well,  11-­‐33.    Nicol,  D.  (2012).  Resituating  feedback  from  the  reactive  to  the  proactive.  Feedback  in  Higher  and  Professional  Education:  Understanding  it  and  Doing  it  Well,  34.    Price,  M.,  Handley,  K.,  O’Donovan,  B.,  Rust,  C.,  &  Millar,  J.  (2013).  Assessment  feedback.  Reconceptualising  Feedback  in  Higher  Education:  Developing  Dialogue  with  Students,  41.    Ross,  P.  M.,  &  Tronson,  D.  A.  (2012,  November).  Providing  quality  feedback-­‐Where  to  from  here?.  In  Proceedings  of  The  Australian  Conference  on  Science  and  Mathematics  Education  (formerly  UniServe  Science  Conference)  (Vol.  11).    Sutton,  P.  (2012).  Conceptualizing  feedback  literacy:  knowing,  being,  and  acting.  Innovations  in  Education  and  Teaching  International,  49(1),  31-­‐40.    Tang,  J.,  &  Harrison,  C.  (2011).  Investigating  university  tutor  perceptions  of  assessment  feedback:  three  types  of  tutor  beliefs.  Assessment  &  Evaluation  in  Higher  Education,  36(5),  583-­‐604.    Tett,  L.,  Hounsell,  J.,  Christie,  H.,  Cree,  V.  E.,  &  McCune,  V.  (2012).  Learning  from  feedback?  Mature  students’  experiences  of  assessment  in  higher  education.  Research  in  Post-­‐Compulsory  Education,  17(2),  247-­‐260.    Tuck,  J.  (2012).  Feedback-­‐giving  as  social  practice:  teachers’  perspectives  on  feedback  as  institutional  requirement,  work  and  dialogue.  Teaching  in  Higher  Education,  17(2),  209-­‐221.    Venning,  J.,  &  Buisman-­‐Pijlman,  F.  (2013).  Integrating  assessment  matrices  in  feedback  loops  to  promote  research  skill  development  in  postgraduate  research  projects.  Assessment  &  Evaluation  in  Higher  Education,  38(5),  567-­‐579.    4.2.1   Feedback  practice    Barker,  M.,  &  Pinard,  M.  (2014).  Closing  the  feedback  loop?  Iterative  feedback  between  tutor  and  student  in  coursework  assessments.  Assessment  &  Evaluation  in  Higher  Education,  DOI:  10.1080/02602938.2013.875985.    Burke,  D.,  &  Pieterick,  J.  (2010).  Giving  students  effective  written  feedback.  Maidenhead:  Open  University  Press/McGraw  Hill.    Burke,  D.  (2011).  Now  I’ve  got  the  feedback,  what  do  I  do  with  it?  Strategies  for  students  to  get  more  out  of  tutor  feedback.  Practice  and  Evidence  of  Scholarship  of  Teaching  and  Learning  in  Higher  Education,  6(1),  43-­‐60.    Cathcart,  A.,  Greer,  D.,  &  Neale,  L.  (2013).  Learner-­‐focused  evaluation  cycles:  facilitating  learning  using  feedforward,  concurrent  and  feedback  evaluation.  Assessment  &  Evaluation  in  Higher  Education,  790-­‐802.  DOI:  10.1080/02602938.2013.870969    

   

Dube,  C.,  Kane,  S.,  &  Lear,  M.  (2012).  The  effectiveness  of  students  redrafting  continuous  assessment  tasks:  the  pivotal  role  of  tutors  and  feedback.  Perspectives  in  Education,  30(3),  50-­‐59.    East,  M.,  Bitchener,  J.,  &  Basturkmen,  H.  (2012).  What  Constitutes  Effective  Feedback  to  Postgraduate  Research  Students?  The  Students'  Perspective.  Journal  of  University  Teaching  and  Learning  Practice,  9(2),  7.    Freeman,  R.,  &  Dobbins,  K.  (2013).  Are  we  serious  about  enhancing  courses?  Using  the  principles  of  assessment  for  learning  to  enhance  course  evaluation.  Assessment  &  Evaluation  in  Higher  Education,  38(2),  142-­‐151.    Hughes,  G.,  Wood,  E.,  &  Kitagawa,  K.  (2014).  Use  of  self-­‐referential  (ipsative)  feedback  to  motivate  and  guide  distance  learners.  Open  Learning:  The  Journal  of  Open,  Distance  and  e-­‐Learning,  29(1),  31-­‐44.    Jonsson,  A.  (2012).  Facilitating  productive  use  of  feedback  in  higher  education.  Active  Learning  in  Higher  Education,  1469787412467125.    Lilly,  J.,  Richter,  I.  U.,  &  Macias,  B.  R.  (2009).  Feeding  Back,  Feeding  Forward:  An  Exchange  of  Ideas.  Networks,  77.    Nordrum,  L.,  Evans,  K.,  &  Gustafsson,  M.  (2013).  Comparing  student  learning  experiences  of  in-­‐text  commentary  and  rubric-­‐articulated  feedback:  strategies  for  formative  assessment.  Assessment  &  Evaluation  in  Higher  Education,  38(8),  919-­‐940.    Scoles,  J.,  Huxham,  M.,  &  McArthur,  J.  (2013).  No  longer  exempt  from  good  practice:  using  exemplars  to  close  the  feedback  gap  for  exams.  Assessment  &  Evaluation  in  Higher  Education,  38(6),  631-­‐645.    4.2.3   Student  engagement  with  feedback    Boud,  D.,  Lawson,  R.,  &  Thompson,  D.  G.  (2013).  Does  student  engagement  in  self-­‐assessment  calibrate  their  judgement  over  time?.  Assessment  &  Evaluation  in  Higher  Education,  38(8),  941-­‐956.    Boud,  D.,  &  Molloy,  E.  (2013).  Rethinking  models  of  feedback  for  learning:  the  challenge  of  design.  Assessment  &  Evaluation  in  Higher  Education,  38(6),  698-­‐712.    Cartney,  P.  (2010).  Exploring  the  use  of  peer  assessment  as  a  vehicle  for  closing  the  gap  between  feedback  given  and  feedback  used.  Assessment  &  Evaluation  in  Higher  Education,  35(5),  551-­‐564.    Donovan,  P.  (2014).  Closing  the  feedback  loop:  physics  undergraduates’  use  of  feedback  comments  on  laboratory  coursework.  Assessment  &  Evaluation  in  Higher  Education,  (ahead-­‐of-­‐print),  1-­‐13.    Fleckhammer,  L.,  &  Wise,  L.  (2011).  The  role  of  tutors  in  facilitating  online  student  engagement.  Changing  Demands,  Changing  Directions.  Proceedings  ascilite  Hobart,  392-­‐397.    

   

Giles,  T.  M.,  Gilbert,  S.,  &  McNeill,  L.  (2014).  Do  students  want  and  need  written  feedback  on  summative  assignments?  Engaging  students  with  the  feedback  process-­‐a  topic  review  activity.  ergo,  3(1).    McMahon,  T.  (2010).  Using  peer-­‐assessment  to  generate  high-­‐quality  and  proximate  feedback  and  feed-­‐forward  on  students’  work.  ICERI2010  Proceedings,  3422-­‐3423.    Nicol,  D.  (2013).  Assessment  and  Feedback  In  the  hands  of  the  student.  JISC  e-­‐Learning  Programme  Webinar.    Nicol,  D.,  Thomson,  A.,  &  Breslin,  C.  (2014).  Rethinking  feedback  practices  in  higher  education:  a  peer  review  perspective.  Assessment  &  Evaluation  in  Higher  Education,  39(1),  102-­‐122.    Robinson,  S.,  Pope,  D.,  &  Holyoak,  L.  (2013).  Can  we  meet  their  expectations?  Experiences  and  perceptions  of  feedback  in  first  year  undergraduate  students.  Assessment  &  Evaluation  in  Higher  Education,  38(3),  260-­‐272.    Rubin,  R.  F.,  &  Turner,  T.  (2012).  Student  performance  on  and  attitudes  toward  peer  assessments  on  Advanced  Pharmacy  Practice  Experience  assignments.  Currents  in  Pharmacy  Teaching  and  Learning,  4(2),  113-­‐121.    4.2.4   Dialogic  approaches  to  feedback    Blair,  A.,  Curtis,  S.,  Goodwin,  M.,  &  Shields,  S.  (2013).  What  feedback  do  students  want?  Politics,  33(1),  66-­‐79.      Blair,  A.,  &  McGinty,  S.  (2013).  Feedback-­‐dialogues:  exploring  the  student  perspective.  Assessment  &  Evaluation  in  Higher  Education,  38(4),  466-­‐476.    Blair,  A.,  Curtis,  S.,  Goodwin,  M.,  &  Shields,  S.  (2013).  The  significance  of  assignment  feedback:  from  consumption  to  construction.  European  Political  Science,  12(2),  231-­‐244.    Blair,  A.,  Wyburn-­‐Powell,  A.,  Goodwin,  M.,  &  Shields,  S.  (2014).  Can  dialogue  help  to  improve  feedback  on  examinations?  Studies  in  Higher  Education,  39(6),  1039-­‐1054.    Carless,  D.  (2013).  Trust  and  its  role  in  facilitating  dialogic  feedback.  Feedback  in  higher  and  professional  education,  90-­‐103.    Merry,  S.,  Price,  M.,  Carless,  D.,  &  Taras,  M.  (Eds.).  (2013).  Reconceptualising  Feedback  in  Higher  Education:  Developing  Dialogue  with  Students.  Routledge.    Yang,  M.,  &  Carless,  D.  (2013).  The  feedback  triangle  and  the  enhancement  of  dialogic  feedback  processes.  Teaching  in  Higher  Education,  18(3),  285-­‐297.    4.2.5   Using  technology  to  provide  peer  and  marker  feedback    Allan,  R.,  &  Bentley,  S.  (2012).  Feedback  mechanisms:  efficient  and  effective  use  of  technology  or  a  waste  of  time  and  effort?  eprints.hud.ac.uk    

   

Alvarez,  I.,  Espasa,  A.,  &  Guasch,  T.  (2012).  The  value  of  feedback  in  improving  collaborative  writing  assignments  in  an  online  learning  environment.  Studies  in  Higher  Education,  37(4),  387-­‐400.    Attenborough,  J.,  Gulati,  S.,  &  Abbott,  S.  (2012).  Audio  feedback  on  student  assignments:  boon  or  burden?  Learning  at  City  Journal,  2(2).  openaccess.city.ac.uk.    Auld,  G.,  Ridgway,  A.,  &  Williams,  J.  (2013).  Digital  oral  feedback  on  written  assignments  as  professional  learning  for  teacher  educators:  A  collaborative  self-­‐study.  Studying  Teacher  Education,  9(1),  31-­‐44.    Bayerlein,  L.  (2014).  Students’  feedback  preferences:  How  do  students  react  to  timely  and  automatically  generated  assessment  feedback?  Assessment  &  Evaluation  in  Higher  Education,  DOI:  10.1080/02602938.2013.870531.    Berlanga,  A.  J.,  Van  Rosmalen,  P.,  Boshuizen,  H.  P.  A.,  &  Sloep,  P.  B.  (2012).  Exploring  formative  feedback  on  textual  assignments  with  the  help  of  automatically  created  visual  representations.  Journal  of  Computer  Assisted  Learning,  28(2),  146–160.  doi:  10.1111/j.1365-­‐2729.2011.00425.    Borup,  J.,  West,  R.  E.,  Thomas,  R.,  &  Graham,  C.  R.  (2014).  Examining  the  impact  of  video  feedback  on  instructor  social  presence  in  blended  courses.  The  International  Review  of  Research  in  Open  and  Distance  Learning,  15(3).  irrodl.org    Bourgault,  A.  M.,  Mundy,  C.,  &  Joshua,  T.  (2013).  Comparison  of  audio  vs.  written  feedback  on  clinical  assignments  of  nursing  students.  Nursing  Education  Perspectives,  34(1),  43-­‐46.  nlnjournals.org.    Brearley,  F.  Q.,  &  Cullen,  W.  R.  (2012).  Providing  students  with  formative  audio  feedback.  -­‐  Bioscience  Education,  20,  22-­‐36.    Carruthers,  C.,  McCarron,  B.,  Bolan,  P.,  Devine,  A.,  &  McMahon-­‐Beattie,  U.  (2014).  Listening  and  Learning:  Reflections  on  the  use  of  Audio  Feedback.  An  Excellence  in  Teaching  and  Learning  Note.  Business  and  Management  Education  in  HE,  1(1),  4-­‐11.    Carruthers,  C.,  McCarron,  B.,  Bolan,  P.,  Devine,  A.,  McMahon-­‐Beattie,  U.,  &  Burns,  A.  (2014).  ‘I  like  the  sound  of  that’–an  evaluation  of  providing  audio  feedback  via  the  virtual  learning  environment  for  summative  assessment.  Assessment  &  Evaluation  in  Higher  Education,  DOI:  10.1080/02602938.2014.917145.    Chang,  N.  (2012).  Pre-­‐Service  Teachers'  Views:  How  Did  E-­‐feedback  through  Assessment  Facilitate  Their  Learning?  Journal  of  the  Scholarship  of  Teaching  and  Learning,  11,  16-­‐33.  josotl.indiana.edu    Chang,  N.,  Watson,  A.  B.,  Bakerson,  M.  A.,  Williams,  E.  E.,  McGoron,  F.  X.,  &  Spitzer,  B.  (2012).  Electronic  feedback  or  handwritten  feedback:  What  do  undergraduate  students  prefer  and  why?  Journal  of  Teaching  and  Learning  with  Technology,  1(1),  1-­‐23.    Crook,  A.,  Mauchline,  A.,  Maw,  S.,  Lawson,  C.,  Drinkwater,  R.,  Lundqvist,  K.,  ...  &  Park,  J.  (2012).  The  use  of  video  technology  for  providing  feedback  to  students:  Can  it  

   

enhance  the  feedback  experience  for  staff  and  students?.  Computers  &  Education,  58(1),  386-­‐396.    Edwards,  K.,  Dujardin,  A.  F.,  &  Williams,  N.  (2012).  Screencast  Feedback  for  Essays  on  a  Distance  Learning  MA  in  Professional  Communication.  Journal  of  Academic  Writing,  2(1),  95-­‐126.    Ferrell,  G.,  &  Sheppard,  M.  (2013).  Supporting  assessment  and  feedback  practice  with  technology:  a  view  of  the  UK  landscape.  jiscdesignstudio.pbworks.com    Frank,  B.  M.,  &  Hamgini,  B.  B.  (2014).  Collaborative  cloud-­‐based  documents  for  real-­‐time  bi-­‐directional  feedback  in  large  lecture  activities.  121st  ASEE  Annual  Conference  &  Exposition.    Gleaves,  A.,  &  Walker,  C.  (2013).  Richness,  redundancy  or  relational  salience?  A  comparison  of  the  effect  of  textual  and  aural  feedback  modes  on  knowledge  elaboration  in  higher  education  students'  work.  Computers  &  Education,  62,  249-­‐261.    Gould,  J.,  &  Day,  P.  (2013).  Hearing  you  loud  and  clear:  student  perspectives  of  audio  feedback  in  higher  education.  Assessment  &  Evaluation  in  Higher  Education,  38(5),  554-­‐566.    Harper,  F.,  Green,  H.,  &  Fernandez-­‐Toro,  M.  (2012,  September).  Evaluating  the  integration  of  Jing®  screencasts  in  feedback  on  written  assignments.  In  Interactive  Collaborative  Learning  (ICL),  2012  15th  International  Conference  on  (pp.  1-­‐7).  IEEE.    Hennessy,  C.,  &  Forrester,  G.  (2014).  Developing  a  framework  for  effective  audio  feedback:  a  case  study.  Assessment  and  Evaluation  in  Higher  Education,  39(7),  777-­‐789.    Hepplestone,  S.,  Holden,  G.,  Irwin,  B.,  Parkin,  H.  J.,  &  Thorpe,  L.  (2011).  Using  technology  to  encourage  student  engagement  with  feedback:  a  literature  review.  Research  in  Learning  Technology,  19(2).    Jones,  N.,  Georghiades,  P.,  &  Gunson,  J.  (2012).  Student  feedback  via  screen  capture  digital  video:  stimulating  student’s  modified  action.  Higher  Education,  64(5),  593-­‐607.    Laughton,  D.  J.  (2013).  Using  audio  feedback  to  enhance  assessment  practice–an  evaluation  of  student  and  tutor  perspectives.  Student  Engagement  and  Experience  Journal,  2(2).    Lee,  K.  (2012).  Technology-­‐mediated  feedback.  Quick  Hits  for  Teaching  with  Technology:  Successful  Strategies  by  Award-­‐Winning  Teachers,  80.    Malau-­‐Aduli,  B.  S.,  Assenheimer,  D.,  Choi-­‐Lundberg,  D.,  &  Zimitat,  C.  (2013).  Using  computer-­‐based  technology  to  improve  feedback  to  staff  and  students  on  MCQ  assessments.  Innovations  in  Education  and  Teaching  International,  51(5),  510-­‐522.    Marriott,  P.,  &  Teoh,  L.  K.  (2012).  Using  screencasts  to  enhance  assessment  feedback:  students’  perceptions  and  preferences.  Accounting  Education,  21(6),  583-­‐598.  

   

 Mathisen,  P.  (2012).  Video  Feedback  in  Higher  Education–A  Contribution  to  Improving  the  Quality  of  Written  Feedback.  Nordic  Journal  of  Digital  Literacy,  7(02),  97-­‐112.    McGregor,  K.,  Merchant,  A.  R.,  &  Butler,  M.  (2012).  Making  feedback  more  immediate.  International  Journal  of  Innovation  in  Science  and  Mathematics  Education  (formerly  CAL-­‐laborate  International),  16(1).    Morrissey,  J.  (2012).  Podcast  steering  of  independent  learning  in  higher  education.  AISHE-­‐J:  The  All  Ireland  Journal  of  Teaching  and  Learning  in  Higher  Education,  4(1).    Mulder,  R.  A.,  Elgar,  M.  A.,  &  Brady,  D.  (2012,  October).  APRES:  Electronically  managed  student  feedback  via  peer  review.  In  Proceedings  of  The  Australian  Conference  on  Science  and  Mathematics  Education  (formerly  UniServe  Science  Conference)  (Vol.  11).    Parkin,  H.  J.,  Hepplestone,  S.,  Holden,  G.,  Irwin,  B.,  &  Thorpe,  L.  (2012).  A  role  for  technology  in  enhancing  students’  engagement  with  feedback.  Assessment  &  Evaluation  in  Higher  Education,  37(8),  963-­‐973.    Ribchester,  C.,  France,  D.,  &  Wheeler,  A.  (2007).  Podcasting:  a  tool  for  enhancing  assessment  feedback.  Education  in  a  Changing  Environment.    Roberts,  S.  (2008).  Pod  casting  feedback  to  students:  Students’  perceptions  of  effectiveness.  Innovations  in  Practice,  44.    Shannon,  S.,  Francis,  R.,  &  Torpey,  G.  (2012,  November).  Barriers  to  adoption  of  blended  learning  and  online  feedback  and  assessment  by  sessional  staff.  In  Paper  accepted  for  ASA  2012  Conference,  Griffith  University,  Gold  Coast  (pp.  14-­‐16).    Singh,  R.,  Gulwani,  S.,  &  Solar-­‐Lezama,  A.  (2013,  June).  Automated  feedback  generation  for  introductory  programming  assignments.  In  ACM  SIGPLAN  Notices  (Vol.  48,  No.  6,  pp.  15-­‐26).  ACM.    Thompson,  R.,  &  Lee,  M.  J.  (2012).  Talking  with  students  through  screencasting:  experimentations  with  video  feedback  to  improve  student  learning.  The  Journal  of  Interactive  Technology  and  Pedagogy,  1(1).    Tong,  V.  C.  (2014).  Towards  technology-­‐and  research-­‐enhanced  education  (TREE):  electronic  feedback  as  a  teaching  tool  in  geoscience.  In  Geoscience  Research  and  Education  (pp.  199-­‐212).  Springer  Netherlands.    Upson-­‐Saia,  K.,  &  Scott,  S.  (2013).  Using  iAnnotate  to  enhance  feedback  on  written  work.  Journal  of  Teaching  and  Learning  with  Technology,  2(2),  43-­‐59.    Walet,  N.,  &  Birch,  M.  (2012,  May).  Using  online  assessment  to  provide  instant  feedback.  In  Proceedings  of  the  HEA-­‐STEM  Annual  Conference,  London.    

   

Walkington,  H.  (2012).  Developing  dialogic  learning  space:  The  case  of  online  undergraduate  research  journals.  Journal  of  Geography  in  Higher  Education,  36(4),  547-­‐562.    Wallace,  I.,  &  Moore,  C.  (2012,  January).  Providing  in-­‐depth  and  personalized  feedback  to  online  students  using  audio  recording.  In  3rd  International  Conference  on  e-­‐education,  e-­‐Business,  e-­‐Management,  and  e-­‐Learning.    Willis,  L.,  &  Webb,  A.  (2010).  Enhancing  feedback  for  engineering  students.  ©  Higher  Education  Academy  Engineering  Subject  Centre,  Loughborough  University.    Wong,  S.  H.  S.,  &  Beaumont,  A.  J.  (2012).  A  quest  for  helpful  feedback  to  programming  coursework.  engineering  education,  7(2),  51-­‐62.    4.2.6   Dealing  with  large  classes    Kulkarni,  C.,  Wei,  K.  P.,  Le,  H.,  Chia,  D.,  Papadopoulos,  K.,  Cheng,  J.,  ...  &  Klemmer,  S.  R.  (2013).  Peer  and  self  assessment  in  massive  online  classes.  ACM  Transactions  on  Computer-­‐Human  Interaction  (TOCHI),  20(6),  33.    Rienties,  B.,  Willis,  A.,  Alcott,  P.,  &  Medland,  E.  (2013).  Student  Experiences  of  Self-­‐Reflection  and  Peer  Assessment  in  Providing  Authentic  Project-­‐Based  Learning  to  Large  Class  Sizes.  In  Facilitating  learning  in  the  21st  century:  leading  through  technology,  diversity  and  authenticity  (pp.  117-­‐136).  Springer  Netherlands.    4.2.7   Feedforward    Backstrom,  M.  (2012).  Enhancing  opportunities  to  feed  forward;  developing  problem  solving  skills  using  web  based  scenarios.  eprints.qut.edu.au    Baker,  D.  J.,  &  Zuvela,  D.  (2013).  Feedforward  strategies  in  the  first-­‐year  experience  of  online  and  distributed  learning  environments.  Assessment  &  Evaluation  in  Higher  Education,  38(6),  687-­‐697.    Bird,  F.  L.,  &  Yucel,  R.  (2014).  Feedback  codes  and  action  plans:  building  the  capacity  of  first-­‐year  students  to  apply  feedback  to  a  scientific  report.  Assessment  &  Evaluation  in  Higher  Education,  DOI:  10.1080/02602938.2014.924476    Bennett,  S.,  &  Barker,  T.  (2013).  Helping  students  self  monitor  through  EVS  based  feed-­‐forward  exercises.  In  e-­‐Learning  and  e-­‐Technologies  in  Education  (ICEEE),  2013  Second  International  Conference  (pp.  1-­‐6).  IEEE.    Deyi,  S.  (2011).  Feedback  or  feed-­‐forward?.  Changing  Spaces:  Writing  Centres  and  Access  to  Higher  Education,  47.    Duncan,  N.  (2007).  ‘Feed-­‐forward’:  improving  students'  use  of  tutors'  comments.  Assessment  &  Evaluation  in  Higher  Education,  32(3),  271-­‐283.    Fladd,  N.  (2011).  Getting  Feedback  to  Feed  Forward:  Incorporating  revision  into  upper-­‐year  English  papers.  Teaching  Innovation  Projects,  1(2),  3.    

   

Greaves,  L.  (2012).  Feed-­‐forward  for  Informed  Learning  (FfIL):  learner-­‐generated  materials  for  personalised  learning.  Enhancing  Learning  in  the  Social  Sciences,  4(3).    Moore,  C.,  &  Wallace,  I.  P.  Personalizing  Feedback  for  Feed-­‐Forward  Opportunities  Utilizing  Audio  Feedback  Technologies  for  Online  Students.    Moore,  S.  A.  (2010)  Does  a  Feedforward  Approach  Improve  Students’  Assignment  Grades?  Innovations  in  Practice,  2(2),  112-­‐122.    Morrell,  L.  J.  (2013).  Use  of  Feed-­‐forward  Mechanisms  in  a  Novel  Research-­‐led  Module.  Bioscience  Education,  (0).    Murtagh,  L.,  &  Baker,  N.  (2009).  Feedback  to  feed  forward:  Student  response  to  tutors’  written  comments  on  assignments.  Practitioner  research  in  higher  education,  3(1),  20-­‐28.    Orsmond,  P.,  Maw,  S.  J.,  Park,  J.  R.,  Gomez,  S.,  &  Crook,  A.  C.  (2013).  Moving  feedback  forward:  theory  to  practice.  Assessment  &  Evaluation  in  Higher  Education,  38(2),  240-­‐252.    Osado,  A.  G.,  Merlo,  L.  S.  B.,  &  del  Campo,  M.  H.  D.  (2013).  Formative  assessment  in  a  Chemistry  course  through  feedback  andfeed-­‐forward  strategies.  INTED2013  Proceedings,  3852-­‐3860.    Pears,  A.,  Harland,  J.,  Hamilton,  M.,  &  Hadgraft,  R.  (2014,  April).  Four  Feed-­‐Forward  Principles  Enhance  Students'  Perception  of  Feedback  as  Meaningful.  In  Teaching  and  Learning  in  Computing  and  Engineering  (LaTiCE),  2014  International  Conference  on  (pp.  272-­‐277).  IEEE.    Quinton,  S.,  &  Smallbone,  T.  (2010).  Feeding  forward:  using  feedback  to  promote  student  reflection  and  learning–a  teaching  model.  Innovations  in  Education  and  Teaching  International,  47(1),  125-­‐135.    Rodway-­‐Dyer,  S.,  &  Dunne,  E.  (2009).  Technology  enhanced  feed-­‐forward  for  learning.  HEA  Final  Report.    Robson,  S.,  Leat,  D.,  Wall,  K.,  &  Lofthouse,  R.  (2012).  Feedback  or  feed  forward?  Supporting  master’s  students  through  effective  assessment  to  enhance  future  learning.  Cross  cultural  teaching  and  learning  for  home  and  international  students:  Internationalisation,  Pedagogy  and  Curriculum  in  Higher  Education:  Internationalisation  of  Pedagogy  and  Curriculum  in  Higher  Education,  53.    Taylor,  C.  (2007).  Integrating  Feedforward  on  Academic  Writing  into  an  Undergraduate  Science  Course.  Assessment  design  for  learner  responsibility  29-­‐31  May  07  http://www.reap.ac.uk    Vardi,  I.  (2013).  Effectively  feeding  forward  from  one  written  assessment  task  to  the  next.  Assessment  &  Evaluation  in  Higher  Education,  38(5),  599-­‐610.    

   

Walker,  S.,  &  Hobson,  J.  (2014).  Interventions  in  teaching  first-­‐year  law:  Feeding  forward  to  improve  learning  outcomes.  Assessment  &  Evaluation  in  Higher  Education,  39(3),  326-­‐338.    Wimshurst,  K.,  &  Manning,  M.  (2013).  Feed-­‐forward  assessment,  exemplars  and  peer  marking:  evidence  of  efficacy.  Assessment  &  Evaluation  in  Higher  Education,  38(4),  451-­‐465.    Arts  and  Social  Sciences:    Wakefield,  C.,  Adie,  J.,  Pitt,  E.,  &  Owens,  T.  (2014).  Feeding  forward  from  summative  assessment:  the  Essay  Feedback  Checklist  as  a  learning  tool.  Assessment  &  Evaluation  in  Higher  Education,  39(2),  253-­‐262.    Walker,  S.,  &  Hobson,  J.  (2014).  Interventions  in  teaching  first-­‐year  law:  Feeding  forward  to  improve  learning  outcomes.  Assessment  &  Evaluation  in  Higher  Education,  39(3),  326-­‐338.  

SCIENCE

CONTACT Meloni M. Muir, PhD Discipline of Physiology Sydney Medical School The University of Sydney NSW 2006 E [email protected]