CSC/ECE 517 Spring 2018- Project E1819: Improve self review, link self and peer review to derive grades: Difference between revisions

From Expertiza_Wiki
Jump to navigation Jump to search
Line 36: Line 36:


== Score Formulas ==
== Score Formulas ==
Several potential formulas to calculate the derived score between self and peer-reviews were discussed.
There are many possible methods of deriving a score between self and peer-reviews. Each formula has the same goal, to derive a meaningful score between the peer and self-reviews. Thus, giving feedback to students on their self-reviewing, hopefully increasing their abilities in this area. Several potential approaches are available, each with different pros and cons. The simpler of approaches involves only the self and peer-review scores, and the more complex approaches may involve several other factors.
A naive approach is to find the difference between the max score and the difference between self and peer-reviews.
 
The simplest approach is to simply take the difference between the average peer-review score, and the self-review score. This rewards the students for good self-reviewing  ability, and evenly punishes students for over and underrating themselves. A similar approach is utilized in several other systems, in which when a students self-review score is within 5% of their average peer-review score, they will receive the greater of the two grades. An extension of this can be made to allow a weight to be applied to the difference between the grades, and to allow instructors to manually enter the weight for each assignment.


[[File:Design Doc Naive Formula.PNG]]
[[File:Design Doc Naive Formula.PNG]]


A slight change to this approach is to take into account the peer score, so the greater your peers believe you performed, the greater your score is (proportional to the difference between reviews still). However, this approach can prove to be harmful. Below it can be seen that even when you perfectly reflect the score of your peers your grade can be very low. This approach punishes students twice for low peer-review grades, and it should be avoided.
A slight change to this approach is to take into account the peer score, so the greater your peers believe you performed, the greater your score is (proportional to the difference between reviews again). However, this, and similar approaches, should be avoided as they are potentially harmful. In the chart below, it can be seen that students with poor performance will be harshly punished for poor performance in their peer-reviews, their self and peer-review derived grade, and finally in their instructor grade. The self and peer-review derived grade should only be indicative of the students ability to self-review, not of the quality of the project.  


[[File:Design Doc Take Score Into Account.PNG]]
[[File:Design Doc Take Score Into Account.PNG]]


A third approach is discussed [https://www.ojs.unisa.edu.au/index.php/atna/article/viewFile/343/258 here.] Which results in a score that should accurately reflect the relation between self and peer scores. However, it can again be seen that when peer-review scores are low students are punished twice unless they can accurately reflect their peer's reviews.
Another approach is to derive the grade based on the percent difference between the self-review score, and the peer-review scores. This results in a smooth gradient which does not punish students as harshly as the previous two approaches. However, at low scores students will see greater swings in this derived grade.
 
[[File:Design Doc Pecentage Diff.PNG]]
 
These previous approaches involve only the peer and self-review scores, and are the simple approaches to the problem, the following options involve greater complexity, but seek to better solve the problem at hand.
 
The SPARK approach is discussed [https://www.ojs.unisa.edu.au/index.php/atna/article/viewFile/343/258 here.] Which results in a score that should accurately reflect the relation between self and peer scores. The scores here will need to be adjusted to match the grading scheme of the project, a score of 1 is equivalent to a 100%, and the difference between 1 and the score is representative of a lower grade, regardless of the score being above or below 1. This method also includes a second score titled SAPA, Self-Assessment to Peer-Assessment factor. This score requires teams to evaluate each team member on the same self-review rubric. After this is completed team member's self-review scores are divided by their average teammate-review score to create the SAPA. A SAPA of > 1.0 indicates that the student rated themselves higher than their team mates did, a score of less than 1.0 indicates that students underrated themselves compared to their peers.


[[File:Design Doc Spark Formula.PNG]]
[[File:Design Doc Spark Formula.PNG]]


A fourth approach is to derive the grade based on the percent difference between the self-review score, and the peer-review scores. This results in a smooth gradient which does not punish students as harshly as the previous two approaches. However, at low scores students will see greater swings in this derived grade.  
Two other approaches involving teammate participation are ranking and FCS(Fair Contribution Scoring).
 
In the ranking approach, students are asked to rank their team, including themselves, from 1 to n (where n is equal to team size), where 1 is the highest score, and n is the lowest score. Students are then assessed based on their average rank from teammates and the rank they awarded themselves. This approach is simple, but it forgoes the rubric approach that garners greater self-reflection in specific areas.
 
In the FCS approach, students are given a finite number of points to award to their teammates, including themselves. The more points that a person receives, the greater their perceived work contribution. At the end of the project students are assessed based on their average number of points received, and the points awarded to themselves. While the approach is very similar to the ranking approach there are important differences. With this approach students will need to weight the contributions of each member on a percentage of points basis, rather than on a, teammate A vs teammate B overall basis.  


[[File:Design Doc Pecentage Diff.PNG]]
{| class="wikitable"
|+ Approach Pros and Cons
! Approach
! Pros
! Cons
|-
! Simple approaches
| Simple to implement, scoring is quick, direct feedback based only on student's self-review ability
| Does not take into account teammate assessments, has potential pitfalls at low peer-review scores
|-
! SPARK approach
| Simple to implement, includes SAPA which shows how self-reviews compare to teammate-reviews
| A separate metric must be calculated, stored, and displayed, team-reviews will slow the process
|-
! Ranking approach
| A very simple approach, can be implemented similar to the bidding page
| Loss of some self-reflection without a rubric
|-
! FCS approach
| A slightly improved ranking approach, scores may be more accurate than just ranking
| Loss of some self-reflection without a rubric
|}


== Testing ==
== Testing ==

Revision as of 03:31, 14 April 2018

Introduction and Problem Statement

Expertiza has the ability to require self-reviews on projects, however, these self-review scores do not have much functionality. In the current system, self-review scores are created and stored, but are not used afterwards. This project seeks to improve upon the self-review system, and extend its functionality. The main points of this project are as follows:

  • Create a method to calculate a derived score between self and peer-reviews. The closer the self-review is to the peer-review, the greater the score. Display this score on the "View Scores" page.
  • Allow instructors to decide if peer-reviews should be visible before self-review completion. Display the correct review scores accordingly.
  • Display the self-review scores in the "View Scores" and heat map of peer-reviews, denoting that they are a different type of review.

Planned Approach

Allow instructors to decide if students should be allowed to view peer-reviewers before completing their self-review.
Currently when instructors create a new assignment, or edit an existing one, they have the option to require self-reviews.
This will be updated to ask instructors if peer-reviews should be visible before self-reviews are completed.
When checked the system will check if a student has completed a self-review when they access the scores view. If they have not completed their self-review they will see a message indicating such. If they have completed the review, they will see their peer-reviews as normal.
Find or create a function that will derive a score from self-reviews and peer-reviews.
This score should be reflective of the difference between the students self-review score, and their peer's reviews. The purpose behind this score is to teach students to become better at reviewing their own work. By providing a score that reflects how similar their reviews are to their peer's reviews, students receive feedback on their self-assessment abilities. More information on the function creation/decision proccess can be found in the section below.
Create the different displays for self-reviews
Instructors and students both have a heat-map view of scores from peer-reviews
(Instructor view pictured) student's self-review scores will be added to this view, but need to be denoted in a way that makes it easy to see this is a self-review not a peer-review. One potential option is to use an icon that shows this is a self-review score.
Which makes it simple for users to understand that particular column represents the scores from a self-review. Other potential options are to title the column "Self-review", or to highlight/border the column with a particular color (ex. blue), both denoting that the score comes from a self-review. Similar approaches can be taken to differentiate between self-review and peer-review scores in the regular "view scores" page.

Score Formulas

There are many possible methods of deriving a score between self and peer-reviews. Each formula has the same goal, to derive a meaningful score between the peer and self-reviews. Thus, giving feedback to students on their self-reviewing, hopefully increasing their abilities in this area. Several potential approaches are available, each with different pros and cons. The simpler of approaches involves only the self and peer-review scores, and the more complex approaches may involve several other factors.

The simplest approach is to simply take the difference between the average peer-review score, and the self-review score. This rewards the students for good self-reviewing ability, and evenly punishes students for over and underrating themselves. A similar approach is utilized in several other systems, in which when a students self-review score is within 5% of their average peer-review score, they will receive the greater of the two grades. An extension of this can be made to allow a weight to be applied to the difference between the grades, and to allow instructors to manually enter the weight for each assignment.

A slight change to this approach is to take into account the peer score, so the greater your peers believe you performed, the greater your score is (proportional to the difference between reviews again). However, this, and similar approaches, should be avoided as they are potentially harmful. In the chart below, it can be seen that students with poor performance will be harshly punished for poor performance in their peer-reviews, their self and peer-review derived grade, and finally in their instructor grade. The self and peer-review derived grade should only be indicative of the students ability to self-review, not of the quality of the project.

Another approach is to derive the grade based on the percent difference between the self-review score, and the peer-review scores. This results in a smooth gradient which does not punish students as harshly as the previous two approaches. However, at low scores students will see greater swings in this derived grade.

These previous approaches involve only the peer and self-review scores, and are the simple approaches to the problem, the following options involve greater complexity, but seek to better solve the problem at hand.

The SPARK approach is discussed here. Which results in a score that should accurately reflect the relation between self and peer scores. The scores here will need to be adjusted to match the grading scheme of the project, a score of 1 is equivalent to a 100%, and the difference between 1 and the score is representative of a lower grade, regardless of the score being above or below 1. This method also includes a second score titled SAPA, Self-Assessment to Peer-Assessment factor. This score requires teams to evaluate each team member on the same self-review rubric. After this is completed team member's self-review scores are divided by their average teammate-review score to create the SAPA. A SAPA of > 1.0 indicates that the student rated themselves higher than their team mates did, a score of less than 1.0 indicates that students underrated themselves compared to their peers.

Two other approaches involving teammate participation are ranking and FCS(Fair Contribution Scoring).

In the ranking approach, students are asked to rank their team, including themselves, from 1 to n (where n is equal to team size), where 1 is the highest score, and n is the lowest score. Students are then assessed based on their average rank from teammates and the rank they awarded themselves. This approach is simple, but it forgoes the rubric approach that garners greater self-reflection in specific areas.

In the FCS approach, students are given a finite number of points to award to their teammates, including themselves. The more points that a person receives, the greater their perceived work contribution. At the end of the project students are assessed based on their average number of points received, and the points awarded to themselves. While the approach is very similar to the ranking approach there are important differences. With this approach students will need to weight the contributions of each member on a percentage of points basis, rather than on a, teammate A vs teammate B overall basis.

Approach Pros and Cons
Approach Pros Cons
Simple approaches Simple to implement, scoring is quick, direct feedback based only on student's self-review ability Does not take into account teammate assessments, has potential pitfalls at low peer-review scores
SPARK approach Simple to implement, includes SAPA which shows how self-reviews compare to teammate-reviews A separate metric must be calculated, stored, and displayed, team-reviews will slow the process
Ranking approach A very simple approach, can be implemented similar to the bidding page Loss of some self-reflection without a rubric
FCS approach A slightly improved ranking approach, scores may be more accurate than just ranking Loss of some self-reflection without a rubric

Testing

Test cases are as follows:

  • Peer-reviews cannot be viewed before a self-review is completed.
  • Peer-reviews can be viewed immediately when self-reviews are not required.
  • The derived scores can be saved to the database.
  • Self-review scores are displayed with peer-review scores
  • Student can only see their own self-review

Files Changed

Reference

Our repo link

our pull request link

Edward F. Gehringer, "Self-assessment to improve learning and evaluation," ASEE Annual Conference and Exposition, Columbus, OH, June 25–28, 2017