analysis of the peerrank
play

Analysis of the PeerRank Method for Peer Grading Joshua Kline - PowerPoint PPT Presentation

Analysis of the PeerRank Method for Peer Grading Joshua Kline Advisors: Matthew Anderson and William Zwicker Benefits of Peer Grading Reduces time instructors spend grading Provides faster feedback for students Increases student


  1. Analysis of the PeerRank Method for Peer Grading Joshua Kline Advisors: Matthew Anderson and William Zwicker

  2. Benefits of Peer Grading • Reduces time instructors spend grading • Provides faster feedback for students • Increases student understanding through analysis of others

  3. Potential Issues with Peer Grading Issues: Ways to Address: • Students may be • Make inaccurate unreliable graders graders count less toward final grade • Inexperience in grading • Lack of understanding of • Provide graders with material an incentive to grade • Students may not care accurately about grading accurately

  4. PeerRank • Algorithm developed by Toby Walsh a • Two factors in final grade: • Weighted combination of b d grades from peers • Individual’s own c accuracy in grading others 𝐵 𝑏,𝑑 𝐵 𝑏,𝑒 𝐵 𝑏,𝑏 𝐵 𝑏,𝑐 • Same linear algebra 𝐵 𝑐,𝑑 𝐵 𝑐,𝑒 𝐵 𝑐,𝑏 𝐵 𝑐,𝑐 foundations as Google 𝐵 = 𝐵 𝑑,𝑏 𝐵 𝑑,𝑐 𝐵 𝑑,𝑑 𝐵 𝑑,𝑒 PageRank 𝐵 𝑒,𝑏 𝐵 𝑒,𝑐 𝐵 𝑒,𝑑 𝐵 𝑒,𝑒 • Original application: Reviewing grant proposals

  5. 0 = 1 𝑌 𝑗 𝑛 𝐵 𝑗,𝑘 PeerRank 𝑘 𝑜+1 = 1 − 𝛽 − 𝛾 ∙ 𝑌 𝑗 𝑜 𝑌 𝑗 • Start with “initial seed” 𝑜 ∙ 𝐵 𝑗,𝑘 𝛽 ∙ 𝑌 + grade vector 𝑌 0 𝑘 𝑘 𝑜 𝑌 𝑘 𝑘 𝛾 𝑜 𝑛 ∙ 1 − 𝐵 𝑘,𝑗 − 𝑌 + • Average of grades 𝑘 𝑘 received • PeerRank equation is Fixed Point evaluated iteratively until fixed point is Initial Seed reached • 𝑌 𝑜+1 ≈ 𝑌 𝑜

  6. Problems with PeerRank • Walsh’s Assumption: a b A grader’s accuracy is assumed to be equal c d e to their grade • Unrealistic assumption? 1 1 0 0 0 1 1 0 0 0 • No way of specifying 0 0 1 1 1 0 0 1 1 1 “correctness” 0 0 1 1 1 • May produce incorrect Correct Result: [1,1,0,0,0] results

  7. Problems with PeerRank • Walsh’s Assumption: a b A grader’s accuracy is assumed to be equal c d e to their grade • Unrealistic assumption? 1 1 0 0 0 1 1 0 0 0 • No way of specifying 0 0 1 1 1 0 0 1 1 1 “correctness” 0 0 1 1 1 • May produce incorrect Correct Result: [1,1,0,0,0] results Actual Result: [0,0,1,1,1]

  8. Project Goal Modify and adapt the PeerRank algorithm so that it can better provide accurate peer grading in a classroom setting

  9. Incorporating “Ground Truth” • Recall: There is no way of specifying “correctness” in PeerRank. • In education, there is a notion of “ground truth” in assignments • Right answer vs. wrong answer • Correct proof • Essay with strong argument and no errors • Ground truth is normally determined by instructor

  10. Incorporating “Ground Truth” • Goal: Give the instructor a role in the PeerRank process that influences the accuracy weights of the students Solution: The instructor submits their own assignment with a known grade. Each student grades the instructor’s assignment, and their grading error determines their accuracy Students do not know which assignment is instructor’s Use these accuracies to produce a weighted combination of the peer grades

  11. Incorporating “Ground Truth” • Goal: Give the instructor a role in the PeerRank process that influences the accuracy weights of the students • Solution: • The instructor submits their own assignment for which they know the correct grade • Each student grades the instructor’s assignment, and their grading error determines their accuracy • Students do not know which assignment is instructor’s • Use these accuracies to produce a weighted combination of the peer grades

  12. Our Method vs. PeerRank PeerRank: Our Method: • Accuracy equal to grade • Accuracy determined by • Walsh’s assumption applies accuracy in grading the instructor • Iterative process • Walsh’s assumption no longer • Final grades are fixed point applies • Non-iterative 0 = 1 𝑌 𝑗 𝑛 𝐵 𝑗,𝑘 • Final grades are a weighted 𝑘 average of the peer grades, weighted by the accuracies 𝑜+1 = 1 − 𝛽 − 𝛾 ∙ 𝑌 𝑗 𝑜 𝑌 𝑗 𝐵𝐷𝐷 𝑗 = 1 − |𝐵 𝐽,𝑗 − 𝑌 𝐽 | 𝑜 ∙ 𝐵 𝑗,𝑘 𝛽 ∙ 𝑌 + 𝑘 𝑘 𝑜 𝑌 𝑘 𝑘 1 𝑌 = 𝐵 ∙ 𝐵𝐷𝐷 𝛾 𝑜 𝑛 ∙ 1 − 𝐵 𝑘,𝑗 − 𝑌 + 𝐵𝐷𝐷 1 𝑘 𝑘

  13. Majority vs. Minority Case • Recall: If a group of a b incorrect students outnumber a group of correct students, the c d e wrong grades are produced by 1 1 0 0 0 PeerRank. 1 1 0 0 0 0 0 1 1 1 What if the instructor 0 0 1 1 1 submits a correct 0 0 1 1 1 assignment in our Correct Result: [1,1,0,0,0] system? Actual Result: [0,0,1,1,1]

  14. Majority vs. Minority Case • Recall: If a group of I a b incorrect students outnumber a group of correct students, the c d e wrong grades are produced by 1 1 0 0 0 − PeerRank. 1 1 0 0 0 − 0 0 1 1 1 − • What if the instructor 0 0 1 1 1 − submits a correct 0 0 1 1 1 − 1 1 0 0 0 1 assignment in our Correct Result: [1,1,0,0,0,1] system?

  15. Majority vs. Minority Case • Recall: If a group of I a b incorrect students outnumber a group of correct students, the c d e wrong grades are produced by 1 1 0 0 0 − PeerRank. 1 1 0 0 0 − 0 0 1 1 1 − • What if the instructor 0 0 1 1 1 − submits a correct 0 0 1 1 1 − 1 1 0 0 0 1 assignment in our Correct Result: [1,1,0,0,0,1] system? Accuracies: [1,1,0,0,0,1]

  16. Majority vs. Minority Case • Recall: If a group of I a b incorrect students outnumber a group of correct students, the c d e wrong grades are produced by 1 1 0 0 0 − PeerRank. 1 1 0 0 0 − 0 0 1 1 1 − • What if the instructor 0 0 1 1 1 − submits a correct 0 0 1 1 1 − 1 1 0 0 0 1 assignment in our Correct Result: [1,1,0,0,0,1] system? Accuracies: [1,1,0,0,0,1] Actual Result: [1,1,0,0,0,1]

  17. 0 = 1 𝑌 𝑗 𝑛 𝐵 𝑗,𝑘 Implementation 𝑘 𝛽 𝑜+1 = 1 − 𝛽 − 𝛾 ∙ 𝑌 𝑗 𝑜 + 𝑜 ∙ 𝐵 𝑗,𝑘 𝑌 𝑗 ∙ 𝑌 𝑜 𝑘 𝑌 • Algorithms for PeerRank and 𝑘 𝑘 𝑘 our method implemented in + 𝛾 𝑜 𝑛 ∙ 1 − 𝐵 𝑘,𝑗 − 𝑌 Sage 𝑘 𝑘 • Based on Python def GeneralPeerRank(A, alpha, beta): • Additional math operations, m = A.nrows() Xlist = [0] * m including matrices and for i in range(0, m): vectors sum = 0.0 for j in range(0, m): • Graphing packages sum += A[i,j] X_i = sum / m Xlist[i] = X_i X = vector(Xlist) fixedpoint = False while not fixedpoint: oldX = X X = (1-alpha-beta)*X + \ (alpha/X.norm(1))*(A*X) for i in range(0, m): X[i] += beta - \ (beta/m)*((A.column(i)- \ oldX).norm(1)) difference = X – oldX if abs(difference) < 10**-10: fixedpoint = True return X

  18. Simulating Data • Real grade data is not easily accessible • Data was simulated using statistical models • Ground truth grades drawn from bimodal distribution • Accuracies drawn from normal distributions centered at grader’s grade • Peer grades drawn from uniform distributions using ground truth grade and accuracies

  19. Experiments • Experiments consisted of generating class/grade data and comparing the performance of PeerRank and our modified version against the ground truth grades. • Variables: • Class size • Grade distribution means, standard deviations Correct Grades • Percentage of students in Grades from Our Method each group PeerRank Grades • Accuracy distribution standard deviation

  20. Reducing Connection Between Grade and Accuracy • Recall: The original version of PeerRank assumes that the grader’s grade is equal to their grading accuracy. • Unrealistic assumption? • Our method does assume any connection between grade and accuracy. • How do the two versions compare as we reduce the connection between grade and accuracy? • We can model this reduction by increasing the standard deviation around the graders’ grades when drawing their accuracies.

  21. Reducing Connection Between Grade and Accuracy Standard Deviation = 0.02 Avg. Error Reduction < 0.1% Correct Grades Grades from Our Method PeerRank Grades

  22. Reducing Connection Between Grade and Accuracy Standard Standard Deviation Deviation = 0.02 = 0.10 Avg. Error Avg. Error Reduction Reduction ≈ 0.2% < 0.1% Correct Grades Grades from Our Method PeerRank Grades

  23. Reducing Connection Between Grade and Accuracy Standard Standard Deviation Deviation = 0.02 = 0.10 Avg. Error Avg. Error Reduction Reduction ≈ 0.2% < 0.1% Standard Deviation = 0.50 Avg. Error Reduction ≈ 2.3% Correct Grades Grades from Our Method PeerRank Grades

  24. Reducing Connection Between Grade and Accuracy Standard Standard Deviation Deviation = 0.02 = 0.10 Avg. Error Avg. Error Reduction Reduction ≈ 0.2% < 0.1% Standard Standard Deviation Deviation = 0.50 = 1.0 Avg. Error Avg. Error Reduction Reduction ≈ 2.3% ≈ 4.0% Correct Grades Grades from Our Method PeerRank Grades

Recommend


More recommend