With EssayTagger's core platform in place, it's time to turn our attention to the incredibly rich data that is generated when you grade your essays in our system.
We've already updated the charts quite a bit and have updated this post to reflect the changes!
Even more improvements and two new charts! Post updated again.
You can now download your grading data to Excel!
- "Section snapshot" overall section-wide aggregate performance graph
- "Section details" chart of all students' performance on each rubric element
- "Individual details" in-depth view of a particular student's performance on the assignment
- Statistically-significant outlier identification to help you focus on the students who are furthest from the pack.
All of these data reports are amazingly useful tools for teachers, but I'm particularly excited about the statistical analysis we're able to provide. You don't have to know the first thing about stats, standard deviation, or z-values; we're computing everything for you and flagging the kids that need your attention the most!
You grade, we crunch the numbers. How awesome is that?!
(see the demo video here: http://youtu.be/WZsEoAJEkv0)
"Section snapshot" overall results
This is the new default view; you'll be routed here automatically when you click "exit grading app" when you're done grading. It's the broadest view of the data and includes two charts. The goal is to provide a rough "snapshot" look at how your class section performed as a whole on the essays graded thus far:
The stacked column graph displays how many of your students fell into which quality levels when you evaluated their essays in the grading app.
Put simply: the more green, the better.
The second chart takes the same data but presents it in a slightly different manner:
Now the rubric elements are sorted from best performance to worst performance so you can quickly hone in on the areas that need the most work.
The average rating for each rubric element is reported in the far right column. The percentages within the grid are an easier way to process how many of your students fell into each quality level for each rubric element.
Where did that average score come from?
The system automatically scales your evaluations based on the number of quality levels you specified when creating your rubric.
- 3 quality levels: 1.0 - 3.0
- 4 quality levels: 1.0 - 4.0
- 5 quality levels: 1.0 - 5.0
Note: Common Core-aligned rubrics are always restricted to 5 pre-configured quality levels.
The numeric value is listed under each quality level (e.g. "Proficient" equates to the 4.00 range).
We can then take each of these numeric values and do aggregate calculations like a determining a class-wide average and even more advanced statistical analysis (more on this below).
"Section details" individual results
This view drills down to the per-student level and gives you a color-coded view of each student's performance on each rubric element:
Note: If an element is evaluated more than once in an essay (e.g. "Textual Evidence / Inferences" in the chart above), the evaluations are averaged together to create a single score for the student for that rubric element.
The down arrows indicate scores that are statistically significant outliers; these students are significantly under-performing relative to their peers. This is where our ability to run the statistics does the work for you. Sure, we'll always have students that are struggling, but now you'll know exactly which students stand out as statistical outliers on specific rubric elements.
And the results may be surprising. For example, in the chart above Bart Connor is struggling on "Overall Organization" (indicated in red), but he's actually further behind his classmates on "Textual Evidence / Inferences" and "Transitions/Links" based on the statistical analysis. This doesn't necessarily mean you shouldn't work "Overall Organization" with him, but the outlier analysis reveals a significant weakness in his skill set that you might have otherwise missed.
"Individual details" full breakdown
Clicking on a student's name will bring up a more detailed view of that student's performance on the assignment:
Each of the evaluations you made in the student's essay appear here as checkmarks in the rubric grid. The multiple checkmarks in "Textual Evidence / Inferences" indicate that the instructor made multiple evaluations of the student's evidence in his essay.
Performance relative to peers
The student's numeric value for each rubric element is reported (this is the same as what was displayed in the previous chart). But this time we add the "diff vs section average" column. This shows how the student performed relative to his or her peers.
In the example here, Brian beat the average on "Thesis" and "Develop Counterclaims" but underperformed in the remaining rubric elements.
You can view other students by selecting them from the "student" droplist above the chart.
A note about these scores
These numbers are just a simple way to quantify the quality level decisions you made while evaluating each essay. They are not intended to be used to calculate a grade for each essay; that is left up to the instructor and his or her own judgment (in fact, numeric scores for essays are optional and are enabled/disabled by the Max Point Value setting in the Assignment setup).
We avoid auto-tabulating a final score based on these numbers because we assume that the assignment will weight certain elements more heavily than others (e.g. "Thesis" might count more than "Citation Format"). Because we currently do not support the weighting of one rubric element over the other, any "final" calculation we could perform would be unlikely to produce satisfactory results for the instructor.
We're also wary of removing the instructor's judgment from the final score. The stats may say that a student underperformed on every aspect of the rubric, but it still might be the best paper he's ever written. Instructors should be able to exercise their judgment and reward that student's effort accordingly. Technology does not have all the answers and it never will.
How to access these reports
The grading app will auto-redirect you to the first report when you click "exit grading app." However, you can directly access the reports from the "analytics" tab on the Assignment Details page. Notice that the original v1 data reports are still available as well:
You can also click on the new "data reports" link at the top right of any member area page:
Future enhancement: Student access
I'd like to add a link in the students' graded version of their essays to their "Individual details" chart. I think it's useful to see how you're performing relative to your peers.
I'm used to teaching seniors and I train them to be tough and face facts, but I can also see some teachers worrying about this comparison against the median; it can be quite disheartening to see that every aspect of your essay underperformed relative to your peers.
Obviously there would also have to be security restrictions so that students would not be able to access anyone else's chart.
Future enhancement: Sharable reports
I think there would be a lot of utility in being able to send a link to these reports via email. Send a report to your administrator, to the other members of your teacher team, to your mentor or coordinator. The reports are currently only available to the logged-in instructor, but a future code enhancement could certainly make sharing possible.