Using Tassomai data to support centre-assessed grades
As teachers around the country face the challenge of grading their own students, Tassomai founder, Murray Morrison, shares some vital tips on how Tassomai data could be used to support decision-making.
It’s an unpleasant responsibility to determine student outcomes this term, when we are more used to arming them with all the knowledge and skills they need to go out and get those grades for themselves. As was discussed at the PiXL virtual conference last week, it’s vital also that the grades given are fair and statistically realistic; if the moderation process deems a school's grades too high, all students, including the most able, may find their grades pulled down.
For those schools who have used Tassomai over the last year or more, the positive news is that you will have accreted a vast amount of data to help finalise grade decisions. The danger is that this data, misinterpreted, could give the wrong outcomes. Having consulted with a number of HoDs over the last couple of weeks, there are a few simple tips that should help get the most of your data, and avoid any false positives or negatives.
Tip 1 - On Tassomai, ‘Accuracy’ can be misleading
The most important factor to bear in mind is that Tassomai is highly adaptive as a program, so continually adjusts the difficulty and spacing coefficients of students’ content in order to give each the right “goldilocks zone” level of challenge or difficulty. Therefore, one student of lower ability may have foundational-level content frequently repeated while their counterpart of higher ability is seeing far more challenging material with greater spacing. Long story short, both students may have similar levels of accuracy, but mean very different things. Consequently, we’d advise strongly against using student accuracy as evidence of potential grade outcome.
Tip 2 - Judge course progress judiciously
By a similar token, different use cases (such as implementation strategies or student engagement) can result in different rates of course completion. We have an enormous amount of evidence from various randomised trials that show that students with a high amount of course coverage (over 25%) will tend to receive grades above the median for their cohort. It’s extremely rare for a student with progress greater than 50% to receive below a 4-5 grade, and likewise extremely likely that they will match or exceed their predicted grade.
The converse is not true: low progress on Tassomai is not an indicator of lower ability or lower predicted outcomes - it could simply indicate that that Tassomai was not pushed as a homework in their class, or that they deliberately ‘spammed’ the program by constantly repeating a single topic quiz rather than following the algorithm’s balanced program.
As a rule of thumb, regardless of prior attainment, students with a course progress of 25% are more likely to meet or exceed targets; below 25% doesn’t tell you very much.
Tip 3 - The Understanding Grid is the best way to support your rankings
Rather than using the stark, raw numbers you find in the dashboard around ‘usage’, we’ve found that the Understanding Grid (“the dots”) is far more informative. It also has compelling evidence from research to support it.
The Understanding Grid can be viewed class by class, and also across an entire year group, so makes the task of ranking a cohort far easier.
Rather than relying solely on accuracy or solely on progress, the dots combine the two in a weighted statistic. If you use the ranking arrow at the top of a column - for example, rank by Biology - you’ll notice larger green dots rank higher than smaller blue ones. This is because the student with a larger dot will have ‘unlocked’ the tougher content. Their accuracy may be a touch lower accordingly, but the expectation is that they will have better prepared themselves to succeed in the exam.
We recommend that you rank students by each subject using the little arrows at the top of the column - this then sorts by a weighted measure of their performance across the whole course - and do so class by class as well as by cohort since those doing foundation courses may suffer in the comparison against those doing higher tier.
Research and Evidence
In 2018-2019, we undertook a trial with 500 students across 12 mock papers in November and February (some 6,000 independently assessed scripts) and correlated their marks in each question with the relevant dots on their understanding grids at the time of assessment; the research, presented at the ASE in Jan 2020, showed a very high correlation between our ranking and their attainment in the examination - enough to give us great confidence in the validity of the results.
There were, however, some instructive outliers which are of note to you: those students who appear on the very bottom of your ranking will be those with very small or non-existent dots. They, again, are like the low-progress exceptions: they didn’t do enough Tassomai to show a true level of ability. Therefore we’d recommend against inferring a lower ranking simply from their lack of Tassomai data.
If you’d like to find out more about this piece of research, the paper is pre-publication, but I’m happy to show you the data and discuss the findings.
Use Tassodata to provide nuance in your rankings, and to provide supportive evidence
One of our Tassomai Ambassadors, conducting the process in his school gave the following advice:
Tassomai performance data could be used with caution as an indicator to justify a slightly higher ranking place within a grade profile, e.g. 4-3 , 4-4. This would give them greater security in grade, as standardisation is most likely to influence the lower rank end with 4-3, 4-4, being most contentious. 4-4 is significant within Gov stats follow up.
While we’d strongly advise against using Tassomai data to issue the grade (as if you were going to do that!) it does provide a superb body of evidence for you to give some hard-working students a grade that shows an improvement on the last sample-point - mock or prediction.
Anyone reading this will have a few students in mind who they know were putting in the hard yards to get a grade that exceeds what they achieved in the PPE; if Tassomai data shows that they were practising, improving, and getting ready for the exam in the right way, then I hope this can be the evidence you need to give them what they deserve.
Again, our Tassomai Ambassador:
Tassomai is also significant in the mocks with significant individual improvements attributed to it. Also with no test data, Tassomai is central in our attitude to learning assessments across the school. Our data is tighter and sharper as a result of using it. For example a pupil in November mock on 3-3 and poor individual effort on Tassomai was unlikely to progress to 4-3 and thus that was our progress grade.
Here to help
I am happy to look through your data with you over screen share at a time that suits you, simply contact us to arrange a convenient time.
I realise you have a million things to think about, but I’ve been looking at these data screens for years…! If I can be of any assistance looking through what we have, I’d be only too happy to help.