I’d like to see a more explicit scoring mechanism. Right now I’ve completed 25 tasks, and I only have a vague idea of what the scoring system is. It seems to be correlated with the volume of neuron that you’ve added, but sometimes not. Just something that says, “you got X points for A, Y points for B, for a total of P points.” would make me feel that the point system isn’t random – that is has a point, if you will :)
(Disclaimer: I’m not a member of the team, just a guy with some thoughts on the matter!)
I’d second aawood’s point about the score. In other posts here, it has been pointed out that the actual answer is not known a-priori, therefore there is no real way other than calculating score by comparing a cohort of possible answers to look for a consensus.
I think you’re right – the goal is to get an independent judgement from the player, and if immediate scoring were provided on a task that others have done before, then you could lazily get points by just clicking until you maximize your score, which doesn’t help anyone.
Actually, you make a very good point about messages regarding previous tasks and points. I’m not sure if they correct scores after they have a sample of answers, but it would be a good way to go.
Yes, normalization is a good idea so that you don’t get lots of points just because you happen to have been dealt a large volume. Also accounting for zero new volume in cases where the neural network got it right and there’s nothing to do.
Great discussion of scoring. You guys are now officially exactly where we are. We want the scored to be fun and engaging which means relatively instant feedback. But we don’t want it to negatively effect the results that we get from people, which means that it can’t be per-click. We don’t know the right answer a-priori, so we’ve got to compare different people’s results with each other. We don’t want to give people 0 points for tasks, but we don’t want to give people points for doing nothing and/or doing the wrong thing. We know better how a person did after a number of other people have done the task, but we don’t want people to wait for their scores. We also don’t want too much overlap for each task, because once we’ve converged on the truth for a particular piece, anyone else working on it is just wasted effort.
- Retroactive Scoring - Once a tasks is completed and closed, we hand out points to everyone
- Retroactive Bonuses - Sort of a hybrid. Completion points for doing a task and bonus points for agreement later
- Team Based Scoring - Teams work together to reconstruct neurons on their own. Teams that do more neurons, or more accurately get more points
- Two Types of Task - Sometimes you will be asked to segment neurons, sometimes you will be asked to synthesize the results of other people’s segmentations. These points would be handed out retroactively too. Huge points if you found a branch that everyone else missed.
- Retroactive scoring.I do think that this is worth trying. Sure, it might put off people who are only here to get ‘points’ by any means, but the more serious players will understand the objective. This would function, in a sense rather like the election nights we have in the UK & US, it would create a sense of anticipation.2. Hybrid Scoring.This could be attempted by correcting scores after the close of a task. Advantage - it gives people a score to work with; Disadvantage - be prepared for very petty people questioning where their points have gone if there is a large discrepancy before and after3 & 4 Team Scoring & segmentation.This is similar to an idea I put in another thread about competitions in which people or teams would construct a neural trace that would be compared against an AI attempt & the consensus. I would not call it a scoring scheme, rather an alternative way of playing the game. I definitely think that something like this should be done.This seems to come down to two main questions :1. Retroactive or running score?2. What should the final score be? One number (accuracy?) or multiple measures (volume of tasks done, and accuracy).Thoughts?
How about this.
(Continued)
You should also be able to click on your score and find out why you got that score for a given task. Player 2’s explanation, for example, would read:
- Rewards players for trailblazing.
- Rewards players for agreeing with each other and postfacto with the synthesis.
- Partially rewards players for agreeing with each other partially and postfacto partially with the synthesis.
- Rewards players even if there is nothing to do.
- Rewards players for being better than the AI.
- Does not reward players who blindly click.
- Is fair, but not 100% fair (people respond better to intermittent rewards than consistent rewards, go figure)
Sidestepping the maths for the moment, I like the idea of AI generated feedback on the scores. It isn’t quite as simple as giving people a number, but with retroactively corrected scores, it would help alleviate the ‘Where’s my score gone, waaa waaa waaa’ response.
I know this hasn’t been central to the discussion, but I think the one main core way the current system is “non-gamified” is with trailblazers getting no points. This can’t be done with any accuracy as it stands without granting them points after the fact. I’d like to suggest a relatively simple way to solve this in the current format, without unbalancing scores, or delaying feedback; when a trailblazer finishes a task, give them a “TB token”. Players can spend a token when being given a task, with it having the effect of doubling their score once they’re done.