Inferers who consistently produce good inferences get a high score over time, and therefore get a large weight assigned to them when computing the network inference. But what happens if an inferer who used to give good results suddenly produces a totally wrong result (say in the wrong order of magnitude)? Then the network inference could be very wrong. Eventually, the network would correct by decreasing the inferer’s score, but this takes a couple of epochs.
There are many reasons such a thing could happen. For example, imagine the inferer (which is most likely a program running without human supervision) loses access to some important data source. Or the data source changes its format and suddenly the program is reading the wrong column. And that leads to its results being completely wrong as well.
So it would be nice to have some kind of outlier-resistant network inference which filters out extreme outliers. How could we achieve that? A solution should have the following properties:
- It is relatively conservative, i.e. only filters out data points which are very clearly outliers
- The outlier detection has to react immediately to an outlier. As I wrote above, the network itself can take care of a bad inferer eventually, but this is about the first few epochs before the network reacts.
- The outlier-resistant network inference should not feed back into the network, i.e. is not used for scoring partipants, distributing awards, or anything like that. It’s just meant to be a convenience function for topic consumers.