Look into your ML black box with LIME
While listening to the data sceptic podcast today I learned about a new python framework called LIME (local interpretable model-agnostic explanations). It tackles a common problem in the field of machine learning: the black box.One thing that always raises some eyebrows when I talk to people without a data science background is the fact that it is often unexplainable how an algorithm arrives at its output.
Why would we invest in/work with/trust in machines we do not understand?
The developer of LIME actually has some great examples where even the data scientist can profit immensly from LIME. His main argument is that without looking into the black box it is impossible to use our own powerful tool: human intuition. For example, if one classifies newsgroup posts and LIME tells you that it classifies by looking at stop words you instantly know something is wrong and you can go back to tweaking your model.
Another more frightening example is a model that returns a probability that indicates if a patient is going to survive if he doesn't get treated immediately. Image that this algorithm decides on the probability by looking at the name (or some other irrelevant feature).