This is Karma. Karma is not a machine learning classifier πβπ¦Ί
Karma is a real dog trained to detect drugs. However, he would fail the simplest tests we apply in ML...
Let me take you through this story from the eyes of an ML engineer.
https://t.co/WAXRUlTvSI
Thread π§΅
Story TLDR π
The story is about police dogs trained to sniff drugs. The problem is that the dogs often signal drugs even if there are none. Then innocent people land in jail for days.
The cops even joke about the βprobable cause on four legsβ.
Let's see why is that π
1. Sampling Bias π€
Drugs were found in 64% of the cars Karma identified, which was praised by the police as very good. In the end, most people don't carry drugs in their cars, so 64% seems solid.
There was a sampling problem though... π
The cars were not sampled at random! The police only did the sniff test if there was a serious suspicion that something is wrong.
The chance there are drugs in the car is much higher in this case!
2. Evaluation Metrics π
The police referred to a 2014 study from Poland measuring the efficacy of sniffer dogs. The problem was that every test actually contained drugs!
This means there was no chance to measure false positives from the dogs! Only recall, not precision π€¦ββοΈ