AI competitions don ’t produce useful models

By LUKE OAKDEN-RAYNER A huge new CT brain dataset was released the other day, with the goal of training models to detect intracranial haemorrhage. So far, it looks pretty good, although I haven’t dug into it in detail yet (and the devil is often in the detail). The dataset has been released for a competition, which obviously lead to the usual friendly rivalry on Twitter: Of course, this lead to cynicism from the usual suspects as well. And the conversation continued from there, with thoughts ranging from “but since there is a hold out test set, how can you overfit?” to “the proposed solutions are never intended to be applied directly” (the latter from a previous competition winner). As the discussion progressed, I realised that while we “all know” that competition results are more than a bit dubious in a clinical sense, I’ve never really seen a compelling explanation for why this is so. Hopefully that is what this post is, an explanation for why competitions are not really about building useful AI systems. DISCLAIMER: I originally wrote this post expecting it to be read by my usual readers, who know my general positions on a range of issues. Instead, it was spread widely on Twitter and HackerNews, and it is pretty clear that I didn’t provide enough context for a number of statements made. I am going to write a follow-up to clarify several things, but as a quick response to several common criticisms: I don’t...
Source: The Health Care Blog - Category: Consumer Health News Authors: Tags: Health Tech AI Luke Oakden-Rayner Source Type: blogs