AI models for medical imaging need labeled data, which can be hard to obtain with the required volume and/or accuracy.
Crowdsourced labels on medical imaging data can help bridge that gap, as suggested by our data showing that crowdsourced classifications and segmentations for B-lines on lung ultrasound videos have comparable or better accuracy than annotations from medical experts with advanced training in lung ultrasound (Duggan 2023) (Jin 2023).
To demonstrate this, we will train deep learning B-line classification and segmentation models using crowdsourced annotations on lung ultrasound videos collected from 483 patients and compare their performance to models trained on manual annotations from experts.
Produce models for classification and segmentation of lung ultrasound videos with comparable or improved performance to models trained on manual annotations from experts.
Collected 330,000 (177,000 at start of PW40) crowdsourced segmentation opinions to form high-quality segmentations of 21,000 (8,500 at the start of PW40) frames within the videos from 483 patients.
In progress:
What are B-lines in lung ultrasound? Here is a picture. The white beams are B-lines, the dark sectors are shadows from ribs.
For crowdsourcing, we collect opinions from 5 experts and combine their opinions, like so. Yellow is the expert consensus:
Then, we collect opinions from crowd using a gamified system very similar to Google’s RECAPTCHA. We take the most reliable opinions and combine them to get a high-quality consensus. Here are the crowd opinions for the same image frame, with the expert consensus in yellow:
For this image, here are the crowd consensus and expert consensus both overlaid:
We will be continuing to crowdsource B-line segmentations throughout Project Week! Here is the visualization of the progress so far (mid-day 2024-01-29):
No response