Agreement Between Experts and an Untrained Crowd for Identifying Dermoscopic Features Using a Gamified App: Reader Feasibility Study.
Jonathan KentleyJochen WeberKonstantinos LiopyrisRalph P BraunAshfaq A MarghoobElizabeth A QuigleyKelly C NelsonKira McCoolErik P DuhaimeAllan C HalpernVeronica M RotembergPublished in: JMIR medical informatics (2023)
This study confirmed that IRR for different dermoscopic features varied among a group of experts; a similar pattern was observed in a nonexpert crowd. There was good or excellent agreement for each of the 6 superfeatures between the crowd and the experts, highlighting the similar reliability of the crowd for labeling dermoscopic images. This confirms the feasibility and dependability of using crowdsourcing as a scalable solution to annotate large sets of dermoscopic images, with several potential clinical and educational applications, including the development of novel, explainable ML tools.