Unknown

Dataset Information

0

An exploration of crowdsourcing citation screening for systematic reviews.


ABSTRACT: Systematic reviews are increasingly used to inform health care decisions, but are expensive to produce. We explore the use of crowdsourcing (distributing tasks to untrained workers via the web) to reduce the cost of screening citations. We used Amazon Mechanical Turk as our platform and 4 previously conducted systematic reviews as examples. For each citation, workers answered 4 or 5 questions that were equivalent to the eligibility criteria. We aggregated responses from multiple workers into an overall decision to include or exclude the citation using 1 of 9 algorithms and compared the performance of these algorithms to the corresponding decisions of trained experts. The most inclusive algorithm (designating a citation as relevant if any worker did) identified 95% to 99% of the citations that were ultimately included in the reviews while excluding 68% to 82% of irrelevant citations. Other algorithms increased the fraction of irrelevant articles excluded at some cost to the inclusion of relevant studies. Crowdworkers completed screening in 4 to 17 days, costing $460 to $2220, a cost reduction of up to 88% compared to trained experts. Crowdsourcing may represent a useful approach to reducing the cost of identifying literature for systematic reviews.

SUBMITTER: Mortensen ML 

PROVIDER: S-EPMC5589498 | biostudies-literature | 2017 Sep

REPOSITORIES: biostudies-literature

altmetric image

Publications

An exploration of crowdsourcing citation screening for systematic reviews.

Mortensen Michael L ML   Adam Gaelen P GP   Trikalinos Thomas A TA   Kraska Tim T   Wallace Byron C BC  

Research synthesis methods 20170704 3


Systematic reviews are increasingly used to inform health care decisions, but are expensive to produce. We explore the use of crowdsourcing (distributing tasks to untrained workers via the web) to reduce the cost of screening citations. We used Amazon Mechanical Turk as our platform and 4 previously conducted systematic reviews as examples. For each citation, workers answered 4 or 5 questions that were equivalent to the eligibility criteria. We aggregated responses from multiple workers into an  ...[more]

Similar Datasets

| S-EPMC6658317 | biostudies-literature
| S-EPMC8077753 | biostudies-literature
| S-EPMC8264476 | biostudies-literature
| S-EPMC3847500 | biostudies-literature
| S-EPMC2824679 | biostudies-literature
| S-EPMC10976661 | biostudies-literature
| S-EPMC3830982 | biostudies-other
| S-EPMC6599339 | biostudies-literature
| S-EPMC7268596 | biostudies-literature
| S-EPMC6959565 | biostudies-literature