Ir para o conteúdo principal

Escrever uma avaliação PREreview

Predicting causal citations without full text

Publicado
Servidor
bioRxiv
DOI
10.1101/2022.07.05.498860

Insights from biomedical citation networks can be used to identify promising avenues for accelerating research and its downstream bench-to-bedside translation. Citation analysis generally assumes that each citation documents causal knowledge transfer that informed the conception, design, or execution of the main experiments. Citations may exist for other reasons. In this paper we identify a subset of citations that are unlikely to represent causal knowledge flow. Using a large, comprehensive feature set of open access data, we train a predictive model to identify such citations. The model relies only on the title, abstract, and reference set and not the full-text or future citations patterns, making it suitable for publications as soon as they are released, or those behind a paywall (the vast majority). We find that the model identifies, with high prediction scores, citations that were likely added during the peer review process, and conversely identifies with low prediction scores citations that are known to represent causal knowledge transfer. Using the model, we find that federally funded biomedical research publications represent 30% of the estimated causal knowledge transfer from basic studies to clinical research, even though these comprise only 10% of the literature, a three-fold overrepresentation in this important type of knowledge transfer. This finding underscores the importance of federal funding as a policy lever to improve human health.

Significance statement

Citation networks document knowledge flow across the literature, and insights from these networks are increasingly used to form science policy decisions. However, many citations are known to be not causally related to the inception, design, and execution of the citing study. This adds noise to the insights derived from these networks. Here, we show that it is possible to train a machine learning model to identify such citations, and that the model learns to identify known causal citations as well. We use this model to show that government funding drives a disproportionate amount of causal knowledge transfer from basic to clinical research. This result highlights a straightforward policy lever for accelerating improvements to human health: federal funding.

Você pode escrever uma avaliação PREreview de Predicting causal citations without full text. Uma avaliação PREreview é uma avaliação de um preprint e pode variar de algumas frases a um parecer extenso, semelhante a um parecer de revisão por pares realizado por periódicos.

Antes de começar

Vamos pedir que você faça login com seu ORCID iD. Se você não tiver um iD, pode criar um.

O que é um ORCID iD?

Um ORCID iD é um identificador único que diferencia você de outras pessoas com o mesmo nome ou nome semelhante.

Começar agora