Testing Causal Theories with Learned Proxies

Dean Knox, Christopher Lucas, Wendy K.Tam Cho

    Research output: Contribution to journalReview articlepeer-review

    20 Scopus citations

    Abstract

    Social scientists commonly use computational models to estimate proxies of unobserved concepts, then incorporate these proxies into subsequent tests of their theories. The consequences of this practice, which occurs in over two-Thirds of recent computational work in political science, are underappreciated. Imperfect proxies can reflect noise and contamination from other concepts, producing biased point estimates and standard errors. We demonstrate how analysts can use causal diagrams to articulate theoretical concepts and their relationships to estimated proxies, then apply straightforward rules to assess which conclusions are rigorously supportable. We formalize and extend common heuristics for quot signing the bias quot mdash a technique for reasoning about unobserved confounding mdash to scenarios with imperfect proxies. Using these tools, we demonstrate how, in often-encountered research settings, proxy-based analyses allow for valid tests for the existence and direction of theorized effects. We conclude with best-practice recommendations for the rapidly growing literature using learned proxies to test causal theories.

    Original languageEnglish
    Pages (from-to)419-441
    Number of pages23
    JournalAnnual Review of Political Science
    Volume25
    DOIs
    StatePublished - Feb 1 2022

    Keywords

    • causal inference
    • machine learning
    • measurement
    • proxies
    • supervised learning

    Fingerprint

    Dive into the research topics of 'Testing Causal Theories with Learned Proxies'. Together they form a unique fingerprint.

    Cite this