Identifying and mitigating bias in machine learning models used in population health

  • Funded by Canadian Institutes of Health Research (CIHR)
  • Total publications:0 publications

Grant number: 202010P13

Grant search

Key facts

  • Disease

    Unspecified
  • Start & end year

    2020
    2021
  • Known Financial Commitments (USD)

    $13,825
  • Funder

    Canadian Institutes of Health Research (CIHR)
  • Principal Investigator

    N/A

  • Research Location

    Canada
  • Lead Research Institution

    Unity Health Toronto
  • Research Priority Alignment

    N/A
  • Research Category

    Research to inform ethical issues

  • Research Subcategory

    Research to inform ethical issues in Research

  • Special Interest Tags

    Data Management and Data Sharing

  • Study Type

    Non-Clinical

  • Clinical Trial Details

    N/A

  • Broad Policy Alignment

    Pending

  • Age Group

    Not Applicable

  • Vulnerable Population

    Not applicable

  • Occupations of Interest

    Not applicable

Abstract

Machine learning (ML) is a form of artificial intelligence that has increasingly been used over the last decade in a variety of sectors, as computing power and the availability of data has increased exponentially. ML applies algorithms to vast amounts of data, and through doing so, are programmed to learn over time in an autonomous fashion. ML has begun to be applied to population-level data, both in health services and in public health. However, it is now well recognized that ML models can replicate biases related to numerous factors including that the data used to train models contains biases, the people developing the models bring their biases to their work, and ideas about causation are biased. As ML is applied in population health, including the COVID-19 pandemic, guidelines to identify and mitigate biases are needed. This project consists of two parts that will occur in tandem. Part 1: We will conduct a scoping review of the literature to identify ML models used in specific areas of population health and to examine whether and how biases were identified. Part 2: We will develop guidelines for model developers to identify and reduce bias in ML models used in population health. Our approach will be informed by standards from the Grading of Recommendations Assessment, Development and Evaluation (GRADE), the UK National Institute for Health and Care Excellence (NICE), and guidelines for statistical model development. We will engage a broad set of stakeholders in the guideline development process. Our diverse team includes national and international experts in computer sciences, statistical modeling, epidemiology, ethics, sociology, public health, and the social determinants of health. This project builds on previous work on public perspectives of AI, ML applied to predicting population health outcomes, and the ethics of AI. Our findings will support efforts to ensure AI applications contribute to reducing, rather than exacerbating, health inequities.