Skip to main content

Software and AI for Risk of Bias assessment.

Project description

Risk of Bias

AI Enabled Risk of Bias Assessment

Risk of bias tools are systematic frameworks primarily used in systematic reviews to evaluate potential flaws in individual studies that could lead to a systematic deviation from the true effect of an intervention. They aim to identify specific mechanisms through which bias might be introduced into study results, such as problems arising during the study design, conduct, or analysis. This package provides AI and software tools to assist researchers in conducting risk of bias analyses.

Key Features

  • Advanced AI provides robust and independent risk of bias analysis
  • Standard Frameworks such as RoB2 are provided out of the box
  • Batch processing for systematic reviews and meta-analyses
  • Web interface for single manuscript analysis with download options
  • Command line interface for integration into research workflows
  • Latest AI models for evidence extraction and bias assessment
  • RobVis-compatible exports for publication-ready visualizations

Getting Started

Installation

You can install the program using the following pip command:

pip install risk_of_bias[all]

Web Interface

A simple web interface is provided to analyse a single manuscript. To start the web interface run:

risk-of-bias web

Then open http://127.0.0.1:8000 and upload your manuscript. After processing you will see the report along with links to download the JSON and Markdown representations.

Command Line Interface

The package comes with an easy to use command line interface (CLI) tool. The CLI tool is installed along with the python package. The CLI tool provides several handy parameters you can adjust. Complete documentation is available in cli

But to get started, you can analyse a manuscript by simply passing the path to the file:

risk-of-bias analyse /path/to/manuscript.pdf

For systematic reviews, you can analyse entire directories and automatically generate RobVis-compatible CSV summaries:

risk-of-bias analyse /path/to/manuscripts/

Frameworks

The Risk of Bias tool currently only supports the RoB 2 framework. However, it is designed to be extensible, please raise an issue if there's another framework you are interested in. See frameworks and api/frameworks for additional details and context.

Statement on the Use of AI in Research

AI does not replace human judgment in risk of bias assessment. Instead, these tools should be viewed as powerful assistants, complementing human expertise. The established gold standard, often involving two independent human reviewers and a third for adjudication, remains paramount for rigorous assessment. However, AI can augment this process; for instance, an AI could serve as an additional reviewer alongside human experts, providing a systematically derived perspective, with a human still making the final adjudication. Recognizing that both human reviewers and AI systems can have inherent biases, incorporating an AI perspective can offer a different lens through which to evaluate studies. Moreover, in situations where resource constraints make achieving the full gold standard difficult, AI tools can provide valuable support, helping to elevate the overall consistency and thoroughness of bias assessment as the field progresses towards universally ideal practices.

References

Sterne JAC, Savović J, Page MJ, Elbers RG, Blencowe NS, Boutron I, Cates CJ,
Cheng H-Y,  Corbett MS, Eldridge SM, Hernán MA, Hopewell S, Hróbjartsson A,
Junqueira DR, Jüni P, Kirkham JJ, Lasserson T, Li T, McAleenan A, Reeves BC,
Shepperd S, Shrier I, Stewart LA, Tilling K, White IR, Whiting PF, Higgins JPT.
RoB 2: a revised tool for assessing risk of bias in randomised trials. 
BMJ 2019; 366: l4898.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

risk_of_bias-0.7.0.tar.gz (22.2 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

risk_of_bias-0.7.0-py3-none-any.whl (17.3 kB view details)

Uploaded Python 3

File details

Details for the file risk_of_bias-0.7.0.tar.gz.

File metadata

  • Download URL: risk_of_bias-0.7.0.tar.gz
  • Upload date:
  • Size: 22.2 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.12.9

File hashes

Hashes for risk_of_bias-0.7.0.tar.gz
Algorithm Hash digest
SHA256 bafdb1f93d2c3e014e3d3f87f93063500c98d9dbe94be37eb516c4fdd163f036
MD5 9d212c04452d7b9c68ae992d8370a3a8
BLAKE2b-256 bedfb0ece2f481254ee4d97538e4b780252d4b6770e3b9a443b580051cd3e837

See more details on using hashes here.

Provenance

The following attestation bundles were made for risk_of_bias-0.7.0.tar.gz:

Publisher: publish.yml on rob-luke/risk-of-bias

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file risk_of_bias-0.7.0-py3-none-any.whl.

File metadata

  • Download URL: risk_of_bias-0.7.0-py3-none-any.whl
  • Upload date:
  • Size: 17.3 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.12.9

File hashes

Hashes for risk_of_bias-0.7.0-py3-none-any.whl
Algorithm Hash digest
SHA256 ee613d4009563f822e7b0308336a07c53fd977214491bdecc8fa51953008f3da
MD5 87541130acfda50e4eeb326e72264745
BLAKE2b-256 6855240a6fc9cb5c549889f181b42f129591bfb216fb98ffa8f4b4e42407f108

See more details on using hashes here.

Provenance

The following attestation bundles were made for risk_of_bias-0.7.0-py3-none-any.whl:

Publisher: publish.yml on rob-luke/risk-of-bias

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page