Skip to main content

CONCORD: Contrastive Learning for Cross-domain Reconciliation and Discovery

Project description

CONCORD: COntrastive learNing for Cross-dOmain Reconciliation and Discovery

Qin Zhu, Gartner Lab, UCSF

Description

Resolving the intricate structure of the cellular state landscape from single-cell RNA sequencing (scRNAseq) experiments remains an outstanding challenge, compounded by technical noise and systematic discrepancies—often referred to as batch effects—across experimental systems and replicate. To address this, we introduce CONCORD (COntrastive learNing for Cross-dOmain Reconciliation and Discovery), a self-supervised contrastive learning framework designed for robust dimensionality reduction and data integration in single-cell analysis. The core innovation of CONCORD lies in its probabilistic, dataset- and neighborhood-aware sampling strategy, which enhances contrastive learning by simultaneously improving the resolution of cell states and mitigating batch artifacts. Operated in a fully unsupervised manner, CONCORD generates denoised cell encodings that faithfully preserve key biological structures, from fine-grained distinctions among closely related cell states to large-scale topological organizations. The resulting high-resolution cell atlas seamlessly integrates data across experimental batches, technologies, and species. Additionally, CONCORD’s latent space capture biologically meaningful gene programs, enabling the exploration of regulatory mechanisms underlying cell state transitions and subpopulation heterogeneity. We demonstrate the utility of CONCORD on a range of topological structures and biological contexts, underscoring its potential to extract meaningful insights from both existing and future single-cell datasets.

Full Documentation available at https://qinzhu.github.io/Concord_documentation/.


Installation

It is recommended to use conda to create and set up a clean virtual environment for CONCORD.

1. Install PyTorch

You must install the correct version of PyTorch based on your system's CUDA setup. Follow the instructions on the official PyTorch website.

  • For CPU:
    pip install torch torchvision torchaudio
    
  • For CUDA (adjust based on your GPU version):
    pip install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cu118
    

2. Install CONCORD (Stable or Development)

Stable Version (PyPI)

pip install concord-sc

Development Version (GitHub)

pip install git+https://github.com/Gartner-Lab/Concord.git

Optional Installations

(Recommended) Enable Additional Functionalities

For GO enrichment, benchmarking, and R integration, install:

pip install "concord-sc[optional]"

(Recommended) Install FAISS for Accelerated KNN Search

Note: If using Mac, you may need to disable FAISS when running Concord:

cur_ccd = ccd.Concord(adata=adata, input_feature=feature_list, use_faiss=False, device=device)
  • FAISS with GPU:
    pip install faiss-gpu
    
  • FAISS with CPU:
    pip install faiss-cpu
    

(Optional) Integration with VisCello

CONCORD integrates with the R package VisCello, a tool for interactive visualization.
To explore results interactively, visit VisCello GitHub for more details.


Getting Started

Concord integrates seamlessly with anndata objects. Single-cell datasets, such as 10x Genomics outputs, can easily be loaded into an annData object using the Scanpy package. If you're using R and have data in a Seurat object, you can convert it to anndata format by following this tutorial. In this quick-start example, we'll demonstrate CONCORD using the pbmc3k dataset provided by the scanpy package.

Load package and data

# Load required packages
import concord as ccd
import scanpy as sc
import torch
# Load and prepare example data
adata = sc.datasets.pbmc3k_processed()
adata = adata.raw.to_adata()  # Store raw counts in adata.X, by default Concord will run standard total count normalization and log transformation internally, not necessary if you want to use your normalized data in adata.X, if so, specify 'X' in cur_ccd.encode_adata(input_layer_key='X', output_key='Concord')

Run CONCORD:

# Set device to cpu or to gpu (if your torch has been set up correctly to use GPU), for mac you can use either torch.device('mps') or torch.device('cpu')
device = torch.device('cuda:0' if torch.cuda.is_available() else 'cpu')

# (Optional) Select top variably expressed/accessible features for analysis (other methods besides seurat_v3 available)
feature_list = ccd.ul.select_features(adata, n_top_features=5000, flavor='seurat_v3')

# Initialize Concord with an AnnData object, skip input_feature to use all features
cur_ccd = ccd.Concord(adata=adata, input_feature=feature_list, device=device) 

# If integrating data across batch, simply add the domain_key argument to indicate the batch key in adata.obs
# cur_ccd = ccd.Concord(adata=adata, input_feature=feature_list, domain_key='batch', device=device) 

# Encode data, saving the latent embedding in adata.obsm['Concord']
cur_ccd.encode_adata(output_key='Concord')

Visualization:

CONCORD latent embeddings can be directly used for downstream analyses such as visualization with UMAP and t-SNE or constructing k-nearest neighbor (kNN) graphs. Unlike PCA, it is important to utilize the full CONCORD latent embedding in downstream analyses, as each dimension is designed to capture meaningful and complementary aspects of the underlying data structure.

ccd.ul.run_umap(adata, source_key='Concord', result_key='Concord_UMAP', n_components=2, n_neighbors=15, min_dist=0.1, metric='euclidean')

# Plot the UMAP embeddings
color_by = ['n_genes', 'louvain'] # Choose which variables you want to visualize
ccd.pl.plot_embedding(
    adata, basis='Concord_UMAP', color_by=color_by, figsize=(10, 5), dpi=600, ncols=2, font_size=6, point_size=10, legend_loc='on data',
    save_path='Concord_UMAP.png'
)

The latent space produced by CONCORD often capture complex biological structures that may not be fully visualized in 2D projections. We recommend exploring the latent space using a 3D UMAP to more effectively capture and examine the intricacies of the data. For example:

ccd.ul.run_umap(adata, source_key='Concord', result_key='Concord_UMAP_3D', n_components=3, n_neighbors=15, min_dist=0.1, metric='euclidean')

# Plot the 3D UMAP embeddings
col = 'louvain'
fig = ccd.pl.plot_embedding_3d(
    adata, basis='Concord_UMAP_3D', color_by=col, 
    save_path='Concord_UMAP_3D.html',
    point_size=3, opacity=0.8, width=1500, height=1000
)

License

This project is licensed under the MIT License.
See the LICENSE file for details.

Citation

If you use CONCORD in your research, please cite the following preprint:

"Revealing a coherent cell state landscape across single-cell datasets with CONCORD"
bioRxiv, 2025

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

concord_sc-0.9.4.tar.gz (123.9 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

concord_sc-0.9.4-py3-none-any.whl (141.0 kB view details)

Uploaded Python 3

File details

Details for the file concord_sc-0.9.4.tar.gz.

File metadata

  • Download URL: concord_sc-0.9.4.tar.gz
  • Upload date:
  • Size: 123.9 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.12.2

File hashes

Hashes for concord_sc-0.9.4.tar.gz
Algorithm Hash digest
SHA256 47e79c273e289bf0ff2cb248e303320352228e44bdfe260f45ab308d09a6934c
MD5 010e33b0467ef6dd7e1db14215a6fbd9
BLAKE2b-256 9f9029e1b0a7797b75f332fe4883b847795fb78af3640b3257acd83387e84724

See more details on using hashes here.

File details

Details for the file concord_sc-0.9.4-py3-none-any.whl.

File metadata

  • Download URL: concord_sc-0.9.4-py3-none-any.whl
  • Upload date:
  • Size: 141.0 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.12.2

File hashes

Hashes for concord_sc-0.9.4-py3-none-any.whl
Algorithm Hash digest
SHA256 3ca45c4812ac4ea5f95745054c0e6b85fd9b771c4a71bc896d89034716bbdafa
MD5 7877fffcefcaef9f339710452d6e2da0
BLAKE2b-256 9612bfb5abf4ec0594cc3daf5b0b9db10357910a9d0547fd15532c22903e12b0

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page