poss_dataset_ids = dataset_info
.map(d => d.dataset_id)
.filter(d => results.map(r => r.dataset_id).includes(d))
poss_method_ids = method_info
.map(d => d.method_id)
.filter(d => results.map(r => r.method_id).includes(d))
poss_metric_ids = metric_info
.map(d => d.metric_id)
.filter(d => results.map(r => Object.keys(r.scaled_scores)).flat().includes(d))
Batch integration feature
Removing batch effects while preserving biological variation (feature output)
3 datasets · 18 methods · 7 control methods · 11 metrics
Task info Method info Metric info Dataset info Results
This is a sub-task of the overall batch integration task. Batch (or data) integration integrates datasets across batches that arise from various biological and technical sources. Methods that integrate batches typically have three different types of output: a corrected feature matrix, a joint embedding across batches, and/or an integrated cell-cell similarity graph (e.g., a kNN graph). This sub-task focuses on all methods that can output feature matrices. Other sub-tasks for batch integration can be found for:
- graphs, and
- embeddings
This sub-task was taken from a benchmarking study of data integration methods.
Summary
Display settings
Filter datasets
Filter methods
Filter metrics
Results
Results table of the scores per method, dataset and metric (after scaling). Use the filters to make a custom subselection of methods and datasets. The “Overall mean” dataset is the mean value across all datasets.
Dataset info
Show
Immune (by batch)
Human immune cells from peripheral blood and bone marrow taken from 5 datasets comprising 10 batches across technologies (10X, Smart-seq2) (Luecken et al. 2021).
Lung (Viera Braga et al.)
Human lung scRNA-seq data from 3 datasets with 32,472 cells. From Vieira Braga et al. Technologies: 10X and Drop-seq (Luecken et al. 2021).
Pancreas (by batch)
Human pancreatic islet scRNA-seq data from 6 datasets across technologies (CEL-seq, CEL-seq2, Smart-seq2, inDrop, Fluidigm C1, and SMARTER-seq) (Luecken et al. 2021).
Method info
Show
Combat (full/scaled)
ComBat uses an Empirical Bayes (EB) approach to correct for batch effects. It estimates batch-specific parameters by pooling information across genes in each batch and shrinks the estimates towards the overall mean of the batch effect estimates across all genes. These parameters are then used to adjust the data for batch effects, leading to more accurate and reproducible results (Johnson, Li, and Rabinovic 2006). Links: Docs.
Combat (full/unscaled)
ComBat uses an Empirical Bayes (EB) approach to correct for batch effects. It estimates batch-specific parameters by pooling information across genes in each batch and shrinks the estimates towards the overall mean of the batch effect estimates across all genes. These parameters are then used to adjust the data for batch effects, leading to more accurate and reproducible results (Johnson, Li, and Rabinovic 2006). Links: Docs.
Combat (hvg/scaled)
ComBat uses an Empirical Bayes (EB) approach to correct for batch effects. It estimates batch-specific parameters by pooling information across genes in each batch and shrinks the estimates towards the overall mean of the batch effect estimates across all genes. These parameters are then used to adjust the data for batch effects, leading to more accurate and reproducible results (Johnson, Li, and Rabinovic 2006). Links: Docs.
Combat (hvg/unscaled)
ComBat uses an Empirical Bayes (EB) approach to correct for batch effects. It estimates batch-specific parameters by pooling information across genes in each batch and shrinks the estimates towards the overall mean of the batch effect estimates across all genes. These parameters are then used to adjust the data for batch effects, leading to more accurate and reproducible results (Johnson, Li, and Rabinovic 2006). Links: Docs.
FastMNN feature (full/scaled)
fastMNN performs a multi-sample PCA to reduce dimensionality, identifying MNN paris in the low-dimensional space, and then correcting the target batch towards the reference using locally weighted correction vectors. The corrected target batch is then merged with the reference. The process is repeated with the next target batch except for the PCA step (Lun 2019). Links: Docs.
FastMNN feature (full/unscaled)
fastMNN performs a multi-sample PCA to reduce dimensionality, identifying MNN paris in the low-dimensional space, and then correcting the target batch towards the reference using locally weighted correction vectors. The corrected target batch is then merged with the reference. The process is repeated with the next target batch except for the PCA step (Lun 2019). Links: Docs.
FastMNN feature (hvg/scaled)
fastMNN performs a multi-sample PCA to reduce dimensionality, identifying MNN paris in the low-dimensional space, and then correcting the target batch towards the reference using locally weighted correction vectors. The corrected target batch is then merged with the reference. The process is repeated with the next target batch except for the PCA step (Lun 2019). Links: Docs.
FastMNN feature (hvg/unscaled)
fastMNN performs a multi-sample PCA to reduce dimensionality, identifying MNN paris in the low-dimensional space, and then correcting the target batch towards the reference using locally weighted correction vectors. The corrected target batch is then merged with the reference. The process is repeated with the next target batch except for the PCA step (Lun 2019). Links: Docs.
MNN (full/scaled)
MNN first detects mutual nearest neighbours in two of the batches and infers a projection of the second onto the first batch. After that, additional batches are added iteratively (Haghverdi et al. 2018). Links: Docs.
MNN (full/unscaled)
MNN first detects mutual nearest neighbours in two of the batches and infers a projection of the second onto the first batch. After that, additional batches are added iteratively (Haghverdi et al. 2018). Links: Docs.
MNN (hvg/scaled)
MNN first detects mutual nearest neighbours in two of the batches and infers a projection of the second onto the first batch. After that, additional batches are added iteratively (Haghverdi et al. 2018). Links: Docs.
MNN (hvg/unscaled)
MNN first detects mutual nearest neighbours in two of the batches and infers a projection of the second onto the first batch. After that, additional batches are added iteratively (Haghverdi et al. 2018). Links: Docs.
SCALEX (full)
SCALEX is a method for integrating heterogeneous single-cell data online using a VAE framework. Its generalised encoder disentangles batch-related components from batch-invariant biological components, which are then projected into a common cell-embedding space (Xiong et al. 2022). Links: Docs.
SCALEX (hvg)
SCALEX is a method for integrating heterogeneous single-cell data online using a VAE framework. Its generalised encoder disentangles batch-related components from batch-invariant biological components, which are then projected into a common cell-embedding space (Xiong et al. 2022). Links: Docs.
Scanorama gene output (full/scaled)
Scanorama is an extension of the MNN method. Other then MNN, it finds mutual nearest neighbours over all batches and embeds observations into a joint hyperplane (Hie, Bryson, and Berger 2019). Links: Docs.
Scanorama gene output (full/unscaled)
Scanorama is an extension of the MNN method. Other then MNN, it finds mutual nearest neighbours over all batches and embeds observations into a joint hyperplane (Hie, Bryson, and Berger 2019). Links: Docs.
Scanorama gene output (hvg/scaled)
Scanorama is an extension of the MNN method. Other then MNN, it finds mutual nearest neighbours over all batches and embeds observations into a joint hyperplane (Hie, Bryson, and Berger 2019). Links: Docs.
Scanorama gene output (hvg/unscaled)
Scanorama is an extension of the MNN method. Other then MNN, it finds mutual nearest neighbours over all batches and embeds observations into a joint hyperplane (Hie, Bryson, and Berger 2019). Links: Docs.
Control method info
Show
Random Integration by Batch
Feature values, embedding coordinates, and graph connectivity are all randomly permuted within each batch label
Random Embedding by Celltype
Cells are embedded as a one-hot encoding of celltype labels
Random Graph by Celltype
Cells are embedded as a one-hot encoding of celltype labels. A graph is then built on this embedding
Random Integration by Celltype
Feature values, embedding coordinates, and graph connectivity are all randomly permuted within each celltype label
No Integration
Cells are embedded by PCA on the unintegrated data. A graph is built on this PCA embedding
No Integration by Batch
Cells are embedded by computing PCA independently on each batch
Random Integration
Feature values, embedding coordinates, and graph connectivity are all randomly permuted
Metric info
Show
ARI
ARI (Adjusted Rand Index) compares the overlap of two clusterings. It considers both correct clustering overlaps while also counting correct disagreements between two clustering (Luecken et al. 2021).
Cell Cycle Score
The cell-cycle conservation score evaluates how well the cell-cycle effect can be captured before and after integration (Luecken et al. 2021).
Graph connectivity
The graph connectivity metric assesses whether the kNN graph representation, G, of the integrated data connects all cells with the same cell identity label (Luecken et al. 2021).
HVG conservation
This metric computes the average percentage of overlapping highly variable genes per batch before and after integration (Luecken et al. 2021).
Isolated label F1
Isolated cell labels are identified as the labels present in the least number of batches in the integration task. The score evaluates how well these isolated labels separate from other cell identities based on clustering (Luecken et al. 2021).
Isolated label Silhouette
This score evaluates the compactness for the label(s) that is(are) shared by fewest batches. It indicates how well rare cell types can be preserved after integration (Luecken et al. 2021).
kBET
kBET determines whether the label composition of a k nearest neighborhood of a cell is similar to the expected (global) label composition. The test is repeated for a random subset of cells, and the results are summarized as a rejection rate over all tested neighborhoods (Büttner et al. 2018).
NMI
NMI compares the overlap of two clusterings. We used NMI to compare the cell-type labels with Louvain clusters computed on the integrated dataset (Luecken et al. 2021).
PC Regression
This compares the explained variance by batch before and after integration. It returns a score between 0 and 1 (scaled=True) with 0 if the variance contribution hasn’t changed. The larger the score, the more different the variance contributions are before and after integration (Luecken et al. 2021).
Silhouette
The absolute silhouette with is computed on cell identity labels, measuring their compactness (Luecken et al. 2021).
Batch ASW
The absolute silhouette width is computed over batch labels per cell. As 0 then indicates that batches are well mixed and any deviation from 0 indicates a batch effect, we use the 1-abs(ASW) to map the score to the scale [0;1] (Luecken et al. 2021).
Quality control results
Show
✓ All checks succeeded!
Normalisation visualisation
Show
Authors
Büttner, Maren, Zhichao Miao, F. Alexander Wolf, Sarah A. Teichmann, and Fabian J. Theis. 2018. “A Test Metric for Assessing Single-Cell RNA-Seq Batch Correction.” Nature Methods 16 (1): 43–49. https://doi.org/10.1038/s41592-018-0254-1.
Haghverdi, Laleh, Aaron T L Lun, Michael D Morgan, and John C Marioni. 2018. “Batch Effects in Single-Cell RNA-Sequencing Data Are Corrected by Matching Mutual Nearest Neighbors.” Nature Biotechnology 36 (5): 421–27. https://doi.org/10.1038/nbt.4091.
Hie, Brian, Bryan Bryson, and Bonnie Berger. 2019. “Efficient Integration of Heterogeneous Single-Cell Transcriptomes Using Scanorama.” Nature Biotechnology 37 (6): 685–91. https://doi.org/10.1038/s41587-019-0113-3.
Johnson, W. Evan, Cheng Li, and Ariel Rabinovic. 2006. “Adjusting Batch Effects in Microarray Expression Data Using Empirical Bayes Methods.” Biostatistics 8 (1): 118–27. https://doi.org/10.1093/biostatistics/kxj037.
Luecken, Malte D., M. Büttner, K. Chaichoompu, A. Danese, M. Interlandi, M. F. Mueller, D. C. Strobl, et al. 2021. “Benchmarking Atlas-Level Data Integration in Single-Cell Genomics.” Nature Methods 19 (1): 41–50. https://doi.org/10.1038/s41592-021-01336-8.
Lun, Aaron. 2019. “A Description of the Theory Behind the fastMNN Algorithm.” https://marionilab.github.io/FurtherMNN2018/theory/description.html.
Xiong, Lei, Kang Tian, Yuzhe Li, Weixi Ning, Xin Gao, and Qiangfeng Cliff Zhang. 2022. “Online Single-Cell Data Integration Through Projecting Heterogeneous Datasets into a Common Cell-Embedding Space.” Nature Communications 13 (1). https://doi.org/10.1038/s41467-022-33758-z.