alandenadel.bsky.social
@alandenadel.bsky.social
Interestingly, we saw improved zero-shot performance when increasing model size (but still no data scaling) for both scVI and Geneformer
November 7, 2025 at 8:07 PM
The Nicheformer authors observed a similar phenomenon, that when Nicheformer was pre-trained on 1% of their 110M cell dataset performance did not decrease dramatically:
November 7, 2025 at 8:07 PM
And for out-of-distribution perturbation response prediction.
November 7, 2025 at 8:07 PM
We also observed similar results for zero-shot batch integration.
November 7, 2025 at 8:07 PM
The learning saturation points were always 25% or less when evaluating the models on zero-shot classification and were always 10% or less when evaluating the models on fine-tuned classification.
November 7, 2025 at 8:07 PM
To assess the extent to which this plateauing generalized across datasets and tasks, we identified the "learning saturation point" for each model. This is the minimum pre-training dataset size for which a model surpassed 95% of the maximum performance observed.
November 7, 2025 at 8:07 PM
Across all model architectures, model performance at cell type classification (both zero-shot and fine-tuned) plateaued at a small fraction of the total pre-training dataset size, regardless of dataset diversity. When fine-tuning, pre-training has almost no impact on performance.
November 7, 2025 at 8:07 PM
We assessed five model architectures pre-trained to perform as single-cell foundation models (scFMs) in the context of single-cell RNA-seq: PCA, scVI, SSL, Geneformer, and SCimilarity. We pre-trained these models on subsets of the scTab corpus using three downsampling schemes.
November 7, 2025 at 8:07 PM
The learning saturation points were always 25% or less when evaluating the models on zero-shot classification and were always 10% or less when evaluating the models on fine-tuned classification. We also observed similar results for zero-shot batch integration.
December 18, 2024 at 6:48 PM
To assess the extent to which this plateauing generalized across datasets and tasks, we identified the "learning saturation point" for each model. This is the minimum pre-training dataset size for which a model surpassed 95% of the maximum performance observed.
December 18, 2024 at 6:48 PM
Model performance at cell type classification (both zero-shot and fine-tuned) tended to plateau at a small fraction of the total pre-training dataset size on a clonal hematopoiesis evaluation dataset, regardless of pre-training dataset diversity.
December 18, 2024 at 6:48 PM
The three downsampling schemes were: (1) random downsampling (2) cell type re-weighting and (3) geometric sketching. (1) conserves diversity, while (2) and (3) increase diversity (relative to the full corpus). Datasets were generated at 1%, 10%, 25%, 50%, and 75% of the total.
December 18, 2024 at 6:48 PM
Current methods in the field are trained on atlases ranging from 1 to 100 million cells. In our newest preprint, we show that these same approaches tend to plateau in performance with pre-training datasets that are only a fraction of the size.
December 18, 2024 at 6:48 PM