|
|
| # Stack: In-context learning of single-cell biology |
|
|
| Stack is a large-scale encoder-decoder foundation model trained on 150 million uniformly-preprocessed single cells. It introduces a novel tabular attention architecture that enables both intra- and inter-cellular information flow, setting cell-by-gene matrix chunks as the basic input data unit. Through in-context learning, Stack offers substantial performance improvements in generalizing biological effects and enables generation of unseen cell profiles in novel contexts. |
|
|
| ## Installation |
|
|
| ### Using pip |
| ```bash |
| # Install from PyPI |
| pip install arc-stack |
| |
| # Or install from source for development |
| git clone https://github.com/ArcInstitute/stack.git |
| cd stack |
| pip install -e . |
| ``` |
|
|
| ### Using uv |
| ```bash |
| # Install from PyPI |
| uv pip install arc-stack |
| |
| # Or install from source for development |
| git clone https://github.com/ArcInstitute/stack.git |
| cd stack |
| uv pip install -e . |
| ``` |
|
|
|
|
| ## Quick Start |
|
|
| - Use Stack to embed your single-cell data: [Notebook](notebooks/tutorial-embed.ipynb) |
| - Use Stack to zero-shot predict unseen perturbation/observation profiles: [Notebook](notebooks/tutorial-predict.ipynb) |
|
|
| ### Training Stack from Scratch |
|
|
| ```bash |
| # Once installed, the console entry point becomes available |
| stack-train \ |
| --dataset_configs "/path/to/data:false:gene_symbols" \ |
| --genelist_path "hvg_genes.pkl" \ |
| --save_dir "./checkpoints" \ |
| --sample_size 256 \ |
| --batch_size 32 \ |
| --n_hidden 100 \ |
| --token_dim 16 \ |
| --n_layers 9 \ |
| --max_epochs 10 |
| |
| # Alternatively, invoke the module directly when working from a cloned repo |
| python -m stack.cli.launch_training [args...] |
| ``` |
|
|
| ### Fine-tuning Stack with Frozen Teacher |
|
|
| ```bash |
| stack-finetune \ |
| --checkpoint_path "./checkpoints/pretrained.ckpt" \ |
| --dataset_configs "human:/path/to/data:donor_id:cell_type:false" \ |
| --genelist_path "hvg_genes.pkl" \ |
| --save_dir "./finetuned_checkpoints" \ |
| --sample_size 512 \ |
| --batch_size 8 \ |
| --replacement_ratio 0.75 \ |
| --max_epochs 8 |
| |
| # Or use uv run |
| uv run stack-finetune [args...] |
| |
| # Repository wrapper remains available for local development |
| python -m stack.cli.launch_finetuning [args...] |
| ``` |
|
|
| ### Running Stack with configuration files |
|
|
| Both `launch_training.py` and `launch_finetuning.py` accept a `--config` flag that points to a YAML or JSON file. Any command line |
| arguments omitted after `--config` inherit their values from the file, while flags provided on the command line override the |
| configuration. Example configs mirroring the provided Slurm scripts live under `configs/`: |
|
|
| ```bash |
| # Train with the preset configuration |
| stack-train --config configs/training/bc_large.yaml |
| |
| # Override a single hyperparameter without editing the file |
| stack-train --config configs/training/bc_large.yaml --learning_rate 5e-5 |
| |
| # Fine-tune using a config file |
| stack-finetune --config configs/finetuning/ft_parsecg.yaml |
| |
| # Direct module invocation is still supported if you prefer python -m |
| python -m stack.cli.launch_training --config configs/training/bc_large.yaml |
| ``` |
|
|
| > **Note:** YAML configs require [`pyyaml`](https://pyyaml.org/). Install it with `pip install pyyaml` or use a JSON config file. |
|
|
| ### Extracting Stack Embeddings |
|
|
| ```bash |
| stack-embedding \ |
| --checkpoint "./checkpoints/pretrained.ckpt" \ |
| --adata "data.h5ad" \ |
| --genelist "hvg_genes.pkl" \ |
| --output "embeddings.h5ad" \ |
| --batch-size 32 |
| |
| # Or use uv run |
| uv run stack-embedding \ |
| --checkpoint "./checkpoints/pretrained.ckpt" \ |
| --adata "data.h5ad" \ |
| --genelist "hvg_genes.pkl" \ |
| --output "embeddings.h5ad" \ |
| --batch-size 32 |
| ``` |
|
|
| ### In-Context Generation with Stack |
|
|
| ```bash |
| stack-generation \ |
| --checkpoint "./checkpoints/pretrained.ckpt" \ |
| --base-adata "base_data.h5ad" \ |
| --test-adata "test_data.h5ad" \ |
| --genelist "hvg_genes.pkl" \ |
| --output-dir "./generations" \ |
| --split-column "donor_id" |
| |
| # Or use uv run |
| uv run stack-generation \ |
| --checkpoint "./checkpoints/pretrained.ckpt" \ |
| --base-adata "base_data.h5ad" \ |
| --test-adata "test_data.h5ad" \ |
| --genelist "hvg_genes.pkl" \ |
| --output-dir "./generations" \ |
| --split-column "donor_id" |
| ``` |
|
|
| ## Model Architecture |
|
|
| - **Tabular Attention**: Alternating cell-wise and gene-wise attention layers |
| - **Token Dimension**: Configurable token embedding dimension (default: 16) |
| - **Hidden Dimension**: Gene dimension reduction (default: 100) |
| - **Masking Strategy**: Rectangular masking with variable rates (0.1-0.8) |
|
|
| ## Data Preparation |
|
|
| ### Computing Highly Variable Genes (HVGs) |
| ```python |
| from stack.data.datasets import DatasetConfig, compute_hvg_union |
| |
| configs = [DatasetConfig(path="/data/path", filter_organism=True)] |
| hvg_genes = compute_hvg_union(configs, n_top_genes=1000, output_path="hvg.pkl") |
| ``` |
|
|
| ### Dataset Configuration Format |
| - **Human datasets**: `human:/path:donor_col:cell_type_col[:filter_organism[:gene_col]]` |
| - **Drug datasets**: `drug:/path:condition_col:cell_line_col:control_condition[:filter_organism[:gene_col]]` |
|
|
| ## Key Features |
|
|
| - **In-Context Learning**: Zero-shot generalization to new biological contexts |
| - **Multi-Dataset Training**: Simultaneous training on multiple single-cell datasets |
| - **Frozen Teacher Fine-tuning**: Novel fine-tuning procedure with stable teacher targets |
| - **Efficient Data Loading**: Optimized HDF5 loading with sparse matrix support |
|
|
| > **Note:** `scShiftAttentionModel` remains available as an alias for backward compatibility. |
|
|
| ## Citation |
|
|
| If you use Stack in your research, please cite the Stack [paper](https://www.biorxiv.org/content/10.64898/2026.01.09.698608v1). |
|
|
| ## Licenses |
| Stack code is [licensed](LICENSE) under the Creative Commons Attribution-NonCommercial-ShareAlike 4.0 (CC BY-NC-SA 4.0). |
|
|
| The model weights and output are licensed under the [Arc Research Institute Stack Model Non-Commercial License](MODEL_LICENSE.md) and subject to the [Arc Research Institute Stack Model Acceptable Use Policy](MODEL_ACCEPTABLE_USE_POLICY.md). |
|
|