ov.fm — Foundation Model Module¶
ov.fm provides a unified API for discovering, selecting, validating, running, and interpreting single-cell foundation models. It wraps 17+ models (scGPT, Geneformer, UCE, scFoundation, CellPLM, etc.) behind a consistent AnnData-based interface with automatic data profiling and model selection.
!!! note “When to use ov.fm”
Use `ov.fm` when you want to apply a pre-trained foundation model to your single-cell data without manually setting up each model's preprocessing pipeline. It handles gene ID conversion, compatibility checks, and output standardization for you.
Quick Start¶
import omicverse as ov
# 1. What models are available?
models = ov.fm.list_models(task="embed")
# 2. Profile your data
profile = ov.fm.profile_data("pbmc3k.h5ad")
# 3. Which model fits best?
selection = ov.fm.select_model("pbmc3k.h5ad", task="embed")
print(selection["recommended"]["name"])
# 4. Is the data ready?
check = ov.fm.preprocess_validate("pbmc3k.h5ad", "scgpt", "embed")
# 5. Run the model
result = ov.fm.run(task="embed", model_name="scgpt", adata_path="pbmc3k.h5ad",
output_path="pbmc3k_embedded.h5ad")
# 6. Visualize & evaluate
metrics = ov.fm.interpret_results("pbmc3k_embedded.h5ad", task="embed")
The 6-Step Workflow¶
ov.fm is designed around six composable steps. You can use any step independently or chain them all together.
Discover ──▸ Profile ──▸ Select ──▸ Validate ──▸ Run ──▸ Interpret
Step |
Function |
Purpose |
|---|---|---|
Discover |
|
Browse available models and their capabilities |
Profile |
|
Detect species, gene scheme, modality, and per-model compatibility |
Select |
|
Score and rank models for your data + task |
Validate |
|
Check data compatibility, get auto-fix suggestions |
Run |
|
Execute model inference (embeddings, annotation, integration, etc.) |
Interpret |
|
Compute metrics (silhouette), generate UMAP visualizations |
API Reference¶
ov.fm.list_models¶
ov.fm.list_models(task=None, skill_ready_only=False) -> dict
List available foundation models with optional filtering.
Parameters:
Parameter |
Type |
Default |
Description |
|---|---|---|---|
|
str | None |
|
Filter by task: |
|
bool |
|
Only return models with fully implemented adapters |
Returns: Dictionary with count (int) and models (list of model summaries).
result = ov.fm.list_models(task="embed")
for m in result["models"]:
print(f"{m['name']:15s} status={m['status']:10s} tasks={m['tasks']}")
ov.fm.describe_model¶
ov.fm.describe_model(model_name: str) -> dict
Get the complete specification for a single model, including input/output contracts, hardware requirements, and resource links.
Returns: Dictionary with keys model, input_contract, output_contract, resources.
spec = ov.fm.describe_model("scgpt")
print(spec["input_contract"]["gene_id_scheme"]) # "symbol"
print(spec["output_contract"]["embedding_key"]) # "X_scGPT"
print(spec["output_contract"]["embedding_dim"]) # 512
ov.fm.profile_data¶
ov.fm.profile_data(adata_path: str) -> dict
Analyze an .h5ad file and return a data profile with automatic species/gene-scheme detection and per-model compatibility assessment.
Returns: Dictionary with n_cells, n_genes, species, gene_scheme, modality, has_raw, layers, obs_columns, obsm_keys, batch_columns, celltype_columns, model_compatibility.
profile = ov.fm.profile_data("pbmc3k.h5ad")
print(f"Species: {profile['species']}")
print(f"Gene IDs: {profile['gene_scheme']}")
# Check which models are compatible
for name, compat in profile["model_compatibility"].items():
status = "OK" if compat["compatible"] else "ISSUES"
print(f" {name}: {status}")
ov.fm.select_model¶
ov.fm.select_model(
adata_path: str,
task: str,
prefer_zero_shot: bool = True,
max_vram_gb: int = None,
) -> dict
Score and rank models for a given dataset and task.
Parameters:
Parameter |
Type |
Default |
Description |
|---|---|---|---|
|
str |
— |
Path to |
|
str |
— |
Task type (required) |
|
bool |
|
Prefer models that don’t require fine-tuning |
|
int | None |
|
Maximum VRAM constraint |
Returns: Dictionary with recommended (name + rationale), fallbacks (list), preprocessing_notes, data_profile.
Scoring logic:
Skill-ready adapter: +100 (ready), +50 (partial), 0 (reference)
Zero-shot match: +30
Gene scheme match: +20
CPU fallback available: +10
Low VRAM: +5
result = ov.fm.select_model("pbmc3k.h5ad", task="embed", prefer_zero_shot=True)
print(f"Recommended: {result['recommended']['name']}")
print(f"Rationale: {result['recommended']['rationale']}")
print(f"Fallbacks: {[f['name'] for f in result['fallbacks']]}")
ov.fm.preprocess_validate¶
ov.fm.preprocess_validate(
adata_path: str,
model_name: str,
task: str,
) -> dict
Validate whether data is compatible with a specific model and task. Returns diagnostic messages and auto-fix suggestions.
Returns: Dictionary with status ("ready" | "needs_preprocessing" | "incompatible"), diagnostics, auto_fixes, data_summary.
result = ov.fm.preprocess_validate("pbmc3k.h5ad", "scgpt", "embed")
if result["status"] == "ready":
print("Data is ready for scGPT")
else:
for diag in result["diagnostics"]:
print(f"[{diag['severity']}] {diag['message']}")
for fix in result["auto_fixes"]:
print(f"Suggested fix: {fix['action']}")
if "code" in fix:
print(fix["code"])
ov.fm.run¶
ov.fm.run(
task: str,
model_name: str,
adata_path: str,
output_path: str = None,
batch_key: str = None,
label_key: str = None,
device: str = "auto",
batch_size: int = None,
checkpoint_dir: str = None,
) -> dict
Execute a foundation model on your data.
Parameters:
Parameter |
Type |
Default |
Description |
|---|---|---|---|
|
str |
— |
Task type (required) |
|
str |
— |
Model name (required) |
|
str |
— |
Path to input |
|
str | None |
|
Path for output (defaults to overwriting input) |
|
str | None |
|
|
|
str | None |
|
|
|
str |
|
|
|
int | None |
|
Override model default batch size |
|
str | None |
|
Path to model checkpoint directory |
Returns: Dictionary with output_path, output_keys, n_cells, status on success; error, status on failure.
Execution flow:
Validates data via
preprocess_validate()Attempts conda subprocess execution (isolated environment)
Falls back to in-process adapter if conda is unavailable
Writes results + provenance metadata to output AnnData
result = ov.fm.run(
task="embed",
model_name="scgpt",
adata_path="pbmc3k.h5ad",
output_path="pbmc3k_embedded.h5ad",
device="cuda",
)
if "error" not in result:
print(f"Output keys: {result['output_keys']}")
print(f"Cells processed: {result['n_cells']}")
ov.fm.interpret_results¶
ov.fm.interpret_results(
adata_path: str,
task: str,
output_dir: str = None,
generate_umap: bool = True,
color_by: list = None,
) -> dict
Generate quality metrics and visualizations for model outputs.
Parameters:
Parameter |
Type |
Default |
Description |
|---|---|---|---|
|
str |
— |
Path to |
|
str |
— |
Task that was executed |
|
str | None |
|
Directory for visualization files |
|
bool |
|
Generate UMAP plots |
|
list | None |
|
|
Metrics computed:
Embedding dimensionality and cell count
Silhouette score (if cell type labels and sklearn are available)
Annotation column detection
Provenance metadata from
adata.uns["fm"]
result = ov.fm.interpret_results(
"pbmc3k_embedded.h5ad",
task="embed",
generate_umap=True,
color_by=["louvain"],
)
for key, info in result["metrics"]["embeddings"].items():
print(f"{key}: dim={info['dim']}, silhouette={info.get('silhouette', 'N/A')}")
Supported Tasks¶
Task |
Description |
Example Models |
|---|---|---|
|
Generate cell embeddings for downstream analysis |
scGPT, Geneformer, UCE, CellPLM |
|
Predict cell type labels |
scGPT (fine-tuned), sccello, ChatCell |
|
Batch integration across datasets |
scGPT, Geneformer, UCE |
|
Perturbation response prediction |
scFoundation, Tabula |
|
Spatial transcriptomics analysis |
Nicheformer |
|
Drug response modeling |
scFoundation |
Model Catalog¶
Skill-Ready Models (full adapter)¶
These models have fully implemented adapters and can be executed directly via ov.fm.run().
Model |
Version |
Tasks |
Species |
Gene IDs |
GPU |
Min VRAM |
|---|---|---|---|---|---|---|
scGPT |
whole-human-2024 |
embed, integrate |
human, mouse |
symbol |
Yes |
8 GB |
Geneformer |
v2-106M |
embed, integrate |
human |
ensembl |
No (CPU OK) |
4 GB |
UCE |
4-layer |
embed, integrate |
7 species |
symbol |
Yes |
16 GB |
Partial-Spec Models¶
These models have partial specifications. They can be used for model selection and profiling; execution depends on adapter availability.
Model |
Tasks |
Modalities |
Key Differentiator |
|---|---|---|---|
scFoundation |
embed, integrate |
RNA |
19K gene vocabulary, perturbation pretraining |
scBERT |
embed, integrate |
RNA |
BERT-style masked language modeling |
GeneCompass |
embed, integrate |
RNA |
120M cell pretraining corpus |
CellPLM |
embed, integrate |
RNA |
Cell-centric (not gene-centric), high throughput |
Nicheformer |
embed, integrate, spatial |
RNA, Spatial |
Niche-aware spatial modeling |
scMulan |
embed, integrate |
RNA, ATAC, Protein, Multi-omics |
Native multi-omics |
Tabula |
embed, annotate, integrate, perturb |
RNA |
Federated learning + FlashAttention |
tGPT |
embed, integrate |
RNA |
Autoregressive next-token prediction |
CellFM |
embed, integrate |
RNA |
MLP architecture, 126M cells |
sccello |
embed, integrate, annotate |
RNA |
Zero-shot annotation via cell ontology |
scPRINT |
embed, integrate |
RNA |
Denoising + protein-coding focus |
ATACformer |
embed, integrate |
ATAC |
ATAC-seq native (peak-based) |
scPlantLLM |
embed, integrate |
RNA |
Plant-specific (Arabidopsis, rice, maize) |
LangCell |
embed, integrate |
RNA |
Text+cell alignment, natural language queries |
!!! tip “Model Selection Cheat Sheet”
- **Default (RNA, human):** scGPT
- **Ensembl IDs / CPU-only:** Geneformer
- **Cross-species:** UCE (supports 7 species)
- **Multi-omics (RNA+ATAC+Protein):** scMulan
- **Spatial transcriptomics:** Nicheformer
- **ATAC-seq only:** ATACformer
- **Plant data:** scPlantLLM
- **Large-scale (1M+ cells):** CellPLM
Data Types & Enums¶
from omicverse.fm import TaskType, Modality, GeneIDScheme, SkillReadyStatus
=== “TaskType”
```python
TaskType.EMBED # "embed"
TaskType.ANNOTATE # "annotate"
TaskType.INTEGRATE # "integrate"
TaskType.PERTURB # "perturb"
TaskType.SPATIAL # "spatial"
TaskType.DRUG_RESPONSE # "drug_response"
```
=== “Modality”
```python
Modality.RNA # "RNA"
Modality.ATAC # "ATAC"
Modality.SPATIAL # "Spatial"
Modality.PROTEIN # "Protein"
Modality.MULTIOMICS # "Multi-omics"
```
=== “GeneIDScheme”
```python
GeneIDScheme.SYMBOL # "symbol" — HGNC symbols (e.g., TP53)
GeneIDScheme.ENSEMBL # "ensembl" — Ensembl IDs (e.g., ENSG00000141510)
GeneIDScheme.CUSTOM # "custom" — Model-specific vocabulary
```
=== “SkillReadyStatus”
```python
SkillReadyStatus.READY # Full adapter implemented
SkillReadyStatus.PARTIAL # Partial spec, needs validation
SkillReadyStatus.REFERENCE # Reference docs only
```
Plugin System¶
You can register custom foundation models by writing a plugin.
Entry Point Plugin (pip-installable)¶
In your pyproject.toml:
[project.entry-points."omicverse.fm"]
my_model = "my_package.fm_plugin:register"
Local Plugin (development)¶
Create a file at ~/.omicverse/plugins/fm/my_model.py:
from omicverse.fm import ModelSpec, SkillReadyStatus, TaskType, Modality, GeneIDScheme
from omicverse.fm.adapters import BaseAdapter
MY_SPEC = ModelSpec(
name="my_model",
version="v1.0",
skill_ready=SkillReadyStatus.PARTIAL,
tasks=[TaskType.EMBED],
modalities=[Modality.RNA],
species=["human"],
gene_id_scheme=GeneIDScheme.SYMBOL,
zero_shot_embedding=True,
embedding_dim=256,
)
class MyAdapter(BaseAdapter):
def run(self, task, adata_path, output_path, **kwargs):
... # Your implementation
def _load_model(self, device):
...
def _preprocess(self, adata, task):
...
def _postprocess(self, adata, embeddings, task):
...
def register():
"""Return (spec, adapter_class) tuple."""
return (MY_SPEC, MyAdapter)
!!! note
Plugins cannot override built-in models. If a name conflict occurs, the plugin is skipped with a warning.
Registry API¶
For advanced use, you can query the model registry directly:
from omicverse.fm import get_registry
registry = get_registry()
# Get a specific model's spec
spec = registry.get("scgpt")
print(spec.embedding_dim) # 512
print(spec.supports_task("embed")) # True
# Find models matching criteria
matches = registry.find_models(
task="embed",
species="human",
gene_scheme="symbol",
zero_shot=True,
max_vram_gb=16,
)
for m in matches:
print(m.name, m.version)
Environment Variables¶
Variable |
Description |
|---|---|
|
Base directory for model checkpoints ( |
|
Model-specific checkpoint directory (works for any model name in uppercase) |
|
Disable conda subprocess execution, use in-process adapters only |
Checkpoint resolution order:
checkpoint_dirparameter inov.fm.run()OV_FM_CHECKPOINT_DIR_<MODEL>environment variableOV_FM_CHECKPOINT_DIR/<model_name>/Default cache:
~/.omicverse/models/<model_name>/
Error Handling¶
All functions return error information in the result dictionary rather than raising exceptions:
result = ov.fm.run(task="embed", model_name="scgpt", adata_path="data.h5ad")
if "error" in result:
print(f"Error: {result['error']}")
print(f"Status: {result['status']}") # "not_implemented", "incompatible", etc.
Common error messages:
Error |
Cause |
|---|---|
|
Model name not in registry |
|
Invalid file path |
|
Wrong file format |
|
No models match the task/data constraints |
|
Model is reference-only |
Hands-On Tutorial¶
For a step-by-step walkthrough with real data (PBMC 3K + scGPT), see the Foundation Model Tutorial Notebook.