Concatenate datasets to a single array store¶
In the previous notebooks, we’ve seen how to incrementally create a collection of scRNA-seq datasets and train models on it.
Sometimes we want to concatenate all datasets into one big array to speed up ad-hoc queries for slices for arbitrary metadata (see this blog post). This is what CELLxGENE does to create Census: a number of .h5ad
files are concatenated to give rise to a single tiledbsoma
array store (CELLxGENE: scRNA-seq).
Note
This notebook shows how lamindb
can be used with tiledbsoma
append mode, also expained in the tiledbsoma documentation.
import lamindb as ln
import pandas as pd
import scanpy as sc
import tiledbsoma.io
from lamindb.integrations import save_tiledbsoma_experiment
from functools import reduce
→ connected lamindb: testuser1/test-scrna
ln.context.uid = "oJN8WmVrxI8m0000"
ln.context.track()
Show code cell output
→ notebook imports: lamindb==0.76.2 pandas==2.2.2 scanpy==1.10.2 tiledbsoma==1.13.1
→ created Transform('oJN8WmVrxI8m0000') & created Run('2024-08-26 16:58:15.915528+00:00')
Query the collection of h5ad
files that we’d like to convert into a single array.
collection = ln.Collection.get(
name="My versioned scRNA-seq collection", version="2"
)
collection.describe()
Show code cell output
Collection(uid='scwQcBg7xRPKgDuD0001', version='2', is_latest=True, name='My versioned scRNA-seq collection', hash='dBJLoG6NFZ8WwlWqnfyFdQ', visibility=1, updated_at='2024-08-26 16:57:38 UTC')
Provenance
.created_by = 'testuser1'
.transform = 'Standardize and append a batch of data'
.run = '2024-08-26 16:57:15 UTC'
.input_of_runs = ["'2024-08-26 16:57:49 UTC'", "'2024-08-26 16:58:08 UTC'"]
Feature sets
'var' = 'MIR1302-2HG', 'FAM138A', 'OR4F5', 'None', 'OR4F29', 'OR4F16', 'LINC01409', 'FAM87B', 'LINC01128', 'LINC00115', 'FAM41C'
'obs' = 'donor', 'tissue', 'cell_type', 'assay'
Prepare the AnnData objects¶
We need to prepare theAnnData
objects in the collection to be concatenated into one tiledbsoma.Experiment
. They need to have the same .var
and .obs
columns, .uns
and .obsp
should be removed.
adatas = [artifact.load() for artifact in collection.ordered_artifacts]
Compute the intersetion of all columns. All AnnData
objects should have the same columns in their .obs
, .var
, .raw.var
to be ingested into one tiledbsoma.Experiment
.
obs_columns = reduce(pd.Index.intersection, [adata.obs.columns for adata in adatas])
var_columns = reduce(pd.Index.intersection, [adata.var.columns for adata in adatas])
var_raw_columns = reduce(pd.Index.intersection, [adata.raw.var.columns for adata in adatas])
Prepare the AnnData
objects for concatenation. Prepare id fields, sanitize index
names, intersect columns, drop slots. Here we have to drop .obsp
, .uns
and also columns from the dataframes that are not in the intersections obtained above, otherwise the ingestion will fail. We will need to provide obs
and var
names in tiledbsoma.io.register_anndatas
, so we create these fileds (obs_id
, var_id
) from the dataframe indices.
for i, adata in enumerate(adatas):
del adata.obsp
del adata.uns
adata.obs = adata.obs.filter(obs_columns)
adata.obs["obs_id"] = adata.obs.index
adata.obs["dataset"] = i
adata.obs.index.name = None
adata.var = adata.var.filter(var_columns)
adata.var["var_id"] = adata.var.index
adata.var.index.name = None
drop_raw_var_columns = adata.raw.var.columns.difference(var_raw_columns)
adata.raw.var.drop(columns=drop_raw_var_columns, inplace=True)
adata.raw.var["var_id"] = adata.raw.var.index
adata.raw.var.index.name = None
Create the array store¶
Ingest the AnnData
objects. This saves the AnnData
objects in one array store, creates Artifact
and saves it. This function also writes current run.uid
to tiledbsoma.Experiment
obs
, under lamin_run_uid
.
soma_artifact = save_tiledbsoma_experiment(
adatas,
measurement_name="RNA",
obs_id_name="obs_id",
var_id_name="var_id",
append_obsm_varm=True,
artifact_kwargs={"description": "tiledbsoma experiment"}
)
Show code cell output
! Mutating in-memory AnnData.
! Mutating in-memory AnnData.
/opt/hostedtoolcache/Python/3.10.14/x64/lib/python3.10/abc.py:119: FutureWarning: SparseDataset is deprecated and will be removed in late 2024. It has been replaced by the public classes CSRDataset and CSCDataset.
For instance checks, use `isinstance(X, (anndata.experimental.CSRDataset, anndata.experimental.CSCDataset))` instead.
For creation, use `anndata.experimental.sparse_dataset(X)` instead.
return _abc_instancecheck(cls, instance)
/opt/hostedtoolcache/Python/3.10.14/x64/lib/python3.10/abc.py:119: FutureWarning: SparseDataset is deprecated and will be removed in late 2024. It has been replaced by the public classes CSRDataset and CSCDataset.
For instance checks, use `isinstance(X, (anndata.experimental.CSRDataset, anndata.experimental.CSCDataset))` instead.
For creation, use `anndata.experimental.sparse_dataset(X)` instead.
return _abc_instancecheck(cls, instance)
Query the array store¶
Open and query the experiment. We can use the registered Artifact
. We query X
and obs
from the array store.
with soma_artifact.open() as soma_store:
obs = soma_store["obs"]
ms_rna = soma_store["ms"]["RNA"]
n_obs = len(obs)
n_var = len(ms_rna["var"])
X = ms_rna["X"]["data"].read().coos((n_obs, n_var)).concat().to_scipy()
print(obs.read().concat().to_pandas())
Show code cell output
soma_joinid cell_type \
0 0 dendritic cell
1 1 B cell, CD19-positive
2 2 dendritic cell
3 3 B cell, CD19-positive
4 4 effector memory CD4-positive, alpha-beta T cel...
... ... ...
1713 1713 naive thymus-derived CD4-positive, alpha-beta ...
1714 1714 naive thymus-derived CD4-positive, alpha-beta ...
1715 1715 naive thymus-derived CD4-positive, alpha-beta ...
1716 1716 CD8-positive, alpha-beta memory T cell
1717 1717 naive thymus-derived CD4-positive, alpha-beta ...
obs_id dataset lamin_run_uid
0 GCAGGGCTGGATTC-1 0 wi9uuEcY08sqnlBrComO
1 CTTTAGTGGTTACG-6 0 wi9uuEcY08sqnlBrComO
2 TGACTGGAACCATG-7 0 wi9uuEcY08sqnlBrComO
3 TCAATCACCCTTCG-8 0 wi9uuEcY08sqnlBrComO
4 CGTTATACAGTACC-8 0 wi9uuEcY08sqnlBrComO
... ... ... ...
1713 Pan_T7991594_CTCACACTCCAGGGCT 1 wi9uuEcY08sqnlBrComO
1714 Pan_T7980358_CGAGCACAGAAGATTC 1 wi9uuEcY08sqnlBrComO
1715 CZINY-0064_AGACCATCACGCTGCA 1 wi9uuEcY08sqnlBrComO
1716 CZINY-0050_TCGATTTAGATGTTGA 1 wi9uuEcY08sqnlBrComO
1717 CZINY-0064_AGTGTTGTCCGAGCTG 1 wi9uuEcY08sqnlBrComO
[1718 rows x 5 columns]
Update the array store¶
Calculate PCA from the queried X
.
pca_array = sc.pp.pca(X, n_comps=2)
soma_artifact
Artifact(uid='IoIuAGjKEqd9abxy0000', is_latest=True, description='tiledbsoma experiment', key='.lamindb/IoIuAGjKEqd9abxy.tiledbsoma', suffix='.tiledbsoma', size=15056693, hash='YIyUkr5kMxEF0ruLE_4mcQ', n_objects=149, _hash_type='md5-d', visibility=1, _key_is_virtual=False, created_by_id=1, storage_id=1, transform_id=6, run_id=6, updated_at='2024-08-26 16:58:23 UTC')
Open the array store in write mode and add PCA. When the store is updated, the corresponding artifact also gets updated with a new version.
with soma_artifact.open(mode="w") as soma_store:
tiledbsoma.io.add_matrix_to_collection(
exp=soma_store,
measurement_name="RNA",
collection_name="obsm",
matrix_name="pca",
matrix_data=pca_array
)
Show code cell output
! The hash of the tiledbsoma store has changed, creating a new version of the artifact.
! artifact version None will _update_ the state of folder /home/runner/work/lamin-usecases/lamin-usecases/docs/test-scrna/.lamindb/IoIuAGjKEqd9abxy.tiledbsoma - to _retain_ the old state by duplicating the entire folder, do _not_ pass `revises`
/opt/hostedtoolcache/Python/3.10.14/x64/lib/python3.10/abc.py:119: FutureWarning: SparseDataset is deprecated and will be removed in late 2024. It has been replaced by the public classes CSRDataset and CSCDataset.
For instance checks, use `isinstance(X, (anndata.experimental.CSRDataset, anndata.experimental.CSCDataset))` instead.
For creation, use `anndata.experimental.sparse_dataset(X)` instead.
return _abc_instancecheck(cls, instance)
Note that the artifact has been changed.
soma_artifact
Artifact(uid='IoIuAGjKEqd9abxy0001', is_latest=True, description='tiledbsoma experiment', key='.lamindb/IoIuAGjKEqd9abxy.tiledbsoma', suffix='.tiledbsoma', size=15077135, hash='Po7gPuiy_Z2qF7CCxVGHtw', n_objects=158, _hash_type='md5-d', visibility=1, _key_is_virtual=False, created_by_id=1, storage_id=1, transform_id=6, run_id=6, updated_at='2024-08-26 16:58:24 UTC')