Importance of normalisation method for Reference Mapping #4627
Unanswered
chris-rands
asked this question in
Q&A
Replies: 1 comment 1 reply
-
|
While you may get reasonable results when using log-normalization for query and SCTransform for reference, we don't recommend this. The following script can be used to convert a log_normalized UMI matrix into a counts matrix. It only works if the original dataset was a UMI matrix, and therefore the smallest non-zero value in each cell vector represents 1 UMI. You should be able to use this to convert the data in your AnnData object to a counts matrix, and then can map as described in the vignette. |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
I'm testing the nice reference mapping feature. It says "The reference was normalized using SCTransform(), so we use the same approach to normalize the query here." My question: is that important? I have a scanpy AnnData object that was processed with a scandard scanpy workflow (normalize_total, log1p, no regression/scaling), which I convert to h5seurat format- is it okay to use this to compare to the Seurat PBMC reference, which was normalised differently? The scanpy object does not hold the raw counts so I cannot re-normalize. The results look promising, thanks
(Cross-posted with issue #4625, feel free to delete one of these threads)
Beta Was this translation helpful? Give feedback.
All reactions