You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
So I had a question about running on multiplexed chemistries. Things like 10x flex, as well as CITE seq. At least with 10x flex, you have a pool of probes that can barcode a few samples. It comes in 16 and 4 sample probe pools. I was curious on whether you should run cellbender on the demultiplexed outputs (meaning cellbender sample 1-Pool1, cellbender sample2-Pool1, etc.). Or whether you should run cellbender on the file of all the probes in the pool. Meaning cellbender Pool1.h5ad, and then seperate the samples after. Intuitively, I would think it would perform better on the second option, because the entire probe pool is loaded onto the machine together, and is essentially one run. However, I get some weird performance when I do this:
on one side, I get a run that looks like it went ok training wise, but seems to call too many cells:
On the other hand, i get runs that seem like they went pretty poorly, and has steep drops in the training curve. I'm wondering if this is due to parameters that need to be adjusted to account for a larger sample, or whether we should run it on demultiplexed samples only.
The text was updated successfully, but these errors were encountered:
So I had a question about running on multiplexed chemistries. Things like 10x flex, as well as CITE seq. At least with 10x flex, you have a pool of probes that can barcode a few samples. It comes in 16 and 4 sample probe pools. I was curious on whether you should run cellbender on the demultiplexed outputs (meaning cellbender sample 1-Pool1, cellbender sample2-Pool1, etc.). Or whether you should run cellbender on the file of all the probes in the pool. Meaning cellbender Pool1.h5ad, and then seperate the samples after. Intuitively, I would think it would perform better on the second option, because the entire probe pool is loaded onto the machine together, and is essentially one run. However, I get some weird performance when I do this:
on one side, I get a run that looks like it went ok training wise, but seems to call too many cells:
On the other hand, i get runs that seem like they went pretty poorly, and has steep drops in the training curve. I'm wondering if this is due to parameters that need to be adjusted to account for a larger sample, or whether we should run it on demultiplexed samples only.
The text was updated successfully, but these errors were encountered: