Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Quality score differences when running dorado on the Gridion vs on a linux server #1110

Open
abridgeland opened this issue Oct 29, 2024 · 3 comments
Labels
models Issues related to model selection

Comments

@abridgeland
Copy link

Hello, I observed that the basecall quality scores are lower when running dorado on Gridion vs off the gridion. Do you know why that might be? For both runs, I used the super high accuracy basecalling option. I specifically noticed that there was a 40 percent increase in the number of reads greater than Q15 when running dorado off the Gridion. Additionally, the mean read quality score was 13.3 when running dorado on the gridion vs 16.6 when running the basecaller off the gridion

@HalfPhoton
Copy link
Collaborator

Hi @abridgeland,
I suspect you're using a different basecalling model in both cases. The most recent (and best) models are available in stand-alone dorado before they're available on device. So you might have used a newer version of the same model architecture and saw the improvement.

You can check which model was used by inspecting the SAM header.

Kind regards,
Rich

@HalfPhoton HalfPhoton added the models Issues related to model selection label Oct 30, 2024
@abridgeland
Copy link
Author

abridgeland commented Oct 31, 2024

Thanks for your suggestion. I did check the headers and found that you were correct and there was a newer model run using the stand-alone dorado. However, when I reran using the same model, I still noticed some differences in the results and higher quality data in the stand alone version. Please see my results below. We used the following model: [email protected]
Dorado_comparison.xlsx

@HalfPhoton
Copy link
Collaborator

Hi @abridgeland, glad to hear they're much closer now - but that's still a greater difference than I'd expect too.

Are these on the same version of dorado and using the all of the same settings?

Kind regards,
Rich

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
models Issues related to model selection
Projects
None yet
Development

No branches or pull requests

2 participants