You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Is your feature request related to a problem? Please describe.
In order to evaluate my error correction tool, I though to build a confusion matrix by filtering one dataset containing bad reads and use as reference one dataset containing the corrected version of the first one reads.
Describe the solution you'd like
I tried to generate two datasets using the same parameters as input (also, the seed in fixed in order to maintain a deterministic generation of sequences) but with quality and glitches values differents. The result is two different collections of reads, coming from different locations of the reference genome.
Describe alternatives you've considered
What I need are two datasets with the same reads, one with perfect quality and one with glitches and errors. In this way, I could compare the perfect dataset with the one corrected by my tool and creating the confusion matrix.
Additional context
I'm honestly not sure if this is already possible, but I have not been able to find the solution to my problem in the documentation or on the internet.
The text was updated successfully, but these errors were encountered:
Is your feature request related to a problem? Please describe.
In order to evaluate my error correction tool, I though to build a confusion matrix by filtering one dataset containing bad reads and use as reference one dataset containing the corrected version of the first one reads.
Describe the solution you'd like
I tried to generate two datasets using the same parameters as input (also, the seed in fixed in order to maintain a deterministic generation of sequences) but with quality and glitches values differents. The result is two different collections of reads, coming from different locations of the reference genome.
Describe alternatives you've considered
What I need are two datasets with the same reads, one with perfect quality and one with glitches and errors. In this way, I could compare the perfect dataset with the one corrected by my tool and creating the confusion matrix.
Additional context
I'm honestly not sure if this is already possible, but I have not been able to find the solution to my problem in the documentation or on the internet.
The text was updated successfully, but these errors were encountered: