-
Notifications
You must be signed in to change notification settings - Fork 1
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
How to produce RV=$tmp/${sample}_R1_filtered.final.fastq.gz #1
Comments
I am also stuck at the same point. Would greatly appreciate any advice! Best wishes |
Hi Duo and Sina, The raw data I uploaded to GEO has already been filtered, with the R1 file for each sample ready to be used as input for the st_pipeline. However, the GEO team has requested the upload of the original, untrimmed raw data, and I am currently working on addressing this request. Best, |
Hi Dr.@Zhiliang-Bai , Thanks for your reply. I think we still need to trim the TSO primer (AAGCAGTGGTATCAACGCAGAGTGAATrGrG+G, denoted in sup table3) as in 10X pipeline before running st_pipeline? st_pipeline doesn't include this trimming. So if we don't trim these TSO primers, it will lead to a lot of mismatchs and may cause these reads cannot be mapped to the reference genome. Did you do this before running st_pipeline? Thank you! Best, |
Hi Zhiliang and Duo Thank you very much for your reply Zhiliang. I agree with Duo, are there any specific trimming requirements needed before alignment? I am having trouble getting any significant mapping of the read 1 transcripts to the reference genome. Best wishes |
Dear @jjznn , Did you show the GSM8454077_MouseE13_Rep1_50um sample? What is the number of reads before mapping (after trimming)? Best, |
I downsampled GSM8454077_MouseE13_Rep1_50um for quick check, so the number of reads are small.
…---- Replied Message ----
| From | Duo ***@***.***> |
| Date | 12/09/2024 23:10 |
| To | ***@***.***> |
| Cc | ***@***.***>***@***.***> |
| Subject | Re: [Zhiliang-Bai/Patho-DBiT] How to produce RV=$tmp/${sample}_R1_filtered.final.fastq.gz (Issue #1) |
Dear @jjznn ,
Did you show the GSM8454077_MouseE13_Rep1_50um sample? What is the number of reads before mapping (after trimming)?
Best,
Duo
—
Reply to this email directly, view it on GitHub, or unsubscribe.
You are receiving this because you were mentioned.Message ID: ***@***.***>
|
Hi @jjznn , |
Hi @xieduo7 , And my log file of cutadapt: === Summary === Total reads processed: 22,961,933 == Read fate breakdown == Total basepairs processed: 3,444,289,950 bp === Adapter 1 === Sequence: AAGCAGTGGTATCAACGCAGAGTGAATGGG; Type: anchored 5'; Length: 30; Trimmed: 21987666 times No. of allowed errors: 3 Overview of removed sequences st_pipeline log file: It seems I have problem in mapping rate. Thank you for taking a look for me. |
Hi @jjznn , My scenario is very similar to yours. I am also trying to find ways to improve the mapping rate. Given that the mapping rate based on trimmed reads is relatively acceptable (about 80%), I think one key to this problem is to improve the number of reads after trimming. I'm still doing some experiments, but haven't made any improvements. Best, |
Dear all, Thank you for your valuable feedback. We utilized our in-house developed algorithms, extending beyond the capabilities of the st-pipeline. However, we were unable to fully finalize this aspect at the time of publication. Our team is actively working on this, and we plan to submit a comprehensive computational work addressing this within the next two months. We kindly ask for your patience as we complete this important component. Thank you for your understanding! Best, |
Hi @Zhiliang-Bai ,
Thanks for your code.
I am wondering how to generate
RV=$tmp/${sample}_R1_filtered.final.fastq.gz
from raw reads before runningst_pipeline_run.py
. Trimming read 1 with Cutadapt? I was confused because I saw thatst_pipeline_run.py
can also trim the reads.Thank you!
Best,
Duo
The text was updated successfully, but these errors were encountered: