-
Notifications
You must be signed in to change notification settings - Fork 4
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
RUM doesn't know it stopped #162
Comments
Couple of questions: What sort of cluster are you running on? A grid-engine cluster? If so, did you use the --qsub flag to submit? |
I am not splitting the chunks over multiple nodes. I am using LFS and it supports MPI but I didn't try here -- just splitting into 8 chunks and requesting 8 cores from the same node. RUM finished successfully using the same configuration on untrimmed reads, so I am wondering whether the trimming or padding messed up RUM. I mentioned the cluster because I submit the job and wait for an email to tell me it finished but after 48h the time limit expires and RUM did not finish.
File 8's log ends with:
so it decided to stop early? In the tmp directory is a file called preproc-error-log which contains:
Could this be the culprit? |
Can you zip or tar up the log directory and send it to me? Also, if you can Thanks, Mike On Tue, Jan 22, 2013 at 10:23 AM, asowalsky [email protected]:
|
There is something odd in your output file. Note that you have 8 chunks but step 6 on only report 7 chunks. XXXXXXXX 5. Make unmapped reads file for blat On Jan 22, 2013, at 10:23 AM, asowalsky [email protected] wrote:
|
Thank you for your feedback. Agreed, this is a major limitation at this
|
Yes, you may want to check the end of your input file. What do you see if On Tue, Jan 22, 2013 at 10:50 AM, asowalsky [email protected]:
|
RUM used to do an integrity check upfront that the two files were the On Tue, 22 Jan 2013, Mike DeLaurentis wrote:
|
I tried trimming first but then the FQ files were not the same size On Tue, Jan 22, 2013 at 10:58 AM, greggrant [email protected]:
|
You found the problem! My post-trim padding script (really a PE fix script) I never had an issue with BWA so I never padded before. I can fix this -- On Tue, Jan 22, 2013 at 11:02 AM, Mike DeLaurentis <[email protected]
|
Or at least verify that the FQ files are formed correctly. In my case, the On Tue, Jan 22, 2013 at 11:05 AM, greggrant [email protected]:
|
It looks like the preprocessing script found that the files ended with an I think the check that makes sure the forward and reverse FQ files are the On Tue, Jan 22, 2013 at 11:32 AM, asowalsky [email protected]:
|
I have a pair of FastQ's from paired end sequencing that had some adapter contamination, so I trimmed them. But since RUM doesn't like FQ files with different sizes (HUGE ISSUE) I padded all the reads with N's and @'s. Before trimming RUM had no problem (just poor mapping).
After trimming, RUM stalls after step 16. I run this on a Cluster and the job is still running but RUM thinks it is stopped.
Here's the screen output:
RUM Version v2.0.4
I'm going to try to figure out how much RAM you have. If you see some error
messages here, don't worry, these are harmless.
Use of uninitialized value $params{"message"} in print at /usr/share/perl5/Log/Log4perl/Appender/File.pm l
ine 245.
Use of uninitialized value $params{"message"} in print at /usr/share/perl5/Log/Log4perl/Appender/File.pm l
ine 245.
It seems like you have 94.00 Gb of RAM on your machine. Unless you have too
much other stuff running, RAM should not be a problem.
Processing as paired-end data
Saving job configuration
If this is a big job, you should keep an eye on the rum_errors*.log files
in the output directory. If all goes well they should be empty. You can
also run "/home/as459/RUM_mapper/bin/rum_runner status -o
/scratch/as459/trimmed_S11-49946_RNA/G3" to check the status of the job.
Preprocessing
Processing as paired-end data
Reformatting reads file... please be patient.
Splitting fastq file into 8 chunks with separate reads and quals
*** Note: I am going to report alignments of length 35, based on a read
length of 50 . If you want to change the minimum size of alignments
reported, use the --min-length option
Processing in 8 chunks
Checking that files were split properly
And here's what happens when I poll the status:
Preprocessing
XXXXXXXX Split input files
Processing in 8 chunks
XXXXXXXX 1. Run Bowtie on genome
XXXXXXXX 2. Run Bowtie on transcriptome
XXXXXXXX 3. Merge unique mappers together
XXXXXXXX 4. Merge non-unique mappers together
XXXXXXXX 5. Make unmapped reads file for blat
XXXXXXX 6. Run BLAT
XXXXXXX 7. Merge bowtie and blat results
XXXXXXX 8. Clean up RUM files
XXXXXXX 9. Produce RUM_Unique
XXXXXXX 10. Sort RUM_Unique by location
XXXXXXX 11. Sort cleaned non-unique mappers by ID
XXXXXXX 12. Remove duplicates from NU
XXXXXXX 13. Create SAM file
XXXXXXX 14. Create non-unique stats
XXXXXXX 15. Sort RUM_NU
XXXXXXX 16. Generate quants
Postprocessing
All error log files are empty. That's good.
RUM is not running.
Any ideas what is going on?
The text was updated successfully, but these errors were encountered: