-
Notifications
You must be signed in to change notification settings - Fork 22
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Cache entry deserialization failed, entry ignored #140
Comments
Thank you for reaching out. Firstly, I consider a total runtime of around 12 hours to be quite normal. I would start to worry if it takes more than a couple of days BR, |
|
Hi @kirilenkobm Thank you very much for your response. Following your suggestion, I did /scrfs/storage/vlamba/home/TOGA/cesar_runner.py /home/vlamba/BD-gene-loss/RERUN_CESAR_JOBS/rerun_job_1_100 /home/vlamba/BD-gene-loss/temp/cesar_results/rerun_job_1_100.txt --check_loss /home/vlamba/BD-gene-loss/temp/inact_mut_data/rerun_job_1_100.txt --rejected_log /home/vlamba/BD-gene-loss/temp/cesar_jobs_crashed_again/rerun_job_1_100.txt and the looked at /scrfs/storage/vlamba/home/TOGA/CESAR_wrapper.py ENSCGOT00000028426.1 28 /home/vlamba/BD-gene-loss/temp/toga_filt_ref_annot.hdf5 /home/vlamba/BD-gene-loss/temp/genome_alignment.bst /home/vlamba/LepNud-Chain-alignment/CG.2bit /home/vlamba/BD.2bit --cesar_binary /scrfs/storage/vlamba/home/TOGA/CESAR2.0/cesar --uhq_flank 50 --temp_dir /home/vlamba/BD-gene-loss/temp/cesar_temp_files --mask_stops --check_loss --alt_frame_del --memlim 10 CESAR JOB FAILURE Input is corrupted! Reference sequence should start with ATG! Error! CESAR output is corrupted, target must start with ATG! Error! CESAR output is corrupted, target must start with ATG! Traceback (most recent call last): File "/scrfs/storage/vlamba/home/TOGA/CESAR_wrapper.py", line 2661, in realign_exons(cmd_args) File "/scrfs/storage/vlamba/home/TOGA/CESAR_wrapper.py", line 2626, in realign_exons loss_report, del_mis_exons = inact_mut_check( File "/scrfs/storage/vlamba/home/TOGA/modules/inact_mut_check.py", line 1660, in inact_mut_check split_stop_codons = detect_split_stops( File "/scrfs/storage/vlamba/home/TOGA/modules/inact_mut_check.py", line 1468, in detect_split_stops position = exon_to_last_codon_of_exon[first_exon] KeyError: 1 It seemes that in my data It would be a great help you could suggest me possible solution for this Thanks |
thanks for checking this.
Otherwise, CESAR (a tool, that realigns reference exons to query loci) might not process such transcripts correctly in the multi-exon mode. Also, post-processing TOGA steps are based on the assumption that the provided reference transcripts have a correct reading frame. (Maybe, at some point, we will find a way to process a bit more diverse variants, and CESAR itself requires some optimisations) BR, |
Hi @kirilenkobm Thank you for your every response. Is there any way to delete these two exons with incomplete reading frames? Any tool or any command you can suggest, I would really appreciate it. Best Rgards, |
I would just do something like |
Hello, excuse me, I've encountered a similar issue as described above. I've placed the tasks on a CPU computing server, but I'm unable to submit the tasks to the slurm system. Therefore, I opted for "local." However, it has been running for five days now, and the log files indicate that it's still in progress. I'd greatly appreciate any suggestions on how to improve its speed. Thank you very much. my command is : The log shows : '''### STEP 7: Execute CESAR jobs: parallel step''' Pushing 2 CESAR job lists |
Hi Authors,
I was able to run TOGA on my couple of species, but somehow, after running for 11 hours, I got this error for one of my species in my sbatch error file.
After loading
nextflow
moduleand installing importlib-metadata using
pip install importlib-metadata
I ran toga with following command./toga.py /home/vlamba/BD-CG.chain /home/vlamba/CG.bed /home/vlamba/CG.2bit /home/vlamba/BD.2bit -i /home/vlamba/CG-isofrom.txt --project_dir /home/vlamba/BD_gene-loss --kt --cb 10,100 --cjn 500 --ms
Cache entry deserialization failed, entry ignored
Command "python setup.py egg_info" failed with error code 1 in /tmp/pip-build-3HJdri/pip/
You are using pip version 8.1.2, however version 23.3.2 is available.
You should consider upgrading via the 'pip install --upgrade pip' command.
Command "python setup.py egg_info" failed with error code 1 in /tmp/pip-build-msMJKV/importlib-metadata/
You are using pip version 8.1.2, however version 23.3.2 is available.
You should consider upgrading via the 'pip install --upgrade pip' command.
Traceback (most recent call last):
File "./toga.py", line 8, in
import importlib.metadata as metadata
ModuleNotFoundError: No module named 'importlib.metadata'
Here are some last lines from my toga .log file
`Polling iteration 653; already waiting 39180 seconds.
Polling iteration 654; already waiting 39240 seconds.
Polling iteration 655; already waiting 39300 seconds.
CESAR jobs done
Checking whether all CESAR results are complete
2 CESAR jobs crashed, trying to run again...
!!RERUN CESAR JOBS: Pushing 2 jobs into 100 GB queue
Selected parallelization strategy: nextflow
Parallel manager: pushing job nextflow /scrfs/storage/vlamba/home/TOGA/execute_joblist.nf --joblist /home/vlamba/BD_gene-loss/_cesar_rerun_batch_100
Monitoring CESAR jobs rerun
Stated polling cluster jobs until they done
CESAR jobs done
`
I would be grateful if you could suggest any solution for this.
My second concern is running time: my first species took 9hr:40min to complete when I ran TOGA for the first time, the second took 10hr:21min, and the third failed after 11 hrs.
Kindly have a look at my shared command and suggest the best way to run this tool faster on my data.
Thank you
The text was updated successfully, but these errors were encountered: