Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

erro: Argument list too long when using latest cenote-Taker2.1.5 #37

Open
hazel914 opened this issue Oct 12, 2022 · 2 comments
Open

erro: Argument list too long when using latest cenote-Taker2.1.5 #37

hazel914 opened this issue Oct 12, 2022 · 2 comments

Comments

@hazel914
Copy link

Hi Mike,
Thank you for Cenote-Taker2! However, I am facing a problem when i run the commond:
python ~/miniconda3/bin/Cenote-Taker2/run_cenote-taker2.py -c ~/hxj/depth_analysis/raw_data_2/contig_megahit/SRR9161502_9000000/final.contigs.fa -r SRR9161502_9000000 -p true -m 32 -t 32 2>&1 | tee output.log

the tail of the output.log file is like:
`SRR9161502_900000074949.fasta has DTRs/circularity
SRR9161502_9000000541967.fasta has DTRs/circularity
SRR9161502_9000000137044.fasta has DTRs/circularity
no reads provided or reads not found
Circular fasta file(s) detected

Putting non-circular contigs in a separate directory
time update: running IRF for ITRs in non-circular contigs 10-12-22---00:21:13
/media/home/user11/miniconda3/bin/Cenote-Taker2/cenote-taker2.1.5.sh: line 464: /usr/bin/find: Argument list too long
time update: running prodigal on linear contigs 10-12-22---01:20:08
`
Could you help me solve this problem?Is that because my input file has too much sequences?

Regards!
Hazel

@Stevenleizheng
Copy link

I also have this question. I think this question is caused by our input big data of fasta format. The size of my fasta data is approximately 250M.

@mtisza1
Copy link
Owner

mtisza1 commented Nov 2, 2022

Hi Hazel and Steven,

I agree that this issue is caused by the find command on line 464 of cenote-taker2.1.5 crashing because there are too many files in the folder. This has been a problem in the past and is a bit tricky for me to code around because some computers/nodes can handle a lot more files than others. I hope it's OK for you to split your input fasta and rerun. Quick way to split the input fasta in to chunks of up to 5000 contigs:

conda activate cenote-taker2_env
seqkit split input_contigs.fasta -s 5000

https://bioinf.shenwei.me/seqkit/usage/#split

I hope this works for you!

Mike

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants