-
Notifications
You must be signed in to change notification settings - Fork 19
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Copying sdata_processed.zarr
to S3 yields java.nio.file.NoSuchFileException
#94
Comments
Hi, I think this might be fixed by nextflow-io/nextflow#3933, as the issue seems to be when publishing a full directory (both |
@cavenel Thank you, this indeed seems to be it. Do you have any suggestions on how to get around it until the fix is in place? |
Hi @R3myG, the answer from @bentsherman on the PR makes me wonder if that's actually the issue here. I have another theory. If we look at your logs, we can see that it tries multiple times before failing. What I think is happening here is that the "attempt index" is not reset between files of the same folder. And So if I am right and this issue comes from your s3 server being too slow or failing for any other reason, then the simplest solution would be to increase the
(It would also eventually be nice to have a "fail index" per file and not per folder, but this looks a bit more tricky on the Nextflow side...) |
@cavenel Thank you so much for the suggestion and for pointing me to the right setting to increase the retry delay and max attempt. Using
Seem to have solve the problem when I ran 1 sample from my samplesheet. Currently rerunning the full samplesheet to double check. PS: You inverted the values in your example, flagging it for anyone who may blindly copy paste the snipped in their |
Great, let us know if it works on the full samplesheet so that we can close this issue. |
Alright some updates on my end. After a successful run using a single sample into a completely different output directory in S3, I reverted back to my original config. I had failures due to an issue with a Docker image but the resumes worked and it all went through. I thought this was sorted for good but after the spatialVi, I've added more steps and I had a failure almost at the end (spatialVi had completed). Thankfully I had enabled the extra debugging logs which I've sanitized for sensitive details and attached. My reading is that when it reran the Clustering process, after completion it deletes certain folders then attempt to upload again? I'm curious to hear what your thoughts are and I'm currently rerunning the workflow from the beginning after clearing all the result folders. |
@cavenel I've now confirmed that the issue persist when the |
Description of the bug
Hello,
I'm running into a very weird issue with Nextflow + Tower + SpatialVi when publishing the
sdata_processed.zarr
to S3.I tested with Nextflow 24.04.3 and 24.04.4. The version of Tower is 23.3.0.
The process fails suddenly, aborting all currently running processes and just reporting a java.nio.file.NoSuchFileException in the Error report.
The error is systematic for this folder. I've checked the
sdata_processed.zarr
in the temporary directory and all is valid, I've tried to manually copy the folder to S3 with anaws s3 cp --recursive
command and all the files were copied to the S3 bucket without problems.I've sanitized the logs which are below.
Thank you,
Command used and terminal output
No response
Relevant files
System information
Nextflow 24.04.3 and 24.04.4.
Tower / Sequera cloud enterprise 23.3.0
Tested with version currently on dev of SpatialVi.
Hardware: HPC
Executor Slurm
The text was updated successfully, but these errors were encountered: