Replies: 4 comments 4 replies
-
Hi @FabianDK , first, I would go through the docs page on caching and resuming, particularly the "tips" at the end. There is a method to compare the task hashes across two runs in a nice diff, it should help us figure out where the resume is failing |
Beta Was this translation helpful? Give feedback.
-
It would be good if the developers can describe the expected behavior. The first question is if resume is supposed to work if the same workflow is launched with Azure batch from a different folder/VM/etc, with all input files, cacheDir and workDir in the cloud blob storage. When I rerun a successfully executed cloud workflow from another VM or folder (not copying over the local .nextflow folder), then I get "WARN: It appears you have never run this project before -- Option I have run the hash compare method that you suggested: Initial run from Folder 1:
Resumed run from Folder 2 (without the .nextflow folder):
It does not seem to contain any relevant info? There is a similar issue here: #1747 |
Beta Was this translation helpful? Give feedback.
-
@bentsherman I looked at the source code in and it seems like it is looking for the last session id in the history file.Since we executed the nextflow run in a container, obviously the subsequent nextflow run with -resume flag in the new container does not know any thing about the previous .nextflow/history . Only the .nextflow/cache is persisted in the cloud (s3) but I don't think the .nextflow/history is persisted along with that cache.
I was under the impression that as long as we have the cloudcache persisted, we should be able to resume the nextflow run without relying on the local .nextflow/history file. Did I miss something here? Thanks! |
Beta Was this translation helpful? Give feedback.
-
Would it be possible to have the Since the |
Beta Was this translation helpful? Give feedback.
-
Dear authors,
We are working with nextflow in the Azure cloud using docker containers and all input/output files and workDir in the blob storage as well.
The issue we encountered is that when the same nextflow run command is launched in different folders, the resume function is not working. This is despite everything being equal (main.nf, config, modules, cloud files) apart from the directory where the nf command is relaunched.
The command we use is:
It is a simple test script that successfully executes when run in
~/nextflow_lib
, but when relaunched in a new folder~test/nextflow_lib
it does not understand the resume (resume = true) defined in the config.The only way to preserve the cache seems to be to copy the local .nextflow folder in
~/nextflow_lib
to the folder where it should be re-executed (~test/nextflow_lib
).Also tried downloading the cache from the cloud to the new folder, but this did not solve the issue.
Isn't the purpose of the cloudcache to enable this?
Thanks,
Daniel
Beta Was this translation helpful? Give feedback.
All reactions