Sorry to ask, I’m sure what I’m trying is trivial but I cannot find an example in docs or on the forum.
I’ve got a nextflow script involving Dragen & AWS Batch where the outputs are written directly to the F1 instance I launch. I stream fastq input from S3.
I’ve a very simple config, with an out_dir like so:
If afterwards, I do the below, I pick up all the inputs and outputs so I can see it exists. However nextflow is not pushing the results automatically. Am I missing a major step?
Have you tested how long it takes for the files to appear in the publish directory? It shouldn’t take too long, but some time after the pipeline run is finished is expected.
Can you think of the TBs of data that would be transferred and stored somewhere if Nextflow transferred everything a task produces? 🫨 That’s why we need the output block to:
a) Throw a warning/error if the task doesn’t deliver what’s supposed to be delivered
b) Make sure the specific set of outputs required by the next task will be available for this next task
c) Guarantee only meaningful output is taken into consideration for reports (e.g. with MultiQC) and published in our results folder.