feat!: timeouts#457
Conversation
7182a85 to
111e53f
Compare
This is a problem; if this is something we are going to use for the benchmarking study we need this to work with singularity because CHTC only uses singularity/apptainer. |
|
I originally assumed this was a more esoteric PR to test the error-handling workflow. Though, based on the meeting just now, I'll look into a nice way to get this working with Singularity. |
There was a problem hiding this comment.
Here is my first round review. I like the empty files if an error occurs, and it is good to have an associated log explaining why.
(Adding this comment again here) My only issue with this is the definition of "error." If a parameter combination fails the heuristics check, I do not want the output to be empty. I want the output to reflect what was actually produced, so I do not have to rerun that combination even though it "failed" the heuristics. I should be able to freely update the heuristics and have that output counted if it now passes, without rerunning combinations that previously produced empty output. In short, a parameter combination failing the heuristics should not be classified as an error as defined in this PR.
This PR not working with singularity is a very big problem because of the chtc integration.
"ML requires at least one pathway, and failing pathways can break ML-work. How do we want to handle downstream analysis when errors occur." This seems like a separate problem that I will fix internally in the ML code (I want to still make figures if we only have one pathway or a set of empty pathways).
This perplexes me but from my tests we do not need --keep-going. I do not know my original intent here
Co-Authored-By: Neha Talluri <78840540+ntalluri@users.noreply.github.com>
The small part here that is adaptable is that errors, in this PR, are only made in the |
|
This works with apptainer now, but this now touches an untested part of profiling, so that part needs a review from @jhiemstrawisc. |
ntalluri
left a comment
There was a problem hiding this comment.
here is half of a review on this pr
Co-Authored-By: Neha Talluri <78840540+ntalluri@users.noreply.github.com>
|
I've made a few changes:
The final bullet point here also means that the conditionals PR will now actually depend on this PR. |
|
does this need to use #464? |
|
No - I was thinking that the |
why did i not do this before???
this is why I didn't do this earlier...
|
I seem to have messed with imports when I went to pass in the entirety of |
Adds
timeoutto algorithms as a demonstration of passing through errors, and introduces run settings, breaking the original configuration design by introducing a newparamsoption. Closes #316.Caveat: ML requires at least one pathway, and failing pathways can break ML-work. How do we want to handle downstream analysis when errors occur (including in the future heuristic errors?)