Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

WW3 does not support concurrent builds #550

Closed
climbfuji opened this issue Apr 30, 2021 · 9 comments · Fixed by #1089
Closed

WW3 does not support concurrent builds #550

climbfuji opened this issue Apr 30, 2021 · 9 comments · Fixed by #1089
Labels
bug Something isn't working

Comments

@climbfuji
Copy link
Collaborator

Description

PR #527 moved the coupled tests to the top of rt.conf, since the coupled tests take longest to wait tin the queue and to run.

This meant that both the CLDP+WAV compile job was running concurrently with the ATMW compile job - which led to compile errors.

Before this move, we were "lucky" in a sense that only ten builds can happen at the same time, and the two builds that involve WW3 are sufficiently far apart to not overlap.

To Reproduce:

Run the two compile jobs that use WW3 at the same time.

Solution

Improve the WW3 build to support clean out of source builds.

@JessicaMeixner-NOAA
Copy link
Collaborator

This is true and was a known issue (NOAA-EMC/WW3#49). However, I had forgotten about this and this explains an issue I was having yesterday. We might be able to leverage this WW3 PR: NOAA-EMC/WW3#331 and get a fix on a weeks time scale. If this doesn't work, then we're looking a months time scale for a fix. We're in the process of cleaning up the WW3 code and then moving to CPP instead of switches (NOAA-EMC/WW3#360) and then updating the build with cmake, etc; but the timeline is in months not days/weeks.

@JessicaMeixner-NOAA
Copy link
Collaborator

Another idea that would not require waiting on any WW3 fixes (which need and will happen), could be to combine the build options so that we just need 1 WW3 compile for all the wave related tests. Then you have one less build too.

@climbfuji
Copy link
Collaborator Author

Another idea that would not require waiting on any WW3 fixes (which need and will happen), could be to combine the build options so that we just need 1 WW3 compile for all the wave related tests. Then you have one less build too.

Not sure if that works, because we are building different apps (ATMW and S2SW).

@JessicaMeixner-NOAA
Copy link
Collaborator

ATMW is just a subset of the components of S2SW, so I think if you added the CCPP suites if they were different, that you could just run the ATMW cases with the S2SW exe.

@climbfuji
Copy link
Collaborator Author

climbfuji commented Apr 30, 2021 via email

@DusanJovic-NOAA
Copy link
Collaborator

There is a logic in the regression test scripts that serializes all COMPILE jobs that are building WW3.

# serialize WW3 builds. FIXME

Unfortunately this logic is now broken since we switched from using compile option flags that turn on individual component (WW3=Y) to using APPS=<appname> option

@SMoorthi-emc
Copy link
Contributor

SMoorthi-emc commented Apr 30, 2021 via email

DusanJovic-NOAA pushed a commit that referenced this issue Apr 30, 2021
…n CCPP (#527)

- Update the submodule pointers for ccpp-physics and fv3atm and the regression test baseline date tag
- Small update of `tests/ci/repo_check.sh` from @MinsukJi-NOAA
- Move coupled tests in `rt.conf` to the top, because these take the longest to wait in the queue and to run; **BUT**: needed to move ATMW tests to end of `rt.conf`, since WW3 is not able to support parallel builds. By moving it to the end, we are as "lucky" as before in a sense that only ten builds can happen at the same time, and the two builds that involve WW3 are sufficiently far apart to not overlap - see issue #550
- Change location of temporary/personal baseline directories on Cheyenne from $WORK to $SCRATCH to avoid disk space limitations
- Increase walltime for compile jobs on wcoss_dell_p3 from 50min to 1hr

Co-authored-by: Dustin Swales <dustin.swales@noaa.gov>
@arunchawla-NOAA
Copy link

@climbfuji @aliabdolali @JessicaMeixner-NOAA is this still an issue ? Does the ww3 cmake build help here ? The issue on ww3 side has been closed

@JessicaMeixner-NOAA
Copy link
Collaborator

Should have a PR to ufs-weather-model by the end of today with a fix for this as CMAKE does resolve this issue.

@JessicaMeixner-NOAA JessicaMeixner-NOAA mentioned this issue Mar 8, 2022
16 tasks
pjpegion pushed a commit to NOAA-PSL/ufs-weather-model that referenced this issue Apr 4, 2023
…munity#550)

* Add 'time_iso' variable to 6-tile history files
* Add '_Encoding' attribute to time_iso

Co-authored-by: jswhit2 <Jeffrey.S.Whitaker@noaa.gov>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
Development

Successfully merging a pull request may close this issue.

5 participants