-
Notifications
You must be signed in to change notification settings - Fork 3
Add missing input data files needed for E3SM tests #32
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
Conversation
|
@mahf708 👋 This repository doesn't have Copilot instructions. With Copilot instructions, I can understand the repository better, work faster and produce higher quality PRs. I can generate a .github/copilot-instructions.md file for you automatically. Click here to open a pre-filled issue and assign it to me. I'll write the instructions, and then tag you for review. |
Co-authored-by: mahf708 <[email protected]>
This reverts commit 7037a4d.
|
@mahf708 Is this meant to fix all the |
|
Yes, just waiting on the input server to be back up and then we run this, update the other containers here and quick pr on the e3sm repo to increment up the version |
|
Which recently added tests need the new data? We can probably reduce it somehow. Lower resolution or fewer time slices. |
|
I can take a deeper look at trimming these containers, but I'm relatively swamped these days, so it will take me a bit of time (weeks)... I hope someone could take a look before we get into trouble with the server people :/ |
|
Would having 2+ containers be sustainable? We can have a "base" one, with just the software stack, and build on top of that a few containers that have the input data needed by different compsets/tests. I don't think it's a beautiful solution, but it might help. Edit: i just googled the gh containers size limit, to see what the cap was, and google AI says the limit is 10GB per layer. Is that correct? If so, do we just need to split data into multiple layers? Seems to good to be true, but I figured I asked... |
|
we can get rid of 20 gb of stuff in these containers, but it will take careful work. I can get to this later... I think your solution makes sense to me. Note that I am kind of doing exactly that now right. I build a bare base now with data only, then add the software stack on top of that. I guess what I am trying to say is, this is not really a hard problem, it will just take time iterating. I will get to it at some point, but I am short on time for the next several weeks (lots of non-software tasks that I delayed over the months ... :/ ... ) |
This PR adds missing input data files that are required for E3SM tests but were not previously included in the container build files.
Background
Recent E3SM test runs were failing due to missing input data files. The test logs showed that several files were being downloaded successfully but were not pre-included in the container, causing tests to fail when network access was limited or when the files weren't available during the test execution.
Changes Made
Files Added to
inputdata/files.txt(12 new files):Ocean/SST data files:
domain.camocn.0.9x1.25_gx1v6_090403.nc- Ocean fraction domain filesst_HadOIBl_bc_0.9x1.25_clim_c040926.nc- SST boundary conditionsDATM forcing files:
nyf.giss.T62.051007.nc,nyf.gxgxs.T62.051007.nc,nyf.ncep.T62.050923.nc- NYF forcing dataCOREv2.correction_factors.T62.121007.nc- CORE2 correction factorsdomain.T62.050609.nc- DATM domain fileSCREAM model files:
vn_table_vals_v2.dat8,vm_table_vals_v2.dat8,revap_table_vals_v2.dat8,mu_r_table_vals_v2.dat8map_ne4pg2_to_ne2pg2_c20240902.ncFiles Added to
inputdata/files-standalone.txt(5 new files):Added the SCREAM v2 table files and mapping file to support standalone SCREAM tests.
Impact
These additions ensure that E3SM tests have access to all required input data files without needing to download them at runtime, improving test reliability and performance. The changes are purely additive - no existing files were removed or modified.
All new URLs follow the established pattern:
https://web.lcrc.anl.gov/public/e3sm/inputdata/...Fixes #31.
Warning
Firewall rules blocked me from connecting to one or more addresses (expand for details)
I tried to connect to the following addresses, but was blocked by firewall rules:
web.lcrc.anl.govcurl -I REDACTED(dns block)If you need me to access, download, or install something from one of these locations, you can either:
✨ Let Copilot coding agent set things up for you — coding agent works faster and does higher quality work when set up for your repo.