The NODD AWS cloud is used to disseminate model outputs from National Ocean Service's operational forecast systems (OFS).
Starting November 20, 2024, model output files will be available in the current and new directory structures. The current file directory structure is: OFS/netcdf/YYYYMM, where one single folder usually contains over 7000 files and a large amount of memory is required for the THREDDS applications.
The current directory and new structures will co-exist until January 1, 2025, when the current directory structure will be decommissioned.
The Lake Superior Operational Forecast System (LSOFS) was jointly developed by NOAA/National Ocean Service's (NOS) Center for Operational Oceanographic Products and Services (CO-OPS) and Office of Coast Survey (OCS), the Great Lakes Environmental Research Laboratory (GLERL), the NOAA/National Weather Service's (NWS) National Centers for Environmental Prediction (NCEP) National Central Operations (NCO), and the University of Massachusetts-Dartmouth.
The new LSOFS uses the Finite Volume Community Ocean Model (FVCOM), coupled with an unstructured grid version of the Los Alamos Sea Ice model (CICE), to provide users with higher resolution of nowcast (analyses of near present) and forecast guidance of water levels, currents, water temperature, ice concentration, ice thickness and ice velocity out to 120 hours, four times per day. By invoking advanced model schemes and algorithms, LSOFS is expected to generate a more accurate model output than the former LSOFS, whose model core was the Princeton Ocean Model (POM). Please note that CO-OPS is distributing ice-related products only in a digital NetCDF format. No graphical products are provided.
The NWS and NOS work together to run LSOFS operationally on NOAA's High Performance Computing System (HPCS). By running on NOAA's HPCS, LSOFS has direct access to National Weather Service operational meteorological products that are required for reliable operations.
For more information about LSOFS, please visit please click here.
For more information about FVCOM, please click here.