GeoscienceAustralia / nwra

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

National Wind Risk Assessment

This repo collects the code developed as part of Geoscience Australia's contribution to the National Wind Risk Assessment (NWRA), a component of the National Climate Risk Assessment (NCRA; 2023-24).

Regional wind hazard

The regional hazard data is generated in ArcGIS Pro, using geoprocessing functions to convert shapefiles to rasters.

The Regional wind hazard provenance.ipynb notebook is used to generate provenance information for the regional hazard data. Running the code in the notebook will generate the provenance file.

applyMultpliers.cmd is a Windows command file to run the applyMultipliers.py script. This will combine the regional wind hazard data with the M3_max site exposure data to generate local wind hazard data. While the Windows command file is the main way to run this script, there's nothing to stop you running the python script on a *NIX machine. Just ensure the conda environment is properly set up (I haven't added a conda environment file yet CA 2023-12-05)

Vulnerability

Validate vulnerability file.ipynb is used to ensure the XML file that holds the vulnerability files is correctly formed NRML v0.5 data.

Vulnerability model NRML file provenance.ipynb generates the provenance information for the vulnerability model file.

Impact calculations

run_hazimp_batch.sh is submitted to the job queue on gadi.nci.org.au to run a collection of HazImp simulations in parallel. This script bulds a list of commands based on the available wind hazard layers in the scratch partition, and then calls run_hazimp.sh to execute HazImp for each wind hazard layer.

run_hazimp_batch.sh is set up to run a single region (i.e. State/Territory), which is specified when submitting the job.

To run, edit the template configuration file hazimp_template.yaml. RPX will be replaced with strings like RP100, while STATE will be replaced with the string value for the region ("WA", "NT", etc.). In most cases, the output folders (and any required parent directories) will be created in run_hazimp.sh.

Submit the job script to the queue, specifying the region::

qsub -v STATE=QLD run_hazimp_batch.sh

There is no checking done on the regions - if you try "NSZ", then the job will fail when it tries to actually run HazImp and finds there's no exposure file that starts with "NSZ". Currently the list of regions is "WA", "SA", "NT", "TAS", "VIC", "NSW", "ACT" and "QLD". "OT" won't work, because there is no hazard data for the other territories.

The resources specified in run_hazimp_batch.sh should be sufficient for all regions. Check the bigger ones ("NSW", "VIC") - increasing memory should do the trick.

Copyright

This work © 2023 by Commonwealth of Australia (Geoscience Australia) is licensed under CC BY 4.0 International

Contacts

Craig Arthur, craig.arthur@ga.gov.au

About

License:Apache License 2.0


Languages

Language:Jupyter Notebook 68.5%Language:Python 23.5%Language:Shell 7.6%Language:Batchfile 0.3%