Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[1 pt] PR: Add CI - GitHub Actions Workflow file #1030

Merged
merged 13 commits into from
Nov 17, 2023
Merged
Show file tree
Hide file tree
Changes from 6 commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
22 changes: 22 additions & 0 deletions .github/workflows/lint_and_format.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,22 @@
name: Lint and Format Using Pre-Commit

on:
pull_request:
branches:
- dev
- main
workflow_dispatch:

permissions:
contents: read

jobs:
lint-and-format:
runs-on: ubuntu-latest
steps:
- uses: actions/checkout@v3
- name: Set up Python
uses: actions/setup-python@v4
with:
python-version-file: pyproject.toml
- uses: pre-commit/action@v3.0.0
5 changes: 2 additions & 3 deletions .pre-commit-config.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -14,11 +14,10 @@ repos:
- id: check-json

- repo: https://github.com/PyCQA/flake8
rev: 6.0.0
rev: 6.1.0
hooks:
- id: flake8
entry: pflake8
additional_dependencies: [pyproject-flake8]
additional_dependencies: [flake8-pyproject]

- repo: https://github.com/psf/black
rev: 23.7.0
Expand Down
2 changes: 1 addition & 1 deletion CONTRIBUTING.md
Original file line number Diff line number Diff line change
Expand Up @@ -65,7 +65,7 @@ If you would like to contribute, please follow these steps:
# Check all files in the repo
pre-commit run -a

# Run only the black formatting tool
# Run only the flake8 formatting tool
pre-commit run -a flake8
```

Expand Down
2 changes: 1 addition & 1 deletion Pipfile
Original file line number Diff line number Diff line change
Expand Up @@ -39,7 +39,7 @@ scipy = "==1.10.1"
gval = "==0.2.3"
flake8 = "==6.0.0"
black = "==23.7.0"
pyproject-flake8 = "==6.0.0.post1"
flake8-pyproject = "==1.2.3"
pre-commit = "==3.3.3"
isort = "==5.12.0"
urllib3 = "==1.26.18"
Expand Down
1,000 changes: 475 additions & 525 deletions Pipfile.lock

Large diffs are not rendered by default.

2 changes: 1 addition & 1 deletion data/wbd/generate_pre_clip_fim_huc8.py
Original file line number Diff line number Diff line change
Expand Up @@ -131,7 +131,7 @@ def pre_clip_hucs_from_wbd(wbd_file, outputs_dir, huc_list, number_of_jobs, over
if number_of_jobs > total_cpus_available:
print(
f'Provided: -j {number_of_jobs}, which is greater than than amount of available cpus -2: '
f'{total_cpus_available -2} will be used instead.'
f'{total_cpus_available - 2} will be used instead.'
)
number_of_jobs = total_cpus_available - 2

Expand Down
2 changes: 1 addition & 1 deletion data/write_parquet_from_calib_pts.py
Original file line number Diff line number Diff line change
Expand Up @@ -231,7 +231,7 @@ def create_parquet_files(
if number_of_jobs > total_cpus_available:
logging.info(
f'Provided: -j {number_of_jobs}, which is greater than than amount of available cpus -1: '
f'{total_cpus_available -1} will be used instead.'
f'{total_cpus_available - 1} will be used instead.'
)
number_of_jobs = total_cpus_available - 1

Expand Down
39 changes: 39 additions & 0 deletions docs/CHANGELOG.md
Original file line number Diff line number Diff line change
@@ -1,6 +1,45 @@
All notable changes to this project will be documented in this file.
We follow the [Semantic Versioning 2.0.0](http://semver.org/) format.

## v4.4.6.0 - 2023-11-9 - [PR#1030](https://github.com/NOAA-OWP/inundation-mapping/pull/1030)

This PR introduces the `.github/workflows/lint_and_format.yaml` file which serves as the first step in developing a Continuous Integration pipeline for this repository.
The `flake8-pyproject` dependency is now used, as it works out of the box with the `pre-commit` GitHub Action in the GitHub Hosted Runner environment.
In switching to this package, a couple of `E721` errors appeared. Modifications were made to the appropriate files to resolve the `flake8` `E721` errors.
Also, updates to the `unit_tests` were necessary since Branch IDs have changed with the latest code.

A small fix was also included where `src_adjust_ras2fim_rating.py` which sometimes fails with an encoding error when the ras2fim csv sometimes is created or adjsuted in windows.

### Changes
- `.pre-commit-config.yaml`: use `flake8-pyproject` package instead of `pyproject-flake8`.
- `Pipfile` and `Pipfile.lock`: updated to use `flake8-pyproject` package instead of `pyproject-flake8`.
- `data`
- `/wbd/generate_pre_clip_fim_huc8.py`: Add space between (-) operator line 134.
- `write_parquet_from_calib_pts.py`: Add space between (-) operator line 234.
- `src`
- `check_huc_inputs.py`: Change `== string` to `is str`, remove `import string`
- `src_adjust_ras2fim_rating.py`: Fixed encoding error.
- `tools`
- `eval_plots.py`: Add space after comma in lines 207 & 208
- `generate_categorical_fim_mapping.py`: Use `is` instead of `==`, line 315
- `hash_compare.py`: Add space after comma, line 153.
- `inundate_mosaic_wrapper.py`: Use `is` instead of `==`, line 73.
- `inundation_wrapper_nwm_flows.py`: Use `is not` instead of `!=`, line 76.
- `mosaic_inundation.py`: Use `is` instead of `==`, line 181.
- `unit_tests`
- `clip_vectors_to_wbd_test.py`: File moved to data/wbd directory, update import statement
- `filter_catchments_and_add_attributes_params.json`: Update Branch ID
- `outputs_cleanup_params.json`: Update Branch ID
- `split_flows_params.json`: Update Branch ID
- `usgs_gage_crosswalk_params.json`: Update Branch ID & update argument to gage_crosswalk.run_crosswalk
- `unit_tests/usgs_gage_crosswalk_test.py`: Update params to gage_crosswalk.run_crosswalk

### Additions
- `.github/workflows/`
- `lint_and_format.yaml`: Add GitHub Actions Workflow file for Continuous Integration environment (lint and format test).

<br/><br/>

## v4.4.5.0 - 2023-10-26 - [PR#1018](https://github.com/NOAA-OWP/inundation-mapping/pull/1018)

During a recent BED attempt which added the new pre-clip system, it was erroring out on a number of hucs. It was issuing an error in the add_crosswalk.py script. While a minor bug does exist there, after a wide number of tests, the true culprit is the memory profile system embedded throughout FIM. This system has been around for at least a few years but not in use. It is not 100% clear why it became a problem with the addition of pre-clip, but that changes how records are loaded which likely affected memory at random times.
Expand Down
5 changes: 2 additions & 3 deletions src/check_huc_inputs.py
Original file line number Diff line number Diff line change
Expand Up @@ -4,7 +4,6 @@
import argparse
import os
import pathlib
import string
from glob import glob
from logging import exception

Expand Down Expand Up @@ -59,8 +58,8 @@ def __clean_huc_value(huc):

def __check_for_membership(hucs, accepted_hucs_set):
for huc in hucs:
if (type(huc) == string) and (not huc.isnumeric()):
msg = f"Huc value of {huc} does not appear to be a number."
if (type(huc) is str) and (not huc.isnumeric()):
msg = f"Huc value of {huc} does not appear to be a number. "
msg += "It could be an incorrect value but also could be that the huc list "
msg += "(if you used one), is not unix encoded."
raise KeyError(msg)
Expand Down
4 changes: 3 additions & 1 deletion src/src_adjust_ras2fim_rating.py
Original file line number Diff line number Diff line change
Expand Up @@ -52,7 +52,9 @@ def create_ras2fim_rating_database(ras_rc_filepath, ras_elev_df, nwm_recurr_file
print('Reading RAS2FIM rating curves from csv...')
log_text = 'Processing database for RAS2FIM flow/WSE at NWM flow recur intervals...\n'
col_filter = ["fid_xs", "flow", "wse"]
ras_rc_df = pd.read_csv(ras_rc_filepath, dtype={'fid_xs': object}, usecols=col_filter) # , nrows=30000)
ras_rc_df = pd.read_csv(
ras_rc_filepath, dtype={'fid_xs': object}, usecols=col_filter, encoding="unicode_escape"
) # , nrows=30000)
ras_rc_df.rename(columns={'fid_xs': 'location_id'}, inplace=True)
# ras_rc_df['location_id'] = ras_rc_df['feature_id'].astype(object)
print('Duration (read ras_rc_csv): {}'.format(dt.datetime.now() - start_time))
Expand Down
4 changes: 2 additions & 2 deletions tools/eval_plots.py
Original file line number Diff line number Diff line change
Expand Up @@ -204,8 +204,8 @@ def scatterplot(dataframe, x_field, y_field, title_text, stats_text=False, annot
axes.tick_params(labelsize='xx-large')

# Define y axis label and x axis label.
axes.set_ylabel(f'{y_field.replace("_"," ")}', fontsize='xx-large', weight='bold')
axes.set_xlabel(f'{x_field.replace("_"," ")}', fontsize='xx-large', weight='bold')
axes.set_ylabel(f'{y_field.replace("_", " ")}', fontsize='xx-large', weight='bold')
axes.set_xlabel(f'{x_field.replace("_", " ")}', fontsize='xx-large', weight='bold')

# Plot diagonal line
diag_range = [0, 1]
Expand Down
2 changes: 1 addition & 1 deletion tools/generate_categorical_fim_mapping.py
Original file line number Diff line number Diff line change
Expand Up @@ -312,7 +312,7 @@ def reformat_inundation_maps(
handle = os.path.split(extent_grid)[1].replace('.tif', '')
diss_extent_filename = os.path.join(gpkg_dir, f"{handle}_{huc}_dissolved.gpkg")
extent_poly_diss["geometry"] = [
MultiPolygon([feature]) if type(feature) == Polygon else feature
MultiPolygon([feature]) if type(feature) is Polygon else feature
for feature in extent_poly_diss["geometry"]
]

Expand Down
2 changes: 1 addition & 1 deletion tools/hash_compare.py
Original file line number Diff line number Diff line change
Expand Up @@ -150,7 +150,7 @@ def compare_gpkg(file1, file2, list_of_failed_files=[], verbose=False):
except AssertionError as e:
print(f"\n {str(e)} \n")
print(" The following files failed assert_geodataframe_equal: ")
print(f" -{file1.rsplit('/',1)[-1]} ")
print(f" {file1.rsplit('/', 1)[-1]} ")
list_of_failed_files.append(f1_gdf)


Expand Down
3 changes: 2 additions & 1 deletion tools/inundate_mosaic_wrapper.py
Original file line number Diff line number Diff line change
Expand Up @@ -70,8 +70,9 @@ def produce_mosaicked_inundation(
raise FileNotFoundError(errno.ENOENT, os.strerror(errno.ENOENT), hydrofabric_dir)

# If the "hucs" argument is really one huc, convert it to a list
if type(hucs) == str:
if type(hucs) is str:
hucs = [hucs]

# Check that huc folder exists in the hydrofabric_dir.
for huc in hucs:
if not os.path.exists(os.path.join(hydrofabric_dir, huc)):
Expand Down
2 changes: 1 addition & 1 deletion tools/inundation_wrapper_nwm_flows.py
Original file line number Diff line number Diff line change
Expand Up @@ -73,7 +73,7 @@ def run_recurr_test(fim_run_dir, branch_name, huc_id, magnitude, mask_type='huc'

# Check if magnitude is list of magnitudes or single value.
magnitude_list = magnitude
if type(magnitude_list) != list:
if type(magnitude_list) is not list:
magnitude_list = [magnitude_list]

for magnitude in magnitude_list:
Expand Down
2 changes: 1 addition & 1 deletion tools/mosaic_inundation.py
Original file line number Diff line number Diff line change
Expand Up @@ -178,7 +178,7 @@ def mosaic_final_inundation_extent_to_poly(inundation_raster, inundation_polygon
extent_poly = gpd.GeoDataFrame.from_features(list(results), crs=src.crs)
extent_poly_diss = extent_poly.dissolve(by="extent")
extent_poly_diss["geometry"] = [
MultiPolygon([feature]) if type(feature) == Polygon else feature
MultiPolygon([feature]) if type(feature) is Polygon else feature
for feature in extent_poly_diss["geometry"]
]

Expand Down
3 changes: 2 additions & 1 deletion unit_tests/clip_vectors_to_wbd_test.py
Original file line number Diff line number Diff line change
Expand Up @@ -4,10 +4,11 @@
import os
import unittest

import clip_vectors_to_wbd as src
import pytest
from unit_tests_utils import FIM_unit_test_helpers as ut_helpers

import data.wbd.clip_vectors_to_wbd as src


class test_clip_vectors_to_wbd(unittest.TestCase):

Expand Down
8 changes: 4 additions & 4 deletions unit_tests/filter_catchments_and_add_attributes_params.json
Original file line number Diff line number Diff line change
@@ -1,10 +1,10 @@
{
"valid_data": {
"outputDestDir": "/outputs/unit_test_data",
"input_catchments_filename": "/data/outputs/unit_test_data/02020005/branches/3246000006/gw_catchments_reaches_3246000006.gpkg",
"input_flows_filename": "/data/outputs/unit_test_data/02020005/branches/3246000006/demDerived_reaches_split_3246000006.gpkg",
"output_catchments_filename": "/data/outputs/unit_test_data/02020005/branches/3246000006/gw_catchments_reaches_filtered_addedAttributes_3246000006.gpkg",
"output_flows_filename": "/data/outputs/unit_test_data/02020005/branches/3246000006/demDerived_reaches_split_filtered_3246000006.gpkg",
"input_catchments_filename": "/data/outputs/unit_test_data/02020005/branches/2274000033/gw_catchments_reaches_2274000033.gpkg",
"input_flows_filename": "/data/outputs/unit_test_data/02020005/branches/2274000033/demDerived_reaches_split_2274000033.gpkg",
"output_catchments_filename": "/data/outputs/unit_test_data/02020005/branches/2274000033/gw_catchments_reaches_filtered_addedAttributes_2274000033.gpkg",
"output_flows_filename": "/data/outputs/unit_test_data/02020005/branches/2274000033/demDerived_reaches_split_filtered_2274000033.gpkg",
"wbd_filename": "/data/outputs/unit_test_data/02020005/wbd8_clp.gpkg",
"huc_code": "02020005"
}
Expand Down
4 changes: 2 additions & 2 deletions unit_tests/outputs_cleanup_params.json
Original file line number Diff line number Diff line change
@@ -1,8 +1,8 @@
{
"valid_specific_branch_data": {
"src_dir": "/data/outputs/unit_test_data/02020005/branches/3246000009",
"src_dir": "/data/outputs/unit_test_data/02020005/branches/2274000018",
"deny_list": "/foss_fim/config/deny_branches.lst",
"branch_id": "3246000009",
"branch_id": "2274000018",
"verbose": true
},
"valid_directory_data": {
Expand Down
8 changes: 4 additions & 4 deletions unit_tests/split_flows_params.json
Original file line number Diff line number Diff line change
Expand Up @@ -4,10 +4,10 @@
"max_length": 1500,
"slope_min": 0.001,
"lakes_buffer_input": 20,
"flows_filename": "/data/outputs/unit_test_data/02020005/branches/3246000005/demDerived_reaches_3246000005.shp",
"dem_filename": "/data/outputs/unit_test_data/02020005/branches/3246000005/dem_thalwegCond_3246000005.tif",
"split_flows_filename": "/data/outputs/unit_test_data/02020005/branches/3246000005/demDerived_reaches_split_3246000005.gpkg",
"split_points_filename": "/data/outputs/unit_test_data/02020005/branches/3246000005/demDerived_reaches_split_points_3246000005.gpkg",
"flows_filename": "/data/outputs/unit_test_data/02020005/branches/2274000031/demDerived_reaches_2274000031.shp",
"dem_filename": "/data/outputs/unit_test_data/02020005/branches/2274000031/dem_thalwegCond_2274000031.tif",
"split_flows_filename": "/data/outputs/unit_test_data/02020005/branches/2274000031/demDerived_reaches_split_2274000031.gpkg",
"split_points_filename": "/data/outputs/unit_test_data/02020005/branches/2274000031/demDerived_reaches_split_points_2274000031.gpkg",
"wbd8_clp_filename": "/data/outputs/unit_test_data/02020005/wbd8_clp.gpkg",
"lakes_filename": "/data/outputs/unit_test_data/02020005/nwm_lakes_proj_subset.gpkg",
"nwm_streams_filename": "/data/outputs/unit_test_data/02020005/nwm_subset_streams_levelPaths.gpkg"
Expand Down
13 changes: 7 additions & 6 deletions unit_tests/usgs_gage_crosswalk_params.json
Original file line number Diff line number Diff line change
@@ -1,11 +1,12 @@
{
"valid_data": {
"usgs_gages_filename": "/data/outputs/unit_test_data/02020005/usgs_subset_gages.gpkg",
"input_flows_filename": "/data/outputs/unit_test_data/02020005/branches/3246000005/demDerived_reaches_split_filtered_3246000005.gpkg",
"input_catchment_filename": "/data/outputs/unit_test_data/02020005/branches/3246000005/gw_catchments_reaches_filtered_addedAttributes_3246000005.gpkg",
"dem_filename": "/data/outputs/unit_test_data/02020005/branches/3246000005/dem_meters_3246000005.tif",
"dem_adj_filename": "/data/outputs/unit_test_data/02020005/branches/3246000005/dem_thalwegCond_3246000005.tif",
"output_table_filename": "/data/outputs/unit_test_data/02020005/branches/3246000005/usgs_elev_table.csv",
"branch_id": "3246000005"
"input_flows_filename": "/data/outputs/unit_test_data/02020005/branches/2274000028/demDerived_reaches_split_filtered_2274000028.gpkg",
"input_catchment_filename": "/data/outputs/unit_test_data/02020005/branches/2274000028/gw_catchments_reaches_filtered_addedAttributes_2274000028.gpkg",
"dem_filename": "/data/outputs/unit_test_data/02020005/branches/2274000028/dem_meters_2274000028.tif",
"dem_adj_filename": "/data/outputs/unit_test_data/02020005/branches/2274000028/dem_thalwegCond_2274000028.tif",
"output_table_filename": "/data/outputs/unit_test_data/02020005/branches/2274000028/usgs_elev_table.csv",
"output_directory": "/data/outputs/unit_test_data/02020005/branches/2274000028",
"branch_id": "2274000028"
}
}
2 changes: 1 addition & 1 deletion unit_tests/usgs_gage_crosswalk_test.py
Original file line number Diff line number Diff line change
Expand Up @@ -52,7 +52,7 @@ def test_GageCrosswalk_success(self):
params["input_flows_filename"],
params["dem_filename"],
params["dem_adj_filename"],
params["output_table_filename"],
params["output_directory"],
)

# Make sure that the usgs_elev_table.csv was written
Expand Down