We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Is your feature request related to a problem? Please describe. This issue described improvements to be done to L0A and L0B processing
Describe the solution you'd like
Check reader_kwargs delimiter is provided and warn otherwise.
delimiter
check df_sanitizer_fun has only lazy and df arguments
df_sanitizer_fun
Remove read_raw_data_zipped function and associated code (currently used for GPM campaigns)
read_raw_data_zipped
Enable saving integers columns to Parquet files. This requires:
FillValue
_FillValue
L0B_encodings.yml
raw_drop*
np.nan
In L0B processing, replace nan_flags from L0_data_format.yml with np.nan
nan_flags
L0_data_format.yml
Feature to drop dates based on issue/station_id.yml file ... .
issue/station_id.yml
In L0B processing, add variable_type (coordinate, count, category, flag, quantity, flux) attribute
variable_type
Enable reader development for stations where data are separated in two files. Example with Grenoble: raw.txt e matrix.txt
check_metadata_compliance strictly!
In L0B processing, check ThiesLPM and OTT_Parsivel raw_drop_number shape: (diameter, velocity) vs (velocity, diameter)
Decide whether to support dask.dataframe or use dask.delayed and save separate Parquets (more efficient)
If supporting dask dataframe, maybe optimize row_partition optmization
row_partition
Decide whether to modify L0B to save each netCDF separately and only add the end (optionally) open again all files, concat and write the full file.
The text was updated successfully, but these errors were encountered:
ghiggi
regislon
No branches or pull requests
Is your feature request related to a problem? Please describe.
This issue described improvements to be done to L0A and L0B processing
Describe the solution you'd like
Check reader_kwargs
delimiter
is provided and warn otherwise.check
df_sanitizer_fun
has only lazy and df argumentsRemove
read_raw_data_zipped
function and associated code (currently used for GPM campaigns)Enable saving integers columns to Parquet files. This requires:
FillValue
flag for integer columns (using_FillValue
ofL0B_encodings.yml
, except forraw_drop*
)np.nan
in L0B processing.In L0B processing, replace
nan_flags
fromL0_data_format.yml
withnp.nan
Feature to drop dates based on
issue/station_id.yml
file ... .In L0B processing, add
variable_type
(coordinate, count, category, flag, quantity, flux) attributeEnable reader development for stations where data are separated in two files. Example with Grenoble: raw.txt e matrix.txt
check_metadata_compliance strictly!
In L0B processing, check ThiesLPM and OTT_Parsivel raw_drop_number shape: (diameter, velocity) vs (velocity, diameter)
Decide whether to support dask.dataframe or use dask.delayed and save separate Parquets (more efficient)
If supporting dask dataframe, maybe optimize
row_partition
optmizationDecide whether to modify L0B to save each netCDF separately and only add the end (optionally) open again all files, concat and write the full file.
The text was updated successfully, but these errors were encountered: