Preprocessor functions#
- esmvalcore.preprocessor.DEFAULT_ORDER = ('fix_file', 'load', 'fix_metadata', 'concatenate', 'cmor_check_metadata', 'clip_timerange', 'fix_data', 'cmor_check_data', 'add_supplementary_variables', 'derive', 'extract_time', 'extract_season', 'extract_month', 'resample_hours', 'resample_time', 'extract_levels', 'weighting_landsea_fraction', 'mask_landsea', 'mask_glaciated', 'mask_landseaice', 'regrid', 'extract_coordinate_points', 'extract_point', 'extract_location', 'mask_multimodel', 'mask_fillvalues', 'mask_above_threshold', 'mask_below_threshold', 'mask_inside_range', 'mask_outside_range', 'clip', 'rolling_window_statistics', 'extract_region', 'extract_shape', 'extract_volume', 'extract_trajectory', 'extract_transect', 'detrend', 'extract_named_regions', 'axis_statistics', 'depth_integration', 'area_statistics', 'volume_statistics', 'local_solar_time', 'amplitude', 'zonal_statistics', 'meridional_statistics', 'accumulate_coordinate', 'hourly_statistics', 'daily_statistics', 'monthly_statistics', 'seasonal_statistics', 'annual_statistics', 'decadal_statistics', 'climate_statistics', 'anomalies', 'regrid_time', 'timeseries_filter', 'linear_trend', 'linear_trend_stderr', 'convert_units', 'ensemble_statistics', 'multi_model_statistics', 'bias', 'remove_supplementary_variables', 'save')#
By default, preprocessor functions are applied in this order.
Preprocessor module.
Functions:
|
Weight data using the bounds from a given coordinate. |
|
Add ancillary variables and/or cell measures. |
|
Calculate amplitude of cycles by aggregating over coordinates. |
|
Compute annual statistics. |
|
Compute anomalies using a mean with the specified granularity. |
|
Apply a statistical operator in the horizontal plane. |
|
Perform statistics along a given axis. |
|
Calculate biases relative to a reference dataset. |
|
Compute climate statistics with the specified granularity. |
|
Clip values at a specified minimum and/or maximum value. |
|
Extract time range with a resolution up to seconds. |
|
Check if data conforms to variable's CMOR definition. |
|
Check if metadata conforms to variable's CMOR definition. |
|
Concatenate all cubes after fixing metadata. |
|
Convert the units of a cube to new ones. |
|
Compute daily statistics. |
|
Compute decadal statistics. |
|
Determine the total sum over the vertical component. |
|
Derive variable. |
|
Detrend data along a given dimension. |
|
Compute ensemble statistics. |
|
Extract points from any coordinate with interpolation. |
|
Perform vertical interpolation. |
|
Extract a point using a location name, with interpolation. |
|
Slice cube to get only the data belonging to a specific month. |
|
Extract a specific named region. |
|
Extract a point, with interpolation. |
|
Extract a region from a cube. |
|
Slice cube to get only the data belonging to a specific season. |
|
Extract a region defined by a shapefile using masking. |
|
Extract a time range from a cube. |
|
Extract data along a trajectory. |
|
Extract data along a line of constant latitude or longitude. |
|
Subset a cube based on a range of values in the z-coordinate. |
|
Fix cube data if fixes are required. |
|
Fix files before ESMValTool can load them. |
|
Fix cube metadata if fixes are required. |
|
Compute hourly statistics. |
|
Calculate linear trend of data along a given coordinate. |
|
Calculate standard error of linear trend along a given coordinate. |
|
Load iris cubes from string or Path objects. |
|
Convert UTC time coordinate to local solar time (LST). |
|
Mask above a specific threshold value. |
|
Mask below a specific threshold value. |
|
Compute and apply a multi-dataset fillvalues mask. |
|
Mask out glaciated areas. |
|
Mask inside a specific threshold range. |
|
Mask out either land mass or sea (oceans, seas and lakes). |
|
Mask out either landsea (combined) or ice. |
|
Apply common mask to all datasets (using logical OR). |
|
Mask outside a specific threshold range. |
|
Compute meridional statistics. |
|
Compute monthly statistics. |
|
Compute multi-model statistics. |
|
Perform horizontal regridding. |
|
Align time axis for cubes so they can be subtracted. |
Remove supplementary variables. |
|
|
Convert x-hourly data to y-hourly by eliminating extra timesteps. |
|
Change frequency of data by resampling it. |
|
Compute rolling-window statistics over a coordinate. |
|
Save iris cubes to file. |
|
Compute seasonal statistics. |
|
Apply a timeseries filter. |
|
Apply a statistical operation over a volume. |
|
Weight fields using land or sea fraction. |
|
Compute zonal statistics. |
- esmvalcore.preprocessor.accumulate_coordinate(cube: Cube, coordinate: str | DimCoord | AuxCoord) Cube [source]#
Weight data using the bounds from a given coordinate.
The resulting cube will then have units given by
cube_units * coordinate_units
.- Parameters:
- Returns:
Cube with the aggregated data
- Return type:
- Raises:
ValueError – If the coordinate is not found in the cube.
NotImplementedError – If the coordinate is multidimensional.
- esmvalcore.preprocessor.add_supplementary_variables(cube: Cube, supplementary_cubes: Iterable[Cube]) Cube [source]#
Add ancillary variables and/or cell measures.
- Parameters:
- Returns:
Cube with added ancillary variables and/or cell measures.
- Return type:
- esmvalcore.preprocessor.amplitude(cube, coords)[source]#
Calculate amplitude of cycles by aggregating over coordinates.
Note
The amplitude is calculated as peak-to-peak amplitude (difference between maximum and minimum value of the signal). Other amplitude types are currently not supported.
- Parameters:
cube (iris.cube.Cube) – Input data.
coords (str or list of str) – Coordinates over which is aggregated. For example, use
'year'
to extract the annual cycle amplitude for each year in the data or['day_of_year', 'year']
to extract the diurnal cycle amplitude for each individual day in the data. If the coordinates are not found incube
, try to add it viairis.coord_categorisation
(at the moment, this only works for the temporal coordinatesday_of_month
,day_of_year
,hour
,month
,month_fullname
,month_number
,season
,season_number
,season_year
,weekday
,weekday_fullname
,weekday_number
oryear
.
- Returns:
Amplitudes.
- Return type:
- Raises:
iris.exceptions.CoordinateNotFoundError – A coordinate is not found in
cube
and cannot be added viairis.coord_categorisation
.
- esmvalcore.preprocessor.annual_statistics(cube: Cube, operator: str = 'mean', **operator_kwargs) Cube [source]#
Compute annual statistics.
Note that this function does not weight the annual mean if uneven time periods are present. Ie, all data inside the year are treated equally.
- Parameters:
cube (Cube) – Input cube.
operator (str) – The operation. Used to determine the
iris.analysis.Aggregator
object used to calculate the statistics. Allowed options are given in this table.**operator_kwargs – Optional keyword arguments for the
iris.analysis.Aggregator
object defined by operator.
- Returns:
Annual statistics cube.
- Return type:
- esmvalcore.preprocessor.anomalies(cube: Cube, period: str, reference: dict | None = None, standardize: bool = False, seasons: Iterable[str] = ('DJF', 'MAM', 'JJA', 'SON')) Cube [source]#
Compute anomalies using a mean with the specified granularity.
Computes anomalies based on hourly, daily, monthly, seasonal or yearly means for the full available period.
- Parameters:
cube (Cube) – Input cube.
period (str) – Period to compute the statistic over. Available periods: full, season, seasonal, monthly, month, mon, daily, day, hourly, hour, hr.
reference (optional) – Period of time to use a reference, as needed for the
extract_time()
preprocessor function. IfNone
, all available data is used as a reference.standardize (optional) – If
True
standardized anomalies are calculated.seasons (optional) – Seasons to use if needed. Defaults to (‘DJF’, ‘MAM’, ‘JJA’, ‘SON’).
- Returns:
Anomalies cube.
- Return type:
- esmvalcore.preprocessor.area_statistics(cube: Cube, operator: str, normalize: Literal['subtract', 'divide'] | None = None, **operator_kwargs) Cube [source]#
Apply a statistical operator in the horizontal plane.
We assume that the horizontal directions are [‘longitude’, ‘latitude’].
This table shows a list of supported operators. All operators that support weights are by default weighted with the grid cell areas. Note that for area-weighted sums, the units of the resulting cube will be multiplied by m \(^2\).
- Parameters:
cube (Cube) – Input cube. The input cube should have a
iris.coords.CellMeasure
named'cell_area'
, unless it has regular 1D latitude and longitude coordinates so the cell areas can be computed usingiris.analysis.cartography.area_weights()
.operator (str) – The operation. Used to determine the
iris.analysis.Aggregator
object used to calculate the statistics. Allowed options are given in this table.normalize (Literal['subtract', 'divide'] | None) – If given, do not return the statistics cube itself, but rather, the input cube, normalized with the statistics cube. Can either be subtract (statistics cube is subtracted from the input cube) or divide (input cube is divided by the statistics cube).
**operator_kwargs – Optional keyword arguments for the
iris.analysis.Aggregator
object defined by operator.
- Returns:
Collapsed cube.
- Return type:
- Raises:
iris.exceptions.CoordinateMultiDimError – Cube has irregular or unstructured grid but supplementary variable cell_area is not available.
- esmvalcore.preprocessor.axis_statistics(cube: Cube, axis: str, operator: str, normalize: Literal['subtract', 'divide'] | None = None, **operator_kwargs) Cube [source]#
Perform statistics along a given axis.
Operates over an axis direction.
Note
The mean, sum and rms operations are weighted by the corresponding coordinate bounds by default. For sum, the units of the resulting cube will be multiplied by corresponding coordinate units.
- Parameters:
cube (Cube) – Input cube.
axis (str) – Direction over where to apply the operator. Possible values are x, y, z, t.
operator (str) – The operation. Used to determine the
iris.analysis.Aggregator
object used to calculate the statistics. Allowed options are given in this table.normalize (Literal['subtract', 'divide'] | None) – If given, do not return the statistics cube itself, but rather, the input cube, normalized with the statistics cube. Can either be subtract (statistics cube is subtracted from the input cube) or divide (input cube is divided by the statistics cube).
**operator_kwargs – Optional keyword arguments for the
iris.analysis.Aggregator
object defined by operator.
- Returns:
Collapsed cube.
- Return type:
- esmvalcore.preprocessor.bias(products: set[PreprocessorFile] | Iterable[Cube], ref_cube: Cube | None = None, bias_type: BiasType = 'absolute', denominator_mask_threshold: float = 0.001, keep_reference_dataset: bool = False) set[PreprocessorFile] | CubeList [source]#
Calculate biases relative to a reference dataset.
The reference dataset needs to be broadcastable to all input products. This supports iris’ rich broadcasting abilities. To ensure this, the preprocessors
esmvalcore.preprocessor.regrid()
and/oresmvalcore.preprocessor.regrid_time()
might be helpful.Notes
The reference dataset can be specified with the ref_cube argument. If ref_cube is
None
, exactly one input dataset in the products set needs to have the facetreference_for_bias: true
defined in the recipe. Please do not specify the option ref_cube when using this preprocessor function in a recipe.- Parameters:
products (set[PreprocessorFile] | Iterable[Cube]) – Input datasets/cubes for which the bias is calculated relative to a reference dataset/cube.
ref_cube (Optional[Cube]) – Cube which is used as reference for the bias calculation. If
None
, products needs to be aset
of ~esmvalcore.preprocessor.PreprocessorFile objects and exactly one dataset in products needs the facetreference_for_bias: true
.bias_type (BiasType) – Bias type that is calculated. Must be one of
'absolute'
(dataset - ref) or'relative'
((dataset - ref) / ref).denominator_mask_threshold (float) – Threshold to mask values close to zero in the denominator (i.e., the reference dataset) during the calculation of relative biases. All values in the reference dataset with absolute value less than the given threshold are masked out. This setting is ignored when
bias_type
is set to'absolute'
. Please note that for some variables with very small absolute values (e.g., carbon cycle fluxes, which are usually \(< 10^{-6}\) kg m \(^{-2}\) s \(^{-1}\)) it is absolutely essential to change the default value in order to get reasonable results.keep_reference_dataset (bool) – If
True
, keep the reference dataset in the output. IfFalse
, drop the reference dataset. Ignored if ref_cube is given.
- Returns:
Output datasets/cubes. Will be a
set
ofPreprocessorFile
objects if products is also one, aCubeList
otherwise.- Return type:
set[PreprocessorFile] | CubeList
- Raises:
ValueError – Not exactly one input datasets contains the facet
reference_for_bias: true
ifref_cube=None
;ref_cube=None
and the input products are given as iterable ofCube
objects;bias_type
is not one of'absolute'
or'relative'
.
- esmvalcore.preprocessor.climate_statistics(cube: Cube, operator: str = 'mean', period: str = 'full', seasons: Iterable[str] = ('DJF', 'MAM', 'JJA', 'SON'), **operator_kwargs) Cube [source]#
Compute climate statistics with the specified granularity.
Computes statistics for the whole dataset. It is possible to get them for the full period or with the data grouped by hour, day, month or season.
Note
The mean, sum and rms operations over the full period are weighted by the time coordinate, i.e., the length of the time intervals. For sum, the units of the resulting cube will be multiplied by corresponding time units (e.g., days).
- Parameters:
cube (Cube) – Input cube.
operator (str) – The operation. Used to determine the
iris.analysis.Aggregator
object used to calculate the statistics. Allowed options are given in this table.period (str) – Period to compute the statistic over. Available periods: full, season, seasonal, monthly, month, mon, daily, day, hourly, hour, hr.
seasons (Iterable[str]) – Seasons to use if needed. Defaults to (‘DJF’, ‘MAM’, ‘JJA’, ‘SON’).
**operator_kwargs – Optional keyword arguments for the
iris.analysis.Aggregator
object defined by operator.
- Returns:
Climate statistics cube.
- Return type:
- esmvalcore.preprocessor.clip(cube, minimum=None, maximum=None)[source]#
Clip values at a specified minimum and/or maximum value.
Values lower than minimum are set to minimum and values higher than maximum are set to maximum.
- Parameters:
cube (iris.cube.Cube) – iris cube to be clipped
minimum (float) – lower threshold to be applied on input cube data.
maximum (float) – upper threshold to be applied on input cube data.
- Returns:
clipped cube.
- Return type:
- esmvalcore.preprocessor.clip_timerange(cube: Cube, timerange: str) Cube [source]#
Extract time range with a resolution up to seconds.
- Parameters:
- Returns:
Sliced cube.
- Return type:
- Raises:
ValueError – Time ranges are outside the cube’s time limits.
- esmvalcore.preprocessor.cmor_check_data(cube: Cube, cmor_table: str, mip: str, short_name: str, frequency: str | None = None, check_level: CheckLevels = CheckLevels.DEFAULT) Cube [source]#
Check if data conforms to variable’s CMOR definition.
- Parameters:
cube (Cube) – Data cube to check.
cmor_table (str) – CMOR definitions to use (i.e., the variable’s project).
mip (str) – Variable’s MIP.
short_name (str) – Variable’s short name
frequency (str | None) – Data frequency. If not given, use the one from the CMOR table of the variable.
check_level (CheckLevels) – Level of strictness of the checks.
- Returns:
Checked cube.
- Return type:
- esmvalcore.preprocessor.cmor_check_metadata(cube: Cube, cmor_table: str, mip: str, short_name: str, frequency: str | None = None, check_level: CheckLevels = CheckLevels.DEFAULT) Cube [source]#
Check if metadata conforms to variable’s CMOR definition.
None of the checks at this step will force the cube to load the data.
- Parameters:
cube (Cube) – Data cube to check.
cmor_table (str) – CMOR definitions to use (i.e., the variable’s project).
mip (str) – Variable’s MIP.
short_name (str) – Variable’s short name.
frequency (str | None) – Data frequency. If not given, use the one from the CMOR table of the variable.
check_level (CheckLevels) – Level of strictness of the checks.
- Returns:
Checked cube.
- Return type:
- esmvalcore.preprocessor.concatenate(cubes, check_level=CheckLevels.DEFAULT)[source]#
Concatenate all cubes after fixing metadata.
- Parameters:
cubes (iterable of iris.cube.Cube) – Data cubes to be concatenated
check_level (CheckLevels) – Level of strictness of the checks in the concatenation.
- Returns:
cube – Resulting concatenated cube.
- Return type:
- Raises:
ValueError – Concatenation was not possible.
- esmvalcore.preprocessor.convert_units(cube, units)[source]#
Convert the units of a cube to new ones.
This converts units of a cube.
Note
Allows special unit conversions which transforms one quantity to another (physically related) quantity. These quantities are identified via their
standard_name
and theirunits
(units convertible to the ones defined are also supported). For example, this enables conversions between precipitation fluxes measured inkg m-2 s-1
and precipitation rates measured inmm day-1
(and vice versa).Currently, the following special conversions are supported:
precipitation_flux
(kg m-2 s-1
) –lwe_precipitation_rate
(mm day-1
)equivalent_thickness_at_stp_of_atmosphere_ozone_content
(m
) –equivalent_thickness_at_stp_of_atmosphere_ozone_content
(DU
)
Names in the list correspond to
standard_names
of the input data. Conversions are allowed from each quantity to any other quantity given in a bullet point. The corresponding target quantity is inferred from the desired target units. In addition, any other units convertible to the ones given are also supported (e.g., instead ofmm day-1
,m s-1
is also supported).Note that for precipitation variables, a water density of
1000 kg m-3
is assumed.- Parameters:
cube (iris.cube.Cube) – Input cube.
units (str) – New units in udunits form.
- Returns:
converted cube.
- Return type:
- esmvalcore.preprocessor.daily_statistics(cube: Cube, operator: str = 'mean', **operator_kwargs) Cube [source]#
Compute daily statistics.
Chunks time in daily periods and computes statistics over them;
- Parameters:
cube (Cube) – Input cube.
operator (str) – The operation. Used to determine the
iris.analysis.Aggregator
object used to calculate the statistics. Allowed options are given in this table.**operator_kwargs – Optional keyword arguments for the
iris.analysis.Aggregator
object defined by operator.
- Returns:
Daily statistics cube.
- Return type:
- esmvalcore.preprocessor.decadal_statistics(cube: Cube, operator: str = 'mean', **operator_kwargs) Cube [source]#
Compute decadal statistics.
Note that this function does not weight the decadal mean if uneven time periods are present. Ie, all data inside the decade are treated equally.
- Parameters:
cube (Cube) – Input cube.
operator (str) – The operation. Used to determine the
iris.analysis.Aggregator
object used to calculate the statistics. Allowed options are given in this table.**operator_kwargs – Optional keyword arguments for the
iris.analysis.Aggregator
object defined by operator.
- Returns:
Decadal statistics cube.
- Return type:
- esmvalcore.preprocessor.depth_integration(cube: Cube) Cube [source]#
Determine the total sum over the vertical component.
Requires a 3D cube. The z-coordinate integration is calculated by taking the sum in the z direction of the cell contents multiplied by the cell thickness. The units of the resulting cube are multiplied by the z-coordinate units.
- Parameters:
cube (Cube) – Input cube.
- Returns:
Collapsed cube.
- Return type:
- esmvalcore.preprocessor.derive(cubes, short_name, long_name, units, standard_name=None)[source]#
Derive variable.
- Parameters:
cubes (iris.cube.CubeList) – Includes all the needed variables for derivation defined in
get_required()
.short_name (str) – short_name
long_name (str) – long_name
units (str) – units
standard_name (str, optional) – standard_name
- Returns:
The new derived variable.
- Return type:
- esmvalcore.preprocessor.detrend(cube, dimension='time', method='linear')[source]#
Detrend data along a given dimension.
- Parameters:
cube (iris.cube.Cube) – input cube.
dimension (str) – Dimension to detrend
method (str) – Method to detrend. Available: linear, constant. See documentation of ‘scipy.signal.detrend’ for details
- Returns:
Detrended cube
- Return type:
- esmvalcore.preprocessor.ensemble_statistics(products: set[PreprocessorFile] | Iterable[Cube], statistics: list[str | dict], output_products, span: str = 'overlap', ignore_scalar_coords: bool = False) dict | set [source]#
Compute ensemble statistics.
An ensemble grouping is performed on the input products (using the ensemble facet of input datasets). The statistics are then computed calling
esmvalcore.preprocessor.multi_model_statistics()
with appropriate groups.- Parameters:
products (set[PreprocessorFile] | Iterable[Cube]) – Cubes (or products) over which the statistics will be computed.
statistics (list[str | dict]) – Statistical operations to be computed, e.g.,
['mean', 'median']
. For some statistics like percentiles, it is also possible to pass additional keyword arguments, e.g.,[{'operator': 'mean', 'percent': 20}]
. All supported options are are given in this table.output_products (dict) – For internal use only. A dict with statistics names as keys and preprocessorfiles as values. If products are passed as input, the statistics cubes will be assigned to these output products.
span (str) – Overlap or full; if overlap, statitstics are computed on common time- span; if full, statistics are computed on full time spans, ignoring missing data.
ignore_scalar_coords (bool) – If True, remove any scalar coordinate in the input datasets before merging the input cubes into the multi-dataset cube. The resulting multi-dataset cube will have no scalar coordinates (the actual input datasets will remain unchanged). If False, scalar coordinates will remain in the input datasets, which might lead to merge conflicts in case the input datasets have different scalar coordinates.
- Returns:
A
dict
of cubes orset
of output_products depending on the type of products.- Return type:
See also
- esmvalcore.preprocessor.extract_coordinate_points(cube, definition, scheme)[source]#
Extract points from any coordinate with interpolation.
Multiple points can also be extracted, by supplying an array of coordinates. The resulting point cube will match the respective coordinates to those of the input coordinates. If the input coordinate is a scalar, the dimension will be a scalar in the output cube.
- Parameters:
- Returns:
Returns a cube with the extracted point(s), and with adjusted latitude and longitude coordinates (see above). If desired point outside values for at least one coordinate, this cube will have fully masked data.
- Return type:
- Raises:
ValueError: – If the interpolation scheme is not provided or is not recognised.
- esmvalcore.preprocessor.extract_levels(cube: Cube, levels: _SupportsArray[dtype[Any]] | _NestedSequence[_SupportsArray[dtype[Any]]] | bool | int | float | complex | str | bytes | _NestedSequence[bool | int | float | complex | str | bytes] | Array, scheme: str, coordinate: str | None = None, rtol: float = 1e-07, atol: float | None = None)[source]#
Perform vertical interpolation.
- Parameters:
cube (Cube) – The source cube to be vertically interpolated.
levels (_SupportsArray[dtype[Any]] | _NestedSequence[_SupportsArray[dtype[Any]]] | bool | int | float | complex | str | bytes | _NestedSequence[bool | int | float | complex | str | bytes] | Array) – One or more target levels for the vertical interpolation. Assumed to be in the same S.I. units of the source cube vertical dimension coordinate. If the requested levels are sufficiently close to the levels of the cube, cube slicing will take place instead of interpolation.
scheme (str) – The vertical interpolation scheme to use. Choose from ‘linear’, ‘nearest’, ‘linear_extrapolate’, ‘nearest_extrapolate’.
coordinate (str | None) – The coordinate to interpolate. If specified, pressure levels (if present) can be converted to height levels and vice versa using the US standard atmosphere. E.g. ‘coordinate = altitude’ will convert existing pressure levels (air_pressure) to height levels (altitude); ‘coordinate = air_pressure’ will convert existing height levels (altitude) to pressure levels (air_pressure).
rtol (float) – Relative tolerance for comparing the levels in cube to the requested levels. If the levels are sufficiently close, the requested levels will be assigned to the cube and no interpolation will take place.
atol (float | None) – Absolute tolerance for comparing the levels in cube to the requested levels. If the levels are sufficiently close, the requested levels will be assigned to the cube and no interpolation will take place. By default, atol will be set to 10^-7 times the mean value of the levels on the cube.
- Returns:
A cube with the requested vertical levels.
- Return type:
See also
regrid
Perform horizontal regridding.
- esmvalcore.preprocessor.extract_location(cube, location, scheme)[source]#
Extract a point using a location name, with interpolation.
Extracts a single location point from a cube, according to the interpolation scheme
scheme
.The function just retrieves the coordinates of the location and then calls the
extract_point
preprocessor.It can be used to locate cities and villages, but also mountains or other geographical locations.
Note
The geolocator needs a working internet connection.
- Parameters:
- Returns:
Returns a cube with the extracted point, and with adjusted
latitude and longitude coordinates.
- Raises:
ValueError: – If location is not supplied as a preprocessor parameter.
ValueError: – If scheme is not supplied as a preprocessor parameter.
ValueError: – If given location cannot be found by the geolocator.
- esmvalcore.preprocessor.extract_month(cube: Cube, month: int) Cube [source]#
Slice cube to get only the data belonging to a specific month.
- Parameters:
- Returns:
Cube for specified month.
- Return type:
- Raises:
ValueError – Requested month is not present in the cube.
- esmvalcore.preprocessor.extract_named_regions(cube: Cube, regions: str | Iterable[str]) Cube [source]#
Extract a specific named region.
The region coordinate exist in certain CMIP datasets. This preprocessor allows a specific named regions to be extracted.
- Parameters:
- Returns:
Smaller cube.
- Return type:
- Raises:
ValueError – regions is not list or tuple or set.
ValueError – region not included in cube.
- esmvalcore.preprocessor.extract_point(cube, latitude, longitude, scheme)[source]#
Extract a point, with interpolation.
Extracts a single latitude/longitude point from a cube, according to the interpolation scheme scheme.
Multiple points can also be extracted, by supplying an array of latitude and/or longitude coordinates. The resulting point cube will match the respective latitude and longitude coordinate to those of the input coordinates. If the input coordinate is a scalar, the dimension will be missing in the output cube (that is, it will be a scalar).
If the point to be extracted has at least one of the coordinate point values outside the interval of the cube’s same coordinate values, then no extrapolation will be performed, and the resulting extracted cube will have fully masked data.
- Parameters:
- Returns:
Returns a cube with the extracted point(s), and with adjusted latitude and longitude coordinates (see above). If desired point outside values for at least one coordinate, this cube will have fully masked data.
- Return type:
- Raises:
ValueError: – If the interpolation scheme is None or unrecognized.
Examples
With a cube that has the coordinates
latitude: [1, 2, 3, 4]
longitude: [1, 2, 3, 4]
- data values: [[[1, 2, 3, 4], [5, 6, …], […], […],
… ]]]
>>> point = extract_point(cube, 2.5, 2.5, 'linear') >>> point.data array([ 8.5, 24.5, 40.5, 56.5])
Extraction of multiple points at once, with a nearest matching scheme. The values for 0.1 will result in masked values, since this lies outside the cube grid.
>>> point = extract_point(cube, [1.4, 2.1], [0.1, 1.1], ... 'nearest') >>> point.data.shape (4, 2, 2) >>> # x, y, z indices of masked values >>> np.where(~point.data.mask) (array([0, 0, 1, 1, 2, 2, 3, 3]), array([0, 1, 0, 1, 0, 1, 0, 1]), array([1, 1, 1, 1, 1, 1, 1, 1])) >>> point.data[~point.data.mask].data array([ 1, 5, 17, 21, 33, 37, 49, 53])
- esmvalcore.preprocessor.extract_region(cube: Cube, start_longitude: float, end_longitude: float, start_latitude: float, end_latitude: float) Cube [source]#
Extract a region from a cube.
Function that subsets a cube on a box (start_longitude, end_longitude, start_latitude, end_latitude).
- Parameters:
- Returns:
Smaller cube.
- Return type:
- esmvalcore.preprocessor.extract_season(cube: Cube, season: str) Cube [source]#
Slice cube to get only the data belonging to a specific season.
- Parameters:
- Returns:
data cube for specified season.
- Return type:
- Raises:
ValueError – Requested season is not present in the cube.
- esmvalcore.preprocessor.extract_shape(cube: Cube, shapefile: str | Path, method: str = 'contains', crop: bool = True, decomposed: bool = False, ids: list | dict | None = None) Cube [source]#
Extract a region defined by a shapefile using masking.
Note that this function does not work for shapes crossing the prime meridian or poles.
- Parameters:
cube (Cube) – Input cube.
A shapefile defining the region(s) to extract. Also accepts the following strings to load special shapefiles:
'ar6'
: IPCC WG1 reference regions (v4) used in Assessment Report 6 (https://doi.org/10.5281/zenodo.5176260). Should be used in combination with adict
for the argument ids, e.g.,ids={'Acronym': ['GIC', 'WNA']}
.
method (str) – Select all points contained by the shape or select a single representative point. Choose either ‘contains’ or ‘representative’. If ‘contains’ is used, but not a single grid point is contained by the shape, a representative point will be selected.
crop (bool) – In addition to masking, crop the resulting cube using
extract_region()
. Data on irregular grids will not be cropped.decomposed (bool) – If set to True, the output cube will have an additional dimension shape_id describing the requested regions.
Shapes to be read from the shapefile. Can be given as:
list
: IDs are assigned from the attributes name, NAME, Name, id, or ID (in that priority order; the first one available is used). If none of these attributes are available in the shapefile, assume that the given ids correspond to the reading order of the individual shapes, e.g.,ids=[0, 2]
corresponds to the first and third shape read from the shapefile. Note: An empty list is interpreted as ids=None.dict
: IDs (dictionary value;list
ofstr
) are assigned from attribute given as dictionary key (str
). Only dictionaries with length 1 are supported. Example:ids={'Acronym': ['GIC', 'WNA']}
forshapefile='ar6'
.None: select all available shapes from the shapefile.
- Returns:
Cube containing the extracted region.
- Return type:
See also
extract_region
Extract a region from a cube.
- esmvalcore.preprocessor.extract_time(cube: Cube, start_year: int, start_month: int, start_day: int, end_year: int, end_month: int, end_day: int) Cube [source]#
Extract a time range from a cube.
Given a time range passed in as a series of years, months and days, it returns a time-extracted cube with data only within the specified time range.
- Parameters:
- Returns:
Sliced cube.
- Return type:
- Raises:
ValueError – Time ranges are outside the cube time limits.
- esmvalcore.preprocessor.extract_trajectory(cube: Cube, latitudes: Sequence[float], longitudes: Sequence[float], number_points: int = 2) Cube [source]#
Extract data along a trajectory.
latitudes and longitudes are the pairs of coordinates for two points. number_points is the number of points between the two points.
This version uses the expensive interpolate method, but it may be necceasiry for irregular grids.
If only two latitude and longitude coordinates are given, extract_trajectory will produce a cube will extrapolate along a line between those two points, and will add number_points points between the two corners.
If more than two points are provided, then extract_trajectory will produce a cube which has extrapolated the data of the cube to those points, and number_points is not needed.
- Parameters:
- Returns:
Collapsed cube.
- Return type:
- Raises:
ValueError – Latitude and longitude have different dimensions.
- esmvalcore.preprocessor.extract_transect(cube: Cube, latitude: None | float | Iterable[float] = None, longitude: None | float | Iterable[float] = None) Cube [source]#
Extract data along a line of constant latitude or longitude.
Both arguments, latitude and longitude, are treated identically. Either argument can be a single float, or a pair of floats, or can be left empty. The single float indicates the latitude or longitude along which the transect should be extracted. A pair of floats indicate the range that the transect should be extracted along the secondairy axis.
For instance ‘extract_transect(cube, longitude=-28)’ will produce a transect along 28 West.
Also, ‘extract_transect(cube, longitude=-28, latitude=[-50, 50])’ will produce a transect along 28 West between 50 south and 50 North.
This function is not yet implemented for irregular arrays - instead try the extract_trajectory function, but note that it is currently very slow. Alternatively, use the regrid preprocessor to regrid along a regular grid and then extract the transect.
- Parameters:
cube (Cube) – Input cube.
latitude (optional) – Transect latitude or range.
longitude (optional) – Transect longitude or range.
- Returns:
Collapsed cube.
- Return type:
- Raises:
ValueError – Slice extraction not implemented for irregular grids.
ValueError – Latitude and longitude are both floats or lists; not allowed to slice on both axes at the same time.
- esmvalcore.preprocessor.extract_volume(cube: Cube, z_min: float, z_max: float, interval_bounds: str = 'open', nearest_value: bool = False) Cube [source]#
Subset a cube based on a range of values in the z-coordinate.
Function that subsets a cube on a box of (z_min, z_max), (z_min, z_max], [z_min, z_max) or [z_min, z_max] Note that this requires the requested z-coordinate range to be the same sign as the iris cube. ie, if the cube has z-coordinate as negative, then z_min and z_max need to be negative numbers. If nearest_value is set to False, the extraction will be performed with the given z_min and z_max values. If nearest_value is set to True, the cube extraction will be performed taking into account the z_coord values that are closest to the z_min and z_max values.
- Parameters:
cube (Cube) – Input cube.
z_min (float) – Minimum depth to extract.
z_max (float) – Maximum depth to extract.
interval_bounds (str) – Sets left bound of the interval to either ‘open’, ‘closed’, ‘left_closed’ or ‘right_closed’.
nearest_value (bool) – Extracts considering the nearest value of z-coord to z_min and z_max.
- Returns:
z-coord extracted cube.
- Return type:
- esmvalcore.preprocessor.fix_data(cube: Cube, short_name: str, project: str, dataset: str, mip: str, frequency: str | None = None, check_level: CheckLevels = CheckLevels.DEFAULT, session: Session | None = None, **extra_facets) Cube [source]#
Fix cube data if fixes are required.
This method assumes that metadata is already fixed and checked.
This method collects all the relevant fixes (including generic ones) for a given variable and applies them.
- Parameters:
cube (Cube) – Cube to fix.
short_name (str) – Variable’s short name.
project (str) – Project of the dataset.
dataset (str) – Name of the dataset.
mip (str) – Variable’s MIP.
frequency (Optional[str]) – Variable’s data frequency, if available.
check_level (CheckLevels) –
Level of strictness of the checks.
Deprecated since version 2.10.0: This option has been deprecated in ESMValCore version 2.10.0 and is scheduled for removal in version 2.12.0. Please use the functions
cmor_check_metadata()
,cmor_check_data()
, orcmor_check()
instead. This function will no longer perform CMOR checks. Fixes and CMOR checks have been clearly separated in ESMValCore version 2.10.0.session (Optional[Session]) – Current session which includes configuration and directory information.
**extra_facets – Extra facets are mainly used for data outside of the big projects like CMIP, CORDEX, obs4MIPs. For details, see Extra Facets.
- Returns:
Fixed cube.
- Return type:
- esmvalcore.preprocessor.fix_file(file: Path, short_name: str, project: str, dataset: str, mip: str, output_dir: Path, add_unique_suffix: bool = False, session: Session | None = None, frequency: str | None = None, **extra_facets) str | Path [source]#
Fix files before ESMValTool can load them.
These fixes are only for issues that prevent iris from loading the cube or that cannot be fixed after the cube is loaded.
Original files are not overwritten.
- Parameters:
file (Path) – Path to the original file.
short_name (str) – Variable’s short name.
project (str) – Project of the dataset.
dataset (str) – Name of the dataset.
mip (str) – Variable’s MIP.
output_dir (Path) – Output directory for fixed files.
add_unique_suffix (bool) – Adds a unique suffix to output_dir for thread safety.
session (Optional[Session]) – Current session which includes configuration and directory information.
frequency (Optional[str]) – Variable’s data frequency, if available.
**extra_facets – Extra facets are mainly used for data outside of the big projects like CMIP, CORDEX, obs4MIPs. For details, see Extra Facets.
- Returns:
Path to the fixed file.
- Return type:
str or pathlib.Path
- esmvalcore.preprocessor.fix_metadata(cubes: Sequence[Cube], short_name: str, project: str, dataset: str, mip: str, frequency: str | None = None, check_level: CheckLevels = CheckLevels.DEFAULT, session: Session | None = None, **extra_facets) CubeList [source]#
Fix cube metadata if fixes are required.
This method collects all the relevant fixes (including generic ones) for a given variable and applies them.
- Parameters:
cubes (Sequence[Cube]) – Cubes to fix.
short_name (str) – Variable’s short name.
project (str) – Project of the dataset.
dataset (str) – Name of the dataset.
mip (str) – Variable’s MIP.
frequency (Optional[str]) – Variable’s data frequency, if available.
check_level (CheckLevels) –
Level of strictness of the checks.
Deprecated since version 2.10.0: This option has been deprecated in ESMValCore version 2.10.0 and is scheduled for removal in version 2.12.0. Please use the functions
cmor_check_metadata()
,cmor_check_data()
, orcmor_check()
instead. This function will no longer perform CMOR checks. Fixes and CMOR checks have been clearly separated in ESMValCore version 2.10.0.session (Optional[Session]) – Current session which includes configuration and directory information.
**extra_facets – Extra facets are mainly used for data outside of the big projects like CMIP, CORDEX, obs4MIPs. For details, see Extra Facets.
- Returns:
Fixed cubes.
- Return type:
- esmvalcore.preprocessor.hourly_statistics(cube: Cube, hours: int, operator: str = 'mean', **operator_kwargs) Cube [source]#
Compute hourly statistics.
Chunks time in x hours periods and computes statistics over them.
- Parameters:
cube (Cube) – Input cube.
hours (int) – Number of hours per period. Must be a divisor of 24, i.e., (1, 2, 3, 4, 6, 8, 12).
operator (str) – The operation. Used to determine the
iris.analysis.Aggregator
object used to calculate the statistics. Allowed options are given in this table.**operator_kwargs – Optional keyword arguments for the
iris.analysis.Aggregator
object defined by operator.
- Returns:
Hourly statistics cube.
- Return type:
- esmvalcore.preprocessor.linear_trend(cube, coordinate='time')[source]#
Calculate linear trend of data along a given coordinate.
The linear trend is defined as the slope of an ordinary linear regression.
- Parameters:
cube (iris.cube.Cube) – Input data.
coordinate (str, optional (default: 'time')) – Dimensional coordinate over which the trend is calculated.
- Returns:
Trends.
- Return type:
- Raises:
iris.exceptions.CoordinateNotFoundError – The dimensional coordinate with the name
coordinate
is not found incube
.
- esmvalcore.preprocessor.linear_trend_stderr(cube, coordinate='time')[source]#
Calculate standard error of linear trend along a given coordinate.
This gives the standard error (not confidence intervals!) of the trend defined as the standard error of the estimated slope of an ordinary linear regression.
- Parameters:
cube (iris.cube.Cube) – Input data.
coordinate (str, optional (default: 'time')) – Dimensional coordinate over which the standard error of the trend is calculated.
- Returns:
Standard errors of trends.
- Return type:
- Raises:
iris.exceptions.CoordinateNotFoundError – The dimensional coordinate with the name
coordinate
is not found incube
.
- esmvalcore.preprocessor.load(file: str | Path, ignore_warnings: list[dict] | None = None) CubeList [source]#
Load iris cubes from string or Path objects.
- Parameters:
file (str | Path) – File to be loaded. Could be string or POSIX Path object.
ignore_warnings (list[dict] | None) – Keyword arguments passed to
warnings.filterwarnings()
used to ignore warnings issued byiris.load_raw()
. Each list element corresponds to one call towarnings.filterwarnings()
.
- Returns:
Loaded cubes.
- Return type:
- Raises:
ValueError – Cubes are empty.
- esmvalcore.preprocessor.local_solar_time(cube: Cube) Cube [source]#
Convert UTC time coordinate to local solar time (LST).
This preprocessor transforms input data with a UTC-based time coordinate to a local solar time (LST) coordinate. In LST, 12:00 noon is defined as the moment when the sun reaches its highest point in the sky. Thus, LST is mainly determined by longitude of a location. LST is particularly suited to analyze diurnal cycles across larger regions of the globe, which would be phase-shifted against each other when using UTC time.
To transform data from UTC to LST, this function shifts data along the time axis based on the longitude. In addition to the cube’s data, this function also considers auxiliary coordinates, cell measures and ancillary variables that span both the time and longitude dimension.
Note
This preprocessor preserves the temporal frequency of the input data. For example, hourly input data will be transformed into hourly output data. For this, a location’s exact LST will be put into corresponding bins defined by the bounds of the input time coordinate (in this example, the bin size is 1 hour). If time bounds are not given or cannot be approximated (only one time step is given), a bin size of 1 hour is assumed.
LST is approximated as UTC_time + 12*longitude/180, where longitude is assumed to be in [-180, 180] (this function automatically calculates the correct format for the longitude). This is only an approximation since the exact LST also depends on the day of year due to the eccentricity of Earth’s orbit (see equation of time). However, since the corresponding error is ~15 min at most, this is ignored here, as most climate model data has a courser temporal resolution and the time scale for diurnal evolution of meteorological phenomena is usually in the order of hours, not minutes.
- Parameters:
cube (Cube) – Input cube. Needs a 1D monotonically increasing dimensional coordinate time (assumed to refer to UTC time) and a 1D coordinate longitude.
- Returns:
Transformed cube of same shape as input cube with an LST coordinate instead of UTC time.
- Return type:
Cube
- Raises:
iris.exceptions.CoordinateNotFoundError – Input cube does not contain valid time and/or longitude coordinate.
iris.exceptions.CoordinateMultiDimError – Input cube has multidimensional longitude coordinate.
ValueError – time coordinate of input cube is not monotonically increasing.
- esmvalcore.preprocessor.mask_above_threshold(cube, threshold)[source]#
Mask above a specific threshold value.
Takes a value ‘threshold’ and masks off anything that is above it in the cube data. Values equal to the threshold are not masked.
- Parameters:
cube (iris.cube.Cube) – iris cube to be thresholded.
threshold (float) – threshold to be applied on input cube data.
- Returns:
thresholded cube.
- Return type:
- esmvalcore.preprocessor.mask_below_threshold(cube, threshold)[source]#
Mask below a specific threshold value.
Takes a value ‘threshold’ and masks off anything that is below it in the cube data. Values equal to the threshold are not masked.
- Parameters:
cube (iris.cube.Cube) – iris cube to be thresholded
threshold (float) – threshold to be applied on input cube data.
- Returns:
thresholded cube.
- Return type:
- esmvalcore.preprocessor.mask_fillvalues(products, threshold_fraction, min_value=None, time_window=1)[source]#
Compute and apply a multi-dataset fillvalues mask.
Construct the mask that fills a certain time window with missing values if the number of values in that specific window is less than a given fractional threshold. This function is the extension of _get_fillvalues_mask and performs the combination of missing values masks from each model (of multimodels) into a single fillvalues mask to be applied to each model.
- Parameters:
products (iris.cube.Cube) – data products to be masked.
threshold_fraction (float) – fractional threshold to be used as argument for Aggregator. Must be between 0 and 1.
min_value (float) – minimum value threshold; default None If default, no thresholding applied so the full mask will be selected.
time_window (float) – time window to compute missing data counts; default set to 1.
- Returns:
Masked iris cubes.
- Return type:
- Raises:
NotImplementedError – Implementation missing for data with higher dimensionality than 4.
- esmvalcore.preprocessor.mask_glaciated(cube, mask_out: str = 'glaciated')[source]#
Mask out glaciated areas.
It applies a Natural Earth mask. Note that for computational reasons only the 10 largest polygons are used for masking.
- Parameters:
cube (iris.cube.Cube) – data cube to be masked.
mask_out (str) – “glaciated” to mask out glaciated areas
- Returns:
Returns the masked iris cube.
- Return type:
- Raises:
ValueError – Error raised if masking on irregular grids is attempted or if mask_out has a wrong value.
- esmvalcore.preprocessor.mask_inside_range(cube, minimum, maximum)[source]#
Mask inside a specific threshold range.
Takes a MINIMUM and a MAXIMUM value for the range, and masks off anything that’s between the two in the cube data.
- Parameters:
cube (iris.cube.Cube) – iris cube to be thresholded
minimum (float) – lower threshold to be applied on input cube data.
maximum (float) – upper threshold to be applied on input cube data.
- Returns:
thresholded cube.
- Return type:
- esmvalcore.preprocessor.mask_landsea(cube, mask_out)[source]#
Mask out either land mass or sea (oceans, seas and lakes).
It uses dedicated ancillary variables (sftlf or sftof) or, in their absence, it applies a Natural Earth mask (land or ocean contours). Note that the Natural Earth masks have different resolutions: 10m for land, and 50m for seas. These are more than enough for masking climate model data.
- Parameters:
cube (iris.cube.Cube) – data cube to be masked. If the cube has an
iris.coords.AncillaryVariable
with standard name'land_area_fraction'
or'sea_area_fraction'
that will be used. If both are present, only the ‘land_area_fraction’ will be used. If the ancillary variable is not available, the mask will be calculated from Natural Earth shapefiles.mask_out (str) – either “land” to mask out land mass or “sea” to mask out seas.
- Returns:
Returns the masked iris cube.
- Return type:
- Raises:
ValueError – Error raised if masking on irregular grids is attempted without an ancillary variable. Irregular grids are not currently supported for masking with Natural Earth shapefile masks.
- esmvalcore.preprocessor.mask_landseaice(cube, mask_out)[source]#
Mask out either landsea (combined) or ice.
Function that masks out either landsea (land and seas) or ice (Antarctica, Greenland and some glaciers).
It uses dedicated ancillary variables (sftgif).
- Parameters:
cube (iris.cube.Cube) – data cube to be masked. It should have an
iris.coords.AncillaryVariable
with standard name'land_ice_area_fraction'
.mask_out (str) – either “landsea” to mask out landsea or “ice” to mask out ice.
- Returns:
Returns the masked iris cube with either land or ice masked out.
- Return type:
- Raises:
ValueError – Error raised if landsea-ice mask not found as an ancillary variable.
- esmvalcore.preprocessor.mask_multimodel(products)[source]#
Apply common mask to all datasets (using logical OR).
- Parameters:
products (iris.cube.CubeList or list of PreprocessorFile) – Data products/cubes to be masked.
- Returns:
Masked data products/cubes.
- Return type:
iris.cube.CubeList or list of PreprocessorFile
- Raises:
ValueError – Datasets have different shapes.
TypeError – Invalid input data.
- esmvalcore.preprocessor.mask_outside_range(cube, minimum, maximum)[source]#
Mask outside a specific threshold range.
Takes a MINIMUM and a MAXIMUM value for the range, and masks off anything that’s outside the two in the cube data.
- Parameters:
cube (iris.cube.Cube) – iris cube to be thresholded
minimum (float) – lower threshold to be applied on input cube data.
maximum (float) – upper threshold to be applied on input cube data.
- Returns:
thresholded cube.
- Return type:
- esmvalcore.preprocessor.meridional_statistics(cube: Cube, operator: str, normalize: Literal['subtract', 'divide'] | None = None, **operator_kwargs) Cube [source]#
Compute meridional statistics.
- Parameters:
cube (Cube) – Input cube.
operator (str) – The operation. Used to determine the
iris.analysis.Aggregator
object used to calculate the statistics. Allowed options are given in this table.normalize (Literal['subtract', 'divide'] | None) – If given, do not return the statistics cube itself, but rather, the input cube, normalized with the statistics cube. Can either be subtract (statistics cube is subtracted from the input cube) or divide (input cube is divided by the statistics cube).
**operator_kwargs – Optional keyword arguments for the
iris.analysis.Aggregator
object defined by operator.
- Returns:
Meridional statistics cube.
- Return type:
- Raises:
ValueError – Error raised if computation on irregular grids is attempted. Zonal statistics not yet implemented for irregular grids.
- esmvalcore.preprocessor.monthly_statistics(cube: Cube, operator: str = 'mean', **operator_kwargs) Cube [source]#
Compute monthly statistics.
Chunks time in monthly periods and computes statistics over them;
- Parameters:
cube (Cube) – Input cube.
operator (str) – The operation. Used to determine the
iris.analysis.Aggregator
object used to calculate the statistics. Allowed options are given in this table.**operator_kwargs – Optional keyword arguments for the
iris.analysis.Aggregator
object defined by operator.
- Returns:
Monthly statistics cube.
- Return type:
- esmvalcore.preprocessor.multi_model_statistics(products: set[PreprocessorFile] | Iterable[Cube], span: str, statistics: list[str | dict], output_products=None, groupby: tuple | None = None, keep_input_datasets: bool = True, ignore_scalar_coords: bool = False) dict | set [source]#
Compute multi-model statistics.
This function computes multi-model statistics on a list of
products
, which can be instances ofCube
orPreprocessorFile
. The latter is used internally by ESMValCore to store workflow and provenance information, and this option should typically be ignored.Cubes must have consistent shapes apart from a potential time dimension. There are two options to combine time coordinates of different lengths, see the
span
argument.Desired statistics need to be given as a list, e.g.,
statistics: ['mean', 'median']
. For some statistics like percentiles, it is also possible to pass additional keyword arguments, for examplestatistics: [{'operator': 'mean', 'percent': 20}]
. A full list of supported statistics is available in the section on Statistical preprocessors.This function can handle cubes with differing metadata:
Cubes with identical
name()
andunits
will get identical values forstandard_name
,long_name
, andvar_name
(which will be arbitrarily set to the first encountered value if different cubes have different values for them).attributes
: Differing attributes are deleted, seeiris.util.equalise_attributes()
.cell_methods
: All cell methods are deleted prior to combining cubes.cell_measures()
: All cell measures are deleted prior to combining cubes, seeesmvalcore.preprocessor.remove_fx_variables()
.ancillary_variables()
: All ancillary variables are deleted prior to combining cubes, seeesmvalcore.preprocessor.remove_fx_variables()
.coords()
: Exactly identical coordinates are preserved. For coordinates with equalname()
andunits
, names are equalized,attributes
deleted andcircular
is set toFalse
. For all other coordinates,long_name
is removed,attributes
deleted andcircular
is set toFalse
. Scalar coordinates can be removed if desired by the optionignore_scalar_coords=True
. Please note that some special scalar coordinates which are expected to differ across cubes (ancillary coordinates for derived coordinates like p0 and ptop) are always removed.
Notes
Some of the operators in
iris.analysis
require additional arguments. Except for percentiles, these operators are currently not supported.Lazy operation is supported for all statistics, except
median
.- Parameters:
products (set[PreprocessorFile] | Iterable[Cube]) – Cubes (or products) over which the statistics will be computed.
span (str) – Overlap or full; if overlap, statitstics are computed on common time- span; if full, statistics are computed on full time spans, ignoring missing data. This option is ignored if input cubes do not have time dimensions.
statistics (list[str | dict]) – Statistical operations to be computed, e.g.,
['mean', 'median']
. For some statistics like percentiles, it is also possible to pass additional keyword arguments, e.g.,[{'operator': 'mean', 'percent': 20}]
. All supported options are are given in this table.output_products (dict) – For internal use only. A dict with statistics names as keys and preprocessorfiles as values. If products are passed as input, the statistics cubes will be assigned to these output products.
groupby (Optional[tuple]) – Group products by a given tag or attribute, e.g., (‘project’, ‘dataset’, ‘tag1’). This is ignored if
products
is a list of cubes.keep_input_datasets (bool) – If True, the output will include the input datasets. If False, only the computed statistics will be returned.
ignore_scalar_coords (bool) – If True, remove any scalar coordinate in the input datasets before merging the input cubes into the multi-dataset cube. The resulting multi-dataset cube will have no scalar coordinates (the actual input datasets will remain unchanged). If False, scalar coordinates will remain in the input datasets, which might lead to merge conflicts in case the input datasets have different scalar coordinates.
- Returns:
A
dict
of cubes orset
of output_products depending on the type of products.- Return type:
- Raises:
ValueError – If span is neither overlap nor full, or if input type is neither cubes nor products.
- esmvalcore.preprocessor.regrid(cube: Cube, target_grid: Cube | Dataset | Path | str | dict, scheme: str | dict, lat_offset: bool = True, lon_offset: bool = True) Cube [source]#
Perform horizontal regridding.
Note that the target grid can be a
Cube
, aDataset
, a path to a cube (Path
orstr
), a grid spec (str
) in the form of MxN, or adict
specifying the target grid.For the latter, the target_grid should be a
dict
with the following keys:start_longitude
: longitude at the center of the first grid cell.end_longitude
: longitude at the center of the last grid cell.step_longitude
: constant longitude distance between grid cellcenters.
start_latitude
: latitude at the center of the first grid cell.end_latitude
: longitude at the center of the last grid cell.step_latitude
: constant latitude distance between grid cell centers.
- Parameters:
cube (Cube) – The source cube to be regridded.
target_grid (Cube | Dataset | Path | str | dict) – The (location of a) cube that specifies the target or reference grid for the regridding operation. Alternatively, a
Dataset
can be provided. Alternatively, a string cell specification may be provided, of the formMxN
, which specifies the extent of the cell, longitude by latitude (degrees) for a global, regular target grid. Alternatively, a dictionary with a regional target grid may be specified (see above).scheme (str | dict) – The regridding scheme to perform. If the source grid is structured (regular or irregular), can be one of the built-in schemes
linear
,nearest
,area_weighted
. If the source grid is unstructured, can be one of the built-in schemesnearest
. Alternatively, a dict that specifies generic regridding can be given (see below).lat_offset (bool) – Offset the grid centers of the latitude coordinate w.r.t. the pole by half a grid step. This argument is ignored if target_grid is a cube or file.
lon_offset (bool) – Offset the grid centers of the longitude coordinate w.r.t. Greenwich meridian by half a grid step. This argument is ignored if target_grid is a cube or file.
- Returns:
Regridded cube.
- Return type:
See also
extract_levels
Perform vertical regridding.
Notes
This preprocessor allows for the use of arbitrary Iris regridding schemes, that is anything that can be passed as a scheme to
iris.cube.Cube.regrid()
is possible. This enables the use of further parameters for existing schemes, as well as the use of more advanced schemes for example for unstructured meshes. To use this functionality, a dictionary must be passed for the scheme with a mandatory entry ofreference
in the form specified for the object reference of the entry point data model, i.e.importable.module:object.attr
. This is used as a factory for the scheme. Any further entries in the dictionary are passed as keyword arguments to the factory.For example, to use the familiar
iris.analysis.Linear
regridding scheme with a custom extrapolation mode, usemy_preprocessor: regrid: target: 1x1 scheme: reference: iris.analysis:Linear extrapolation_mode: nanmask
To use the area weighted regridder available in
esmf_regrid.schemes.ESMFAreaWeighted
usemy_preprocessor: regrid: target: 1x1 scheme: reference: esmf_regrid.schemes:ESMFAreaWeighted
- esmvalcore.preprocessor.regrid_time(cube: Cube, frequency: str) Cube [source]#
Align time axis for cubes so they can be subtracted.
Operations on time units, time points and auxiliary coordinates so that any cube from cubes can be subtracted from any other cube from cubes. Currently this function supports yearly (frequency=yr), monthly (frequency=mon), daily (frequency=day), 6-hourly (frequency=6hr), 3-hourly (frequency=3hr) and hourly (frequency=1hr) data time frequencies.
- Parameters:
- Returns:
Cube with converted time axis and units.
- Return type:
- esmvalcore.preprocessor.remove_supplementary_variables(cube: Cube)[source]#
Remove supplementary variables.
Strip cell measures or ancillary variables from the cube containing the data.
- Parameters:
cube (iris.cube.Cube) – Iris cube with data and cell measures or ancillary variables.
- Returns:
Cube without cell measures or ancillary variables.
- Return type:
- esmvalcore.preprocessor.resample_hours(cube: Cube, interval: int, offset: int = 0) Cube [source]#
Convert x-hourly data to y-hourly by eliminating extra timesteps.
Convert x-hourly data to y-hourly (y > x) by eliminating the extra timesteps. This is intended to be used only with instantaneous values.
For example:
resample_hours(cube, interval=6): Six-hourly intervals at 0:00, 6:00, 12:00, 18:00.
resample_hours(cube, interval=6, offset=3): Six-hourly intervals at 3:00, 9:00, 15:00, 21:00.
resample_hours(cube, interval=12, offset=6): Twelve-hourly intervals at 6:00, 18:00.
- Parameters:
- Returns:
Cube with the new frequency.
- Return type:
- Raises:
ValueError: – The specified frequency is not a divisor of 24.
- esmvalcore.preprocessor.resample_time(cube: Cube, month: int | None = None, day: int | None = None, hour: int | None = None) Cube [source]#
Change frequency of data by resampling it.
Converts data from one frequency to another by extracting the timesteps that match the provided month, day and/or hour. This is meant to be used with instantaneous values when computing statistics is not desired.
For example:
resample_time(cube, hour=6): Daily values taken at 6:00.
resample_time(cube, day=15, hour=6): Monthly values taken at 15th 6:00.
resample_time(cube, month=6): Yearly values, taking in June
resample_time(cube, month=6, day=1): Yearly values, taking 1st June
The condition must yield only one value per interval: the last two samples above will produce yearly data, but the first one is meant to be used to sample from monthly output and the second one will work better with daily.
- Parameters:
cube (Cube) – Input cube.
month (optional) – Month to extract.
day (optional) – Day to extract.
hour (optional) – Hour to extract.
- Returns:
Cube with the new frequency.
- Return type:
- esmvalcore.preprocessor.rolling_window_statistics(cube: Cube, coordinate: str, operator: str, window_length: int, **operator_kwargs)[source]#
Compute rolling-window statistics over a coordinate.
- Parameters:
cube (Cube) – Input cube.
coordinate (str) – Coordinate over which the rolling-window statistics is calculated.
operator (str) – The operation. Used to determine the
iris.analysis.Aggregator
object used to calculate the statistics. Allowed options are given in this table.**operator_kwargs – Optional keyword arguments for the
iris.analysis.Aggregator
object defined by operator.window_length (int) – Size of the window to use.
- Returns:
Rolling-window statistics cube.
- Return type:
- esmvalcore.preprocessor.save(cubes, filename, optimize_access='', compress=False, alias='', **kwargs)[source]#
Save iris cubes to file.
- Parameters:
cubes (iterable of iris.cube.Cube) – Data cubes to be saved
filename (str) – Name of target file
optimize_access (str) –
Set internal NetCDF chunking to favour a reading scheme
Values can be map or timeseries, which improve performance when reading the file one map or time series at a time. Users can also provide a coordinate or a list of coordinates. In that case the better performance will be avhieved by loading all the values in that coordinate at a time
compress (bool, optional) – Use NetCDF internal compression.
alias (str, optional) – Var name to use when saving instead of the one in the cube.
- Returns:
filename
- Return type:
- Raises:
ValueError – cubes is empty.
- esmvalcore.preprocessor.seasonal_statistics(cube: Cube, operator: str = 'mean', seasons: Iterable[str] = ('DJF', 'MAM', 'JJA', 'SON'), **operator_kwargs) Cube [source]#
Compute seasonal statistics.
Chunks time seasons and computes statistics over them.
- Parameters:
cube (Cube) – Input cube.
operator (str) – The operation. Used to determine the
iris.analysis.Aggregator
object used to calculate the statistics. Allowed options are given in this table.seasons (Iterable[str]) – Seasons to build. Available: (‘DJF’, ‘MAM’, ‘JJA’, SON’) (default) and all sequentially correct combinations holding every month of a year: e.g. (‘JJAS’,’ONDJFMAM’), or less in case of prior season extraction.
**operator_kwargs – Optional keyword arguments for the
iris.analysis.Aggregator
object defined by operator.
- Returns:
Seasonal statistic cube.
- Return type:
- esmvalcore.preprocessor.timeseries_filter(cube: Cube, window: int, span: int, filter_type: str = 'lowpass', filter_stats: str = 'sum', **operator_kwargs) Cube [source]#
Apply a timeseries filter.
Method borrowed from iris example
Apply each filter using the rolling_window method used with the weights keyword argument. A weighted sum is required because the magnitude of the weights are just as important as their relative sizes.
See also the iris rolling window
iris.cube.Cube.rolling_window
.- Parameters:
cube (Cube) – Input cube.
window (int) – The length of the filter window (in units of cube time coordinate).
span (int) – Number of months/days (depending on data frequency) on which weights should be computed e.g. 2-yearly: span = 24 (2 x 12 months). Span should have same units as cube time coordinate.
filter_type (optional) – Type of filter to be applied; default ‘lowpass’. Available types: ‘lowpass’.
filter_stats (optional) – Type of statistic to aggregate on the rolling window; default: sum. Used to determine the
iris.analysis.Aggregator
object used for aggregation. Allowed options are given in this table.**operator_kwargs – Optional keyword arguments for the
iris.analysis.Aggregator
object defined by filter_stats.
- Returns:
Cube time-filtered using ‘rolling_window’.
- Return type:
- Raises:
iris.exceptions.CoordinateNotFoundError: – Cube does not have time coordinate.
NotImplementedError: – filter_type is not implemented.
- esmvalcore.preprocessor.volume_statistics(cube: Cube, operator: str, normalize: Literal['subtract', 'divide'] | None = None, **operator_kwargs) Cube [source]#
Apply a statistical operation over a volume.
The volume average is weighted according to the cell volume.
- Parameters:
cube (Cube) – Input cube. The input cube should have a
iris.coords.CellMeasure
named'ocean_volume'
, unless it has regular 1D latitude and longitude coordinates so the cell volumes can be computed by usingiris.analysis.cartography.area_weights()
to compute the cell areas and multiplying those by the cell thickness, computed from the bounds of the vertical coordinate.operator (str) – The operation. Used to determine the
iris.analysis.Aggregator
object used to calculate the statistics. Currently, only mean is allowed.normalize (Literal['subtract', 'divide'] | None) – If given, do not return the statistics cube itself, but rather, the input cube, normalized with the statistics cube. Can either be subtract (statistics cube is subtracted from the input cube) or divide (input cube is divided by the statistics cube).
**operator_kwargs – Optional keyword arguments for the
iris.analysis.Aggregator
object defined by operator.
- Returns:
Collapsed cube.
- Return type:
- esmvalcore.preprocessor.weighting_landsea_fraction(cube, area_type)[source]#
Weight fields using land or sea fraction.
This preprocessor function weights a field with its corresponding land or sea area fraction (value between 0 and 1). The application of this is important for most carbon cycle variables (and other land-surface outputs), which are e.g. reported in units of kgC m-2. This actually refers to ‘per square meter of land/sea’ and NOT ‘per square meter of gridbox’. So in order to integrate these globally or regionally one has to both area-weight the quantity but also weight by the land/sea fraction.
- Parameters:
cube (iris.cube.Cube) – Data cube to be weighted. It should have an
iris.coords.AncillaryVariable
with standard name'land_area_fraction'
or'sea_area_fraction'
. If both are present, only the'land_area_fraction'
will be used.area_type (str) – Use land (
'land'
) or sea ('sea'
) fraction for weighting.
- Returns:
Land/sea fraction weighted cube.
- Return type:
- Raises:
TypeError –
area_type
is not'land'
or'sea'
.ValueError – Land/sea fraction variables
sftlf
orsftof
not found.
- esmvalcore.preprocessor.zonal_statistics(cube: Cube, operator: str, normalize: Literal['subtract', 'divide'] | None = None, **operator_kwargs) Cube [source]#
Compute zonal statistics.
- Parameters:
cube (Cube) – Input cube.
operator (str) – The operation. Used to determine the
iris.analysis.Aggregator
object used to calculate the statistics. Allowed options are given in this table.normalize (Literal['subtract', 'divide'] | None) – If given, do not return the statistics cube itself, but rather, the input cube, normalized with the statistics cube. Can either be subtract (statistics cube is subtracted from the input cube) or divide (input cube is divided by the statistics cube).
**operator_kwargs – Optional keyword arguments for the
iris.analysis.Aggregator
object defined by operator.
- Returns:
Zonal statistics cube or input cube normalized by statistics cube (see normalize).
- Return type:
- Raises:
ValueError – Error raised if computation on irregular grids is attempted. Zonal statistics not yet implemented for irregular grids.