The dataset viewer is not available for this split.
Error code: FeaturesError Exception: ParserError Message: Error tokenizing data. C error: Expected 16 fields in line 6, saw 17 Traceback: Traceback (most recent call last): File "/src/services/worker/src/worker/job_runners/split/first_rows.py", line 322, in compute compute_first_rows_from_parquet_response( File "/src/services/worker/src/worker/job_runners/split/first_rows.py", line 88, in compute_first_rows_from_parquet_response rows_index = indexer.get_rows_index( File "/src/libs/libcommon/src/libcommon/parquet_utils.py", line 640, in get_rows_index return RowsIndex( File "/src/libs/libcommon/src/libcommon/parquet_utils.py", line 521, in __init__ self.parquet_index = self._init_parquet_index( File "/src/libs/libcommon/src/libcommon/parquet_utils.py", line 538, in _init_parquet_index response = get_previous_step_or_raise( File "/src/libs/libcommon/src/libcommon/simple_cache.py", line 591, in get_previous_step_or_raise raise CachedArtifactError( libcommon.simple_cache.CachedArtifactError: The previous step failed. During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/src/services/worker/src/worker/job_runners/split/first_rows.py", line 240, in compute_first_rows_from_streaming_response iterable_dataset = iterable_dataset._resolve_features() File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/iterable_dataset.py", line 2216, in _resolve_features features = _infer_features_from_batch(self.with_format(None)._head()) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/iterable_dataset.py", line 1239, in _head return _examples_to_batch(list(self.take(n))) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/iterable_dataset.py", line 1389, in __iter__ for key, example in ex_iterable: File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/iterable_dataset.py", line 1044, in __iter__ yield from islice(self.ex_iterable, self.n) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/iterable_dataset.py", line 282, in __iter__ for key, pa_table in self.generate_tables_fn(**self.kwargs): File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/packaged_modules/csv/csv.py", line 195, in _generate_tables for batch_idx, df in enumerate(csv_file_reader): File "/src/services/worker/.venv/lib/python3.9/site-packages/pandas/io/parsers/readers.py", line 1843, in __next__ return self.get_chunk() File "/src/services/worker/.venv/lib/python3.9/site-packages/pandas/io/parsers/readers.py", line 1985, in get_chunk return self.read(nrows=size) File "/src/services/worker/.venv/lib/python3.9/site-packages/pandas/io/parsers/readers.py", line 1923, in read ) = self._engine.read( # type: ignore[attr-defined] File "/src/services/worker/.venv/lib/python3.9/site-packages/pandas/io/parsers/c_parser_wrapper.py", line 234, in read chunks = self._reader.read_low_memory(nrows) File "parsers.pyx", line 850, in pandas._libs.parsers.TextReader.read_low_memory File "parsers.pyx", line 905, in pandas._libs.parsers.TextReader._read_rows File "parsers.pyx", line 874, in pandas._libs.parsers.TextReader._tokenize_rows File "parsers.pyx", line 891, in pandas._libs.parsers.TextReader._check_tokenize_status File "parsers.pyx", line 2061, in pandas._libs.parsers.raise_parser_error pandas.errors.ParserError: Error tokenizing data. C error: Expected 16 fields in line 6, saw 17
Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.
Dataset: Low-to-High-Resolution Weather Forecasting using Topography
The dataset is intended and structured for the problem of transforming/interpolating low-resolution weather forecasts into higher resolution using topography data.
The dataset consists of 3 different types of data (as illustrated above):
- Historical weather observation data (SMHI)
- Historical weather observation data from selected SMHI observation stations (evaluation points)
- Historical low-resolution weather forecasts (ECMWF)
- For a given SMHI station: Historical, (relatively) low-resolution ECMWF weather forecasts from the 4 nearest ECMWF grid points
- Topography/elevation data (Copernicus DEM GLO-30):
- Topography/elevation data around a given SMHI station, grid enclosed by the 4 ECMWF points
- Mean elevation around each of the nearest 4 ECMWF grid points for a given SMHI station (mean over +/-0.05 arc degree in each direction from the grid point center)
The dataset is meant to facilitate the following modeling pipeline:
- Weather forecasts for a set of 4 neighboring ECMWF points are combined with topography/elevation data and turned into higher resolution forecasts grid (corresponding to the resolution of the topography data)
- SMHI weather observation data is used as a sparse evaluation point of the produced higher-resolution forecasts
License and terms
The terms of usage of the source data and their corresponding licenses:
- SMHI weather observation data: (Creative Common Attribution 4.0)
- ECMWF historical weather forecasts:
- Copernicus DEM GLO-30:
Acknowledgment
The dataset has been developed as part of the OWGRE project, funded within the ERA-Net SES Joint Call 2020 for transnational research, development and demonstration projects.
Data details & samples
SMHI weather observation data
SMHI weather observation data is structured in csv files, separately for each weather parameter and weather observation station. See the sample below:
Datum;Tid (UTC);Lufttemperatur;Kvalitet
2020-01-01;06:00:00;-2.2;G
2020-01-01;12:00:00;-2.7;G
2020-01-01;18:00:00;0.2;G
2020-01-02;06:00:00;0.3;G
2020-01-02;12:00:00;4.3;G
2020-01-02;18:00:00;4.9;G
2020-01-03;06:00:00;6.0;G
2020-01-03;12:00:00;2.7;G
2020-01-03;18:00:00;1.7;G
2020-01-04;06:00:00;-4.6;G
2020-01-04;12:00:00;0.6;G
2020-01-04;18:00:00;-5.9;G
2020-01-05;06:00:00;-7.9;G
2020-01-05;12:00:00;-3.1;G
ECMWF historical weather forecasts
Historical ECMWF weather forecasts contain a number of forecasted weather variables at 4 nearest grid points around each SMHI observation station:
<xarray.Dataset>
Dimensions: (reference_time: 2983, valid_time: 54,
corner_index: 4, station_index: 275)
Coordinates:
* reference_time (reference_time) datetime64[ns] 2020-01-01 ... 20...
latitude (corner_index, station_index) float64 55.3 ... 68.7
longitude (corner_index, station_index) float64 ...
point (corner_index, station_index) int64 ...
* valid_time (valid_time) int32 0 1 2 3 4 5 ... 48 49 50 51 52 53
* station_index (station_index) int64 0 1 2 3 4 ... 271 272 273 274
* corner_index (corner_index) <U3 'llc' 'lrc' 'ulc' 'urc'
station_names (station_index) <U29 ...
station_ids (station_index) int64 ...
Data variables:
PressureReducedMSL (reference_time, valid_time, corner_index, station_index) float32 ...
RelativeHumidity (reference_time, valid_time, corner_index, station_index) float32 ...
SolarDownwardRadiation (reference_time, valid_time, corner_index, station_index) float64 ...
Temperature (reference_time, valid_time, corner_index, station_index) float32 ...
WindDirection:10 (reference_time, valid_time, corner_index, station_index) float32 ...
WindSpeed:10 (reference_time, valid_time, corner_index, station_index) float32 ...
Topography data
The topography data is provided in the chunks cut around each of the SMHI stations. The corners of each chunk correspond to ECMWF forecast grid points.
Each chunk consists approximately 361 x 361 points, spanning across 0.1° x 0.1°. (Some of the values across longitudes are NaN since apparently the Earth is not square [citation needed]).
Loading the data
The dependencies can be installed through conda
or mamba
in the following way:
mamba create -n ourenv python pandas xarray dask netCDF4
Below, for a given SMHI weather observation station, we read the following data:
- weather observations
- historical ECMWF weather forecasts
- topography/elevation
import pickle
import pandas as pd
import xarray as xr
smhi_weather_observation_station_index = 153
smhi_weather_observation_station_id = pd.read_csv(
'./smhi_weather_observation_stations.csv',
index_col='station_index'
).loc[smhi_weather_observation_station_index]['id'] # 102540
weather_parameter = 1 # temperature
# NOTE: Need to unzip the file first!
smhi_observation_data = pd.read_csv(
'./weather_observations/smhi_observations_from_2020/'
f'parameter_{weather_parameter}'
f'/smhi_weather_param_{weather_parameter}_station_{smhi_weather_observation_station_id}.csv',
sep=';',
)
print('SMHI observation data:')
print(smhi_observation_data)
ecmwf_data = xr.open_dataset(
'./ecmwf_historical_weather_forecasts/ECMWF_HRES-reindexed.nc'
).sel(station_index=smhi_weather_observation_station_index)
print('ECMWF data:')
print(ecmwf_data)
topography_data = xr.open_dataset(
'topography/sweden_chunks_copernicus-dem-30m'
f'/topography_chunk_station_index-{smhi_weather_observation_station_index}.nc'
)
print('Topography chunk:')
print(topography_data)
mean_elevations = pickle.load(open('./topography/ecmwf_grid_mean_elevations.pkl', 'rb'))
print('Mean elevation:')
print(mean_elevations[0])
- Downloads last month
- 355