Dataset Viewer
Full Screen
The dataset viewer is not available for this split.
Cannot extract the features (columns) for the split 'train' of the config 'default' of the dataset.
Error code:   FeaturesError
Exception:    ParserError
Message:      Error tokenizing data. C error: Expected 16 fields in line 6, saw 17

Traceback:    Traceback (most recent call last):
                File "/src/services/worker/src/worker/job_runners/split/first_rows.py", line 322, in compute
                  compute_first_rows_from_parquet_response(
                File "/src/services/worker/src/worker/job_runners/split/first_rows.py", line 88, in compute_first_rows_from_parquet_response
                  rows_index = indexer.get_rows_index(
                File "/src/libs/libcommon/src/libcommon/parquet_utils.py", line 640, in get_rows_index
                  return RowsIndex(
                File "/src/libs/libcommon/src/libcommon/parquet_utils.py", line 521, in __init__
                  self.parquet_index = self._init_parquet_index(
                File "/src/libs/libcommon/src/libcommon/parquet_utils.py", line 538, in _init_parquet_index
                  response = get_previous_step_or_raise(
                File "/src/libs/libcommon/src/libcommon/simple_cache.py", line 591, in get_previous_step_or_raise
                  raise CachedArtifactError(
              libcommon.simple_cache.CachedArtifactError: The previous step failed.
              
              During handling of the above exception, another exception occurred:
              
              Traceback (most recent call last):
                File "/src/services/worker/src/worker/job_runners/split/first_rows.py", line 240, in compute_first_rows_from_streaming_response
                  iterable_dataset = iterable_dataset._resolve_features()
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/iterable_dataset.py", line 2216, in _resolve_features
                  features = _infer_features_from_batch(self.with_format(None)._head())
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/iterable_dataset.py", line 1239, in _head
                  return _examples_to_batch(list(self.take(n)))
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/iterable_dataset.py", line 1389, in __iter__
                  for key, example in ex_iterable:
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/iterable_dataset.py", line 1044, in __iter__
                  yield from islice(self.ex_iterable, self.n)
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/iterable_dataset.py", line 282, in __iter__
                  for key, pa_table in self.generate_tables_fn(**self.kwargs):
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/packaged_modules/csv/csv.py", line 195, in _generate_tables
                  for batch_idx, df in enumerate(csv_file_reader):
                File "/src/services/worker/.venv/lib/python3.9/site-packages/pandas/io/parsers/readers.py", line 1843, in __next__
                  return self.get_chunk()
                File "/src/services/worker/.venv/lib/python3.9/site-packages/pandas/io/parsers/readers.py", line 1985, in get_chunk
                  return self.read(nrows=size)
                File "/src/services/worker/.venv/lib/python3.9/site-packages/pandas/io/parsers/readers.py", line 1923, in read
                  ) = self._engine.read(  # type: ignore[attr-defined]
                File "/src/services/worker/.venv/lib/python3.9/site-packages/pandas/io/parsers/c_parser_wrapper.py", line 234, in read
                  chunks = self._reader.read_low_memory(nrows)
                File "parsers.pyx", line 850, in pandas._libs.parsers.TextReader.read_low_memory
                File "parsers.pyx", line 905, in pandas._libs.parsers.TextReader._read_rows
                File "parsers.pyx", line 874, in pandas._libs.parsers.TextReader._tokenize_rows
                File "parsers.pyx", line 891, in pandas._libs.parsers.TextReader._check_tokenize_status
                File "parsers.pyx", line 2061, in pandas._libs.parsers.raise_parser_error
              pandas.errors.ParserError: Error tokenizing data. C error: Expected 16 fields in line 6, saw 17

Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.

Dataset: Low-to-High-Resolution Weather Forecasting using Topography

The dataset is intended and structured for the problem of transforming/interpolating low-resolution weather forecasts into higher resolution using topography data.

Dataset diagram: 4 ECMWF grid points at the corners, 1 SMHI observation station, topographical data in the background

The dataset consists of 3 different types of data (as illustrated above):

  • Historical weather observation data (SMHI)
    • Historical weather observation data from selected SMHI observation stations (evaluation points)
  • Historical low-resolution weather forecasts (ECMWF)
    • For a given SMHI station: Historical, (relatively) low-resolution ECMWF weather forecasts from the 4 nearest ECMWF grid points
  • Topography/elevation data (Copernicus DEM GLO-30):
    • Topography/elevation data around a given SMHI station, grid enclosed by the 4 ECMWF points
    • Mean elevation around each of the nearest 4 ECMWF grid points for a given SMHI station (mean over +/-0.05 arc degree in each direction from the grid point center)

The dataset is meant to facilitate the following modeling pipeline:

  • Weather forecasts for a set of 4 neighboring ECMWF points are combined with topography/elevation data and turned into higher resolution forecasts grid (corresponding to the resolution of the topography data)
  • SMHI weather observation data is used as a sparse evaluation point of the produced higher-resolution forecasts

License and terms

The terms of usage of the source data and their corresponding licenses:

Acknowledgment

The dataset has been developed as part of the OWGRE project, funded within the ERA-Net SES Joint Call 2020 for transnational research, development and demonstration projects.

Data details & samples

SMHI weather observation data

SMHI weather observation data is structured in csv files, separately for each weather parameter and weather observation station. See the sample below:

Datum;Tid (UTC);Lufttemperatur;Kvalitet
2020-01-01;06:00:00;-2.2;G
2020-01-01;12:00:00;-2.7;G
2020-01-01;18:00:00;0.2;G
2020-01-02;06:00:00;0.3;G
2020-01-02;12:00:00;4.3;G
2020-01-02;18:00:00;4.9;G
2020-01-03;06:00:00;6.0;G
2020-01-03;12:00:00;2.7;G
2020-01-03;18:00:00;1.7;G
2020-01-04;06:00:00;-4.6;G
2020-01-04;12:00:00;0.6;G
2020-01-04;18:00:00;-5.9;G
2020-01-05;06:00:00;-7.9;G
2020-01-05;12:00:00;-3.1;G

ECMWF historical weather forecasts

Historical ECMWF weather forecasts contain a number of forecasted weather variables at 4 nearest grid points around each SMHI observation station:

<xarray.Dataset>
Dimensions:                 (reference_time: 2983, valid_time: 54,
                             corner_index: 4, station_index: 275)
Coordinates:
  * reference_time          (reference_time) datetime64[ns] 2020-01-01 ... 20...
    latitude                (corner_index, station_index) float64 55.3 ... 68.7
    longitude               (corner_index, station_index) float64 ...
    point                   (corner_index, station_index) int64 ...
  * valid_time              (valid_time) int32 0 1 2 3 4 5 ... 48 49 50 51 52 53
  * station_index           (station_index) int64 0 1 2 3 4 ... 271 272 273 274
  * corner_index            (corner_index) <U3 'llc' 'lrc' 'ulc' 'urc'
    station_names           (station_index) <U29 ...
    station_ids             (station_index) int64 ...
Data variables:
    PressureReducedMSL      (reference_time, valid_time, corner_index, station_index) float32 ...
    RelativeHumidity        (reference_time, valid_time, corner_index, station_index) float32 ...
    SolarDownwardRadiation  (reference_time, valid_time, corner_index, station_index) float64 ...
    Temperature             (reference_time, valid_time, corner_index, station_index) float32 ...
    WindDirection:10        (reference_time, valid_time, corner_index, station_index) float32 ...
    WindSpeed:10            (reference_time, valid_time, corner_index, station_index) float32 ...

Topography data

The topography data is provided in the chunks cut around each of the SMHI stations. The corners of each chunk correspond to ECMWF forecast grid points.

Each chunk consists approximately 361 x 361 points, spanning across 0.1° x 0.1°. (Some of the values across longitudes are NaN since apparently the Earth is not square [citation needed]).

Sample topography map

Loading the data

The dependencies can be installed through conda or mamba in the following way:

mamba create -n ourenv python pandas xarray dask netCDF4

Below, for a given SMHI weather observation station, we read the following data:

  • weather observations
  • historical ECMWF weather forecasts
  • topography/elevation
import pickle
import pandas as pd
import xarray as xr

smhi_weather_observation_station_index = 153
smhi_weather_observation_station_id = pd.read_csv(
    './smhi_weather_observation_stations.csv',
    index_col='station_index'
).loc[smhi_weather_observation_station_index]['id'] # 102540

weather_parameter = 1 # temperature

# NOTE: Need to unzip the file first!
smhi_observation_data = pd.read_csv(
    './weather_observations/smhi_observations_from_2020/'
    f'parameter_{weather_parameter}'
    f'/smhi_weather_param_{weather_parameter}_station_{smhi_weather_observation_station_id}.csv',
    sep=';',
)
print('SMHI observation data:')
print(smhi_observation_data)

ecmwf_data = xr.open_dataset(
    './ecmwf_historical_weather_forecasts/ECMWF_HRES-reindexed.nc'
).sel(station_index=smhi_weather_observation_station_index)
print('ECMWF data:')
print(ecmwf_data)

topography_data = xr.open_dataset(
    'topography/sweden_chunks_copernicus-dem-30m'
    f'/topography_chunk_station_index-{smhi_weather_observation_station_index}.nc'
)
print('Topography chunk:')
print(topography_data)

mean_elevations = pickle.load(open('./topography/ecmwf_grid_mean_elevations.pkl', 'rb'))
print('Mean elevation:')
print(mean_elevations[0])
Downloads last month
355