calculator / utils /indicators.py
pgzmnk
docstring
bccf193
raw
history blame
11.5 kB
import datetime
import json
import os
from itertools import repeat
import ee
import numpy as np
import pandas as pd
import plotly.graph_objects as go
import yaml
from utils import duckdb_queries as dq
from . import logging
GEE_SERVICE_ACCOUNT = (
"climatebase-july-2023@ee-geospatialml-aquarry.iam.gserviceaccount.com"
)
INDICES_FILE = "indices.yaml"
class IndexGenerator:
"""
A class to generate indices and compute zonal means.
Args:
indices (string[], required): Array of index names to include in aggregate index generation.
"""
def __init__(
self,
indices,
):
# Authenticate to GEE & DuckDB
self._authenticate_ee(GEE_SERVICE_ACCOUNT)
# Use defined subset of indices
all_indices = self._load_indices(INDICES_FILE)
self.indices = {k: all_indices[k] for k in indices}
def _cloudfree(self, gee_path, daterange):
"""
Internal method to generate a cloud-free composite.
Args:
gee_path (str): The path to the Google Earth Engine (GEE) image or image collection.
Returns:
ee.Image: The cloud-free composite clipped to the region of interest.
"""
# Load a raw Landsat ImageCollection for a single year.
collection = (
ee.ImageCollection(gee_path).filterDate(*daterange).filterBounds(self.roi)
)
# Create a cloud-free composite with custom parameters for cloud score threshold and percentile.
composite_cloudfree = ee.Algorithms.Landsat.simpleComposite(
**{"collection": collection, "percentile": 75, "cloudScoreRange": 5}
)
return composite_cloudfree.clip(self.roi)
def _load_indices(self, indices_file):
# Read index configurations
with open(indices_file, "r") as stream:
try:
return yaml.safe_load(stream)
except yaml.YAMLError as e:
logging.error(e)
return None
def generate_index(self, index_config, year):
"""
Generates an index based on the provided index configuration.
Args:
index_config (dict): Configuration for generating the index.
Returns:
ee.Image: The generated index clipped to the region of interest.
"""
# Calculate date range, assume 1 year
start_date = str(datetime.date(year, 1, 1))
end_date = str(datetime.date(year, 12, 31))
daterange = [start_date, end_date]
# Calculate index based on type
logging.info(
f"Generating index: {index_config['name']} of type {index_config['gee_type']}"
)
match index_config["gee_type"]:
case "image":
dataset = ee.Image(index_config["gee_path"]).clip(self.roi)
if index_config.get("select"):
dataset = dataset.select(index_config["select"])
case "image_collection":
dataset = (
ee.ImageCollection(index_config["gee_path"])
.filterBounds(self.roi)
.map(lambda image: image.clip(self.roi))
.mean()
)
if index_config.get("select"):
dataset = dataset.select(index_config["select"])
case "feature_collection":
dataset = (
ee.Image()
.float()
.paint(
ee.FeatureCollection(index_config["gee_path"]),
index_config["select"],
)
.clip(self.roi)
)
case "algebraic":
image = self._cloudfree(index_config["gee_path"], daterange)
# to-do: params should come from index_config
dataset = image.normalizedDifference(["B4", "B3"])
case _:
dataset = None
if not dataset:
raise Exception("Failed to generate dataset.")
logging.info(f"Generated index: {index_config['name']}")
return dataset
def zonal_mean_index(self, index_key, year):
index_config = self.indices[index_key]
dataset = self.generate_index(index_config, year)
logging.info(f"Calculating zonal mean for {index_key}...")
out = dataset.reduceRegion(
**{
"reducer": ee.Reducer.mean(),
"geometry": self.roi,
"scale": 2000, # map scale
"bestEffort": True,
"maxPixels": 1e3,
}
).getInfo()
if index_config.get("bandname"):
return out[index_config.get("bandname")]
logging.info(f"Calculated zonal mean for {index_key}.")
return out
def generate_composite_index_df(self, year, project_geometry, indices=[]):
data = {
"metric": indices,
"year": year,
"centroid": "",
"project_name": "",
"value": list(map(self.zonal_mean_index, indices, repeat(year))),
# to-do: calculate with duckdb; also, should be part of project table instead
"area": self.roi.area().getInfo(), # m^2
"geojson": "",
# to-do: coefficient
}
logging.info("data", data)
df = pd.DataFrame(data)
return df
@staticmethod
def _authenticate_ee(ee_service_account):
"""
Huggingface Spaces does not support secret files, therefore authenticate with an environment variable containing the JSON.
"""
logging.info("Authenticating to Google Earth Engine...")
credentials = ee.ServiceAccountCredentials(
ee_service_account, key_data=os.environ["ee_service_account"]
)
ee.Initialize(credentials)
logging.info("Authenticated to Google Earth Engine.")
def _calculate_yearly_index(self, years, project_name):
dfs = []
logging.info(years)
project_geometry = dq.get_project_geometry(project_name)
project_centroid = dq.get_project_centroid(project_name)
# to-do: refactor to involve less transformations
_polygon = json.dumps(
json.loads(project_geometry[0][0])["features"][0]["geometry"]
)
# to-do: don't use self.roi and instead pass patameter strategically
self.roi = ee.Geometry.Polygon(json.loads(_polygon)["coordinates"])
# to-do: pararelize?
for year in years:
logging.info(year)
self.project_name = project_name
df = self.generate_composite_index_df(
year, project_geometry, list(self.indices.keys())
)
dfs.append(df)
# Concatenate all dataframes
df_concat = pd.concat(dfs)
df_concat["centroid"] = str(project_centroid)
df_concat["project_name"] = project_name
df_concat["geojson"] = str(project_geometry)
return df_concat.round(2)
# h/t: https://community.plotly.com/t/dynamic-zoom-for-mapbox/32658/12
def _latlon_to_config(self, longitudes=None, latitudes=None):
"""Function documentation:\n
Basic framework adopted from Krichardson under the following thread:
https://community.plotly.com/t/dynamic-zoom-for-mapbox/32658/7
# NOTE:
# THIS IS A TEMPORARY SOLUTION UNTIL THE DASH TEAM IMPLEMENTS DYNAMIC ZOOM
# in their plotly-functions associated with mapbox, such as go.Densitymapbox() etc.
Returns the appropriate zoom-level for these plotly-mapbox-graphics along with
the center coordinate tuple of all provided coordinate tuples.
"""
# Check whether both latitudes and longitudes have been passed,
# or if the list lenghts don't match
if (latitudes is None or longitudes is None) or (
len(latitudes) != len(longitudes)
):
# Otherwise, return the default values of 0 zoom and the coordinate origin as center point
return 0, (0, 0)
# Get the boundary-box
b_box = {}
b_box["height"] = latitudes.max() - latitudes.min()
b_box["width"] = longitudes.max() - longitudes.min()
b_box["center"] = (np.mean(longitudes), np.mean(latitudes))
# get the area of the bounding box in order to calculate a zoom-level
area = b_box["height"] * b_box["width"]
# * 1D-linear interpolation with numpy:
# - Pass the area as the only x-value and not as a list, in order to return a scalar as well
# - The x-points "xp" should be in parts in comparable order of magnitude of the given area
# - The zpom-levels are adapted to the areas, i.e. start with the smallest area possible of 0
# which leads to the highest possible zoom value 20, and so forth decreasing with increasing areas
# as these variables are antiproportional
zoom = np.interp(
x=area,
xp=[0, 5**-10, 4**-10, 3**-10, 2**-10, 1**-10, 1**-5],
fp=[20, 15, 14, 13, 12, 7, 5],
)
# Finally, return the zoom level and the associated boundary-box center coordinates
return zoom, b_box["center"]
def show_project_map(self, project_name):
project_geometry = dq.get_project_geometry(project_name)
features = json.loads(project_geometry[0][0].replace("'", '"'))["features"]
geometry = features[0]["geometry"]
longitudes = np.array(geometry["coordinates"])[0, :, 0]
latitudes = np.array(geometry["coordinates"])[0, :, 1]
zoom, bbox_center = self._latlon_to_config(longitudes, latitudes)
fig = go.Figure(
go.Scattermapbox(
mode="markers",
lon=[bbox_center[0]],
lat=[bbox_center[1]],
marker={"size": 20, "color": ["cyan"]},
)
)
fig.update_layout(
mapbox={
"style": "stamen-terrain",
"center": {"lon": bbox_center[0], "lat": bbox_center[1]},
"zoom": zoom,
"layers": [
{
"source": {
"type": "FeatureCollection",
"features": [{"type": "Feature", "geometry": geometry}],
},
"type": "fill",
"below": "traces",
"color": "royalblue",
}
],
},
margin={"l": 0, "r": 0, "b": 0, "t": 0},
)
return fig
def calculate_biodiversity_score(self, start_year, end_year, project_name):
years = []
for year in range(start_year, end_year):
row_exists = dq.check_if_project_exists_for_year(project_name, year)
if not row_exists:
years.append(year)
if len(years) > 0:
df = self._calculate_yearly_index(years, project_name)
# Write score table to `_temptable`
dq.write_score_to_temptable(df)
# Create `bioindicator` table IF NOT EXISTS.
dq.get_or_create_bioindicator_table()
# UPSERT project record
dq.upsert_project_record()
logging.info("upserted records into motherduck")
scores = dq.get_project_scores(project_name, start_year, end_year)
return scores