nsthorat-lilac's picture
Duplicate from lilacai/nikhil_staging
bfc0ec6
raw
history blame
9.78 kB
"""Utils for the python server."""
import asyncio
import functools
import itertools
import logging
import os
import pathlib
import re
import shutil
import threading
import time
import uuid
from asyncio import AbstractEventLoop
from concurrent.futures import Executor, ThreadPoolExecutor
from datetime import timedelta
from functools import partial, wraps
from typing import IO, Any, Awaitable, Callable, Iterable, Optional, TypeVar, Union
import numpy as np
import requests
import yaml
from google.cloud.storage import Blob, Client
from pydantic import BaseModel
from .env import data_path, env
from .schema import Path
GCS_PROTOCOL = 'gs://'
GCS_REGEX = re.compile(f'{GCS_PROTOCOL}(.*?)/(.*)')
GCS_COPY_CHUNK_SIZE = 1_000
IMAGES_DIR_NAME = 'images'
DATASETS_DIR_NAME = 'datasets'
@functools.cache
def _get_storage_client(thread_id: Optional[int] = None) -> Client:
# The storage client is not thread safe so we use a thread_id to make sure each thread gets a
# separate storage client.
del thread_id
return Client()
def _parse_gcs_path(filepath: str) -> tuple[str, str]:
# match a regular expression to extract the bucket and filename
if matches := GCS_REGEX.match(filepath):
bucket_name, object_name = matches.groups()
return bucket_name, object_name
raise ValueError(f'Failed to parse GCS path: {filepath}')
def _get_gcs_blob(filepath: str) -> Blob:
bucket_name, object_name = _parse_gcs_path(filepath)
storage_client = _get_storage_client(threading.get_ident())
bucket = storage_client.bucket(bucket_name)
return bucket.blob(object_name)
def open_file(filepath: str, mode: str = 'r') -> IO:
"""Open a file handle. It works with both GCS and local paths."""
if filepath.startswith(GCS_PROTOCOL):
blob = _get_gcs_blob(filepath)
return blob.open(mode)
write_mode = 'w' in mode
binary_mode = 'b' in mode
if write_mode:
base_path = os.path.dirname(filepath)
os.makedirs(base_path, exist_ok=True)
encoding = None if binary_mode else 'utf-8'
return open(filepath, mode=mode, encoding=encoding)
def download_http_files(filepaths: list[str]) -> list[str]:
"""Download files from HTTP(s) URLs."""
out_filepaths: list[str] = []
for filepath in filepaths:
if filepath.startswith(('http://', 'https://')):
tmp_filename = uuid.uuid4().hex
tmp_filepath = f'/tmp/{data_path()}/local_cache/{tmp_filename}'
log(f'Downloading from url {filepath} to {tmp_filepath}')
dl = requests.get(filepath, timeout=10000, allow_redirects=True)
with open_file(tmp_filepath, 'wb') as f:
f.write(dl.content)
filepath = tmp_filepath
out_filepaths.append(filepath)
return out_filepaths
def makedirs(dir_path: str) -> None:
"""Recursively makes the directories. It works with both GCS and local paths."""
if dir_path.startswith(GCS_PROTOCOL):
return
os.makedirs(dir_path, exist_ok=True)
def get_datasets_dir(base_dir: Union[str, pathlib.Path]) -> str:
"""Return the output directory that holds all datasets."""
return os.path.join(base_dir, DATASETS_DIR_NAME)
def get_dataset_output_dir(base_dir: Union[str, pathlib.Path], namespace: str,
dataset_name: str) -> str:
"""Return the output directory for a dataset."""
return os.path.join(get_datasets_dir(base_dir), namespace, dataset_name)
def get_lilac_cache_dir(base_dir: Union[str, pathlib.Path]) -> str:
"""Return the output directory for a dataset."""
return os.path.join(base_dir, '.cache', 'lilac')
class CopyRequest(BaseModel):
"""A request to copy a file from source to destination path. Used to copy media files to GCS."""
from_path: str
to_path: str
def copy_batch(copy_requests: list[CopyRequest]) -> None:
"""Copy a single item from a CopyRequest."""
storage_client = _get_storage_client(threading.get_ident())
with storage_client.batch():
for copy_request in copy_requests:
from_gcs = False
if GCS_REGEX.match(copy_request.from_path):
from_gcs = True
to_gcs = False
if GCS_REGEX.match(copy_request.to_path):
to_gcs = True
makedirs(os.path.dirname(copy_request.to_path))
# When both source and destination are local, use the shutil copy.
if not from_gcs and not to_gcs:
shutil.copyfile(copy_request.from_path, copy_request.to_path)
continue
from_bucket: Any = None
to_bucket: Any = None
from_gcs_blob: Any = None
to_object_name: Optional[str] = None
if from_gcs:
from_bucket_name, from_object_name = _parse_gcs_path(copy_request.from_path)
from_bucket = storage_client.bucket(from_bucket_name)
from_gcs_blob = from_bucket.blob(from_object_name)
if to_gcs:
to_bucket_name, to_object_name = _parse_gcs_path(copy_request.to_path)
to_bucket = storage_client.bucket(to_bucket_name)
if from_gcs and to_gcs:
from_bucket.copy_blob(from_gcs_blob, from_bucket, to_object_name)
elif from_gcs and not to_gcs:
from_gcs_blob.download_to_filename(copy_request.to_path)
elif not from_gcs and to_gcs:
to_gcs_blob = to_bucket.blob(to_object_name)
to_gcs_blob.upload_from_filename(copy_request.from_path)
def copy_files(copy_requests: Iterable[CopyRequest], input_gcs: bool, output_gcs: bool) -> None:
"""Copy media files from an input gcs path to an output gcs path."""
start_time = time.time()
chunk_size = 1
if output_gcs and input_gcs:
# When downloading or uploading locally, batching greatly slows down the parallelism as GCS
# batching with storage.batch() has no effect.
# When copying files locally, storage.batch() has no effect and it's better to run each copy in
# separate thread.
chunk_size = GCS_COPY_CHUNK_SIZE
batched_copy_requests = chunks(copy_requests, chunk_size)
with ThreadPoolExecutor() as executor:
executor.map(copy_batch, batched_copy_requests)
log(f'Copy took {time.time() - start_time} seconds.')
Tchunk = TypeVar('Tchunk')
def chunks(iterable: Iterable[Tchunk], size: int) -> Iterable[list[Tchunk]]:
"""Split a list of items into equal-sized chunks. The last chunk might be smaller."""
it = iter(iterable)
chunk = list(itertools.islice(it, size))
while chunk:
yield chunk
chunk = list(itertools.islice(it, size))
def delete_file(filepath: str) -> None:
"""Delete a file. It works for both GCS and local paths."""
if filepath.startswith(GCS_PROTOCOL):
blob = _get_gcs_blob(filepath)
blob.delete()
return
os.remove(filepath)
def file_exists(filepath: Union[str, pathlib.PosixPath]) -> bool:
"""Return true if the file exists. It works with both GCS and local paths."""
str_filepath = str(filepath)
if str_filepath.startswith(GCS_PROTOCOL):
return _get_gcs_blob(str_filepath).exists()
return os.path.exists(filepath)
def get_image_path(output_dir: str, path: Path, row_id: bytes) -> str:
"""Return the GCS file path to an image associated with a specific row."""
path_subdir = '_'.join([str(p) for p in path])
filename = row_id.hex()
return os.path.join(output_dir, IMAGES_DIR_NAME, path_subdir, filename)
Tout = TypeVar('Tout')
def async_wrap(func: Callable[..., Tout],
loop: Optional[AbstractEventLoop] = None,
executor: Optional[Executor] = None) -> Callable[..., Awaitable[Tout]]:
"""Wrap a sync function into an async function."""
@wraps(func)
async def run(*args: Any, **kwargs: Any) -> Any:
current_loop = loop or asyncio.get_running_loop()
pfunc: Callable = partial(func, *args, **kwargs)
return await current_loop.run_in_executor(executor, pfunc)
return run
def is_primitive(obj: object) -> bool:
"""Returns True if the object is a primitive."""
if isinstance(obj, (str, bytes, np.ndarray, int, float)):
return True
if isinstance(obj, Iterable):
return False
return True
def log(log_str: str) -> None:
"""Print and logs a message so it shows up in the logs on cloud."""
if env('DISABLE_LOGS'):
return
print(log_str)
logging.info(log_str)
class DebugTimer:
"""A context manager that prints the time elapsed in a block of code.
```py
with DebugTimer('dot product'):
np.dot(np.random.randn(1000), np.random.randn(1000))
```
$ dot product took 0.001s.
"""
def __init__(self, name: str) -> None:
self.name = name
def __enter__(self) -> 'DebugTimer':
"""Start a timer."""
self.start = time.perf_counter()
return self
def __exit__(self, *args: list[Any]) -> None:
"""Stop the timer and print the elapsed time."""
log(f'{self.name} took {(time.perf_counter() - self.start):.3f}s.')
def pretty_timedelta(delta: timedelta) -> str:
"""Pretty-prints a `timedelta`."""
seconds = delta.total_seconds()
days, seconds = divmod(seconds, 86400)
hours, seconds = divmod(seconds, 3600)
minutes, seconds = divmod(seconds, 60)
if days > 0:
return '%dd%dh%dm%ds' % (days, hours, minutes, seconds)
elif hours > 0:
return '%dh%dm%ds' % (hours, minutes, seconds)
elif minutes > 0:
return '%dm%ds' % (minutes, seconds)
else:
return '%ds' % (seconds,)
def to_yaml(input: dict) -> str:
"""Convert a dictionary to a pretty yaml representation."""
return yaml.dump(input, default_flow_style=None)
def get_hf_dataset_repo_id(hf_org: str, hf_space_name: str, namespace: str,
dataset_name: str) -> str:
"""Returns the repo name for a given dataset. This does not include the namespace."""
if hf_space_name == 'lilac':
# Don't include the space name for lilac datasets to shorten the linked dataset name.
return f'{hf_org}/{namespace}-{dataset_name}'
return f'{hf_org}/{hf_space_name}-{namespace}-{dataset_name}'