Tymec's picture
Tokenization rework
2c1f9dd
raw
history blame
7.64 kB
from __future__ import annotations
from pathlib import Path
from typing import Literal
import click
__all__ = ["cli_wrapper"]
DONE_STR = click.style("DONE", fg="green")
@click.group()
def cli() -> None: ...
@cli.command()
@click.option(
"--model",
"model_path",
required=True,
help="Path to the trained model",
type=click.Path(exists=True, file_okay=True, dir_okay=False, readable=True, resolve_path=True, path_type=Path),
)
@click.option(
"--share/--no-share",
default=False,
help="Whether to create a shareable link",
)
def gui(model_path: Path, share: bool) -> None:
"""Launch the Gradio GUI"""
import os
from app.gui import launch_gui
os.environ["MODEL_PATH"] = model_path.as_posix()
launch_gui(share)
@cli.command()
@click.option(
"--model",
"model_path",
required=True,
help="Path to the trained model",
type=click.Path(exists=True, file_okay=True, dir_okay=False, readable=True, resolve_path=True, path_type=Path),
)
@click.argument("text", nargs=-1)
def predict(model_path: Path, text: list[str]) -> None:
"""Perform sentiment analysis on the provided text.
Note: Piped input takes precedence over the text argument
"""
import sys
import joblib
from app.model import infer_model
text = " ".join(text).strip()
if not sys.stdin.isatty():
piped_text = sys.stdin.read().strip()
text = piped_text or text
if not text:
msg = "No text provided"
raise click.UsageError(msg)
click.echo("Loading model... ", nl=False)
model = joblib.load(model_path)
click.echo(DONE_STR)
click.echo("Performing sentiment analysis... ", nl=False)
prediction = infer_model(model, [text])[0]
# prediction = model.predict([text])[0]
if prediction == 0:
sentiment = click.style("NEGATIVE", fg="red")
elif prediction == 1:
sentiment = click.style("POSITIVE", fg="green")
else:
sentiment = click.style("NEUTRAL", fg="yellow")
click.echo(sentiment)
@cli.command()
@click.option(
"--dataset",
default="test",
help="Dataset to evaluate the model on",
type=click.Choice(["test", "sentiment140", "amazonreviews", "imdb50k"]),
)
@click.option(
"--model",
"model_path",
required=True,
help="Path to the trained model",
type=click.Path(exists=True, file_okay=True, dir_okay=False, readable=True, resolve_path=True, path_type=Path),
)
@click.option(
"--cv",
default=5,
help="Number of cross-validation folds",
show_default=True,
type=click.IntRange(1, 50),
)
@click.option(
"--batch-size",
default=512,
help="Size of the batches used in tokenization",
show_default=True,
)
@click.option(
"--processes",
default=8,
help="Number of parallel jobs during tokenization",
show_default=True,
)
@click.option(
"--verbose",
is_flag=True,
help="Show verbose output",
)
def evaluate(
dataset: Literal["test", "sentiment140", "amazonreviews", "imdb50k"],
model_path: Path,
cv: int,
batch_size: int,
processes: int,
verbose: bool,
) -> None:
"""Evaluate the model on the the specified dataset"""
import joblib
from app.constants import CACHE_DIR
from app.data import load_data, tokenize
from app.model import evaluate_model
cached_data_path = CACHE_DIR / f"{dataset}_tokenized.pkl"
use_cached_data = False
if cached_data_path.exists():
use_cached_data = click.confirm(f"Found existing tokenized data for '{dataset}'. Use it?", default=True)
if use_cached_data:
click.echo("Loading cached data... ", nl=False)
token_data, label_data = joblib.load(cached_data_path)
click.echo(DONE_STR)
else:
click.echo("Loading dataset... ", nl=False)
text_data, label_data = load_data(dataset)
click.echo(DONE_STR)
click.echo("Tokenizing data... ", nl=False)
token_data = tokenize(text_data, batch_size=batch_size, n_jobs=processes, show_progress=True)
joblib.dump((token_data, label_data), cached_data_path, compress=3)
click.echo(DONE_STR)
del text_data
click.echo("Loading model... ", nl=False)
model = joblib.load(model_path)
click.echo(DONE_STR)
click.echo("Evaluating model... ", nl=False)
acc_mean, acc_std = evaluate_model(model, token_data, label_data, folds=cv, verbose=verbose)
click.secho(f"{acc_mean:.2%} ± {acc_std:.2%}", fg="blue")
@cli.command()
@click.option(
"--dataset",
required=True,
help="Dataset to train the model on",
type=click.Choice(["sentiment140", "amazonreviews", "imdb50k"]),
)
@click.option(
"--max-features",
default=20000,
help="Maximum number of features",
show_default=True,
type=click.IntRange(1, None),
)
@click.option(
"--cv",
default=5,
help="Number of cross-validation folds",
show_default=True,
type=click.IntRange(1, 50),
)
@click.option(
"--batch-size",
default=512,
help="Size of the batches used in tokenization",
show_default=True,
)
@click.option(
"--processes",
default=8,
help="Number of parallel jobs during tokenization",
show_default=True,
)
@click.option(
"--seed",
default=42,
help="Random seed (-1 for random seed)",
show_default=True,
type=click.IntRange(-1, None),
)
@click.option(
"--force",
is_flag=True,
help="Overwrite the model file if it already exists",
)
@click.option(
"--verbose",
is_flag=True,
help="Show verbose output",
)
def train(
dataset: Literal["sentiment140", "amazonreviews", "imdb50k"],
max_features: int,
cv: int,
batch_size: int,
processes: int,
seed: int,
force: bool,
verbose: bool,
) -> None:
"""Train the model on the provided dataset"""
import joblib
from app.constants import CACHE_DIR, MODELS_DIR
from app.data import load_data, tokenize
from app.model import create_model, train_model
model_path = MODELS_DIR / f"{dataset}_tfidf_ft-{max_features}.pkl"
if model_path.exists() and not force:
click.confirm(f"Model file '{model_path}' already exists. Overwrite?", abort=True)
cached_data_path = CACHE_DIR / f"{dataset}_tokenized.pkl"
use_cached_data = False
if cached_data_path.exists():
use_cached_data = click.confirm(f"Found existing tokenized data for '{dataset}'. Use it?", default=True)
if use_cached_data:
click.echo("Loading cached data... ", nl=False)
token_data, label_data = joblib.load(cached_data_path)
click.echo(DONE_STR)
else:
click.echo("Loading dataset... ", nl=False)
text_data, label_data = load_data(dataset)
click.echo(DONE_STR)
click.echo("Tokenizing data... ", nl=False)
token_data = tokenize(text_data, batch_size=batch_size, n_jobs=processes, show_progress=True)
joblib.dump((token_data, label_data), cached_data_path, compress=3)
click.echo(DONE_STR)
del text_data
click.echo("Training model... ")
model = create_model(max_features, seed=None if seed == -1 else seed, verbose=verbose)
trained_model, accuracy = train_model(model, token_data, label_data, folds=cv, seed=seed, verbose=verbose)
click.echo("Model accuracy: ", nl=False)
click.secho(f"{accuracy:.2%}", fg="blue")
click.echo("Model saved to: ", nl=False)
joblib.dump(trained_model, model_path, compress=3)
click.secho(str(model_path), fg="blue")
def cli_wrapper() -> None:
cli(max_content_width=120)
if __name__ == "__main__":
cli_wrapper()