person-thumbs-up / README.md
asrimanth's picture
Update README.md
5fb03e6
|
raw
history blame
5.22 kB
metadata
title: Person Thumbs Up
emoji: 🐠
colorFrom: blue
colorTo: purple
sdk: streamlit
sdk_version: 1.21.0
app_file: app.py
pinned: false

Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference

Stable diffusion finetune using LoRA

HuggingFace Spaces URL: https://huggingface.co/spaces/asrimanth/person-thumbs-up

Please note that the app on spaces is very slow due to compute constraints. For good results, please try locally.

Approach

The key resource in this endeavor: https://huggingface.co/blog/lora

Training

All of the following models were trained on stable-diffusion-v1-5

  • Several different training strategies and found LoRA to be the best for my needs.
  • In the dataset, the thumbs up dataset had 121 images for training, which I found to be adequate.
  • First, I scraped ~50 images of "sachin tendulkar". This experiment failed, since the model gave a player with cricket helmet.
  • For training on "Tom cruise", I've scraped ~100 images from images.google.com, using the javascript code from pyimagesearch.com
  • For training on "srimanth", I've put 50 images of myself.

For the datasets, I started as follows:

  • Use an image captioning model from HuggingFace - In our case it is the Salesforce/blip-image-captioning-large model.
  • Once captioned, If the caption has "thumbs up", we replace it with #thumbsup, otherwise we attach the word #thumbsup to the caption.
  • If the model recognizes the person or says the word "man", we replace it with <person>. Otherwise, we attach the word <person> to the caption.
  • No-cap dataset: For the no-cap models, we don't use the captioning models. We simply add the <person> and the #thumbsup tag.
  • Plain dataset: For the plain models, we leave the words as is - the "thumbs up" and the person name are without special characters.

The wandb dashboard for the models are as follows: Initial experiments: I've tried training only on the thumbs up first. The results were good. The thumbs up was mostly accurate, with 4 fingers folded and the thumb raised. However, the model trained on sachin had several issues, including occlusion by cricket gear. I've tried several different learning rates (from 1e-4 to 1e-6 with cosine scheduler), but the loss curve did not change much. Number of epochs : 50-60 Augmentations used : Center crop, Random Flip Gradient accumulation steps : Tried 1, 3, and 4 for different experiments. 4 gave decent results.

text2image_fine-tune :

wandb dashboard : https://wandb.ai/asrimanth/text2image_fine-tune

Model card for asrimanth/person-thumbs-up-lora: https://huggingface.co/asrimanth/person-thumbs-up-lora

Prompt: <tom_cruise> #thumbsup

Deployed models:

When the above experiment failed, I had to try different datasets. One of them was "tom cruise".

srimanth-thumbs-up-lora-plain : We use the plain dataset with srimanth mentioned above.

wandb link: https://wandb.ai/asrimanth/srimanth-thumbs-up-lora-plain

Model card for srimanth-thumbs-up-lora-plain: https://huggingface.co/asrimanth/srimanth-thumbs-up-lora-plain

Prompt: srimanth thumbs up

person-thumbs-up-plain-lora wandb : We use the plain dataset with tom cruise images.

wandb link: https://wandb.ai/asrimanth/person-thumbs-up-plain-lora

Model card for asrimanth/person-thumbs-up-plain-lora: https://huggingface.co/asrimanth/person-thumbs-up-plain-lora

Prompt: tom cruise thumbs up

person-thumbs-up-lora-no-cap wandb dashboard: We use the no-cap dataset with tom cruise images.

https://wandb.ai/asrimanth/person-thumbs-up-lora-no-cap

Model card for asrimanth/person-thumbs-up-lora-no-cap: https://huggingface.co/asrimanth/person-thumbs-up-lora-no-cap

Prompt: <tom_cruise> #thumbsup

Inference

  • Inference works best for 25 steps in the pipeline.
  • Since the huggingface space built by Streamlit is slow due to low compute, please perform local inference using GPU.
  • During local inference (25 steps), I found the person-thumbs-up-plain-lora to show 35 out of 50 images with a decent thumbs up result for tom cruise, 5 incomplete thumbs up.
  • While I could not evaluate the model with metrics due to insufficient time, I chose the visual approach. To view the inference images, check the results folder.
  • To evaulate diffusion models, I would start with this: https://huggingface.co/docs/diffusers/conceptual/evaluation
  • The half-precision inference was not working on CPU, so we've used torch.float32 instead.

Deployment

To run inference locally, choose a model and run the command:

python3 inference.py

To run the streamlit app locally, run the command:

streamlit run app.py
  • I chose streamlit to deploy the application on HuggingFace spaces. It was developer friendly and the app logic can be found in app.py
  • Streamlit app would be a great choice for an MVP.
  • AWS sagemaker would be a good choice for deploying models, since it supports huggingface models with minimal friction.
  • A docker container orchestrated in a kubernetes cluster would be ideal.
  • In practice, evaluation of models in real-time would let us know if there is model drift and retrain accordingly.