Spaces:
Sleeping
Sleeping
title: Llama Nallm Test | |
emoji: 💬 | |
colorFrom: yellow | |
colorTo: purple | |
sdk: gradio | |
sdk_version: 4.36.1 | |
app_file: app.py | |
pinned: false | |
license: mit | |
A simple interface for testing out NA-LLM-qa model based on Llama-3-8B checkpoint. | |
# Use | |
```sh | |
python -m venv .venv | |
source .venv/bin/activate | |
pip install -r requirements.txt | |
gradio app.py | |
``` | |
# Notes | |
- The model is hosted with HuggingFace Inference Endpoint. | |
- The Endpoint may be paused due to inactivity. In that case, calling a signal will "wake up" the endpoint, but it would take around several minutes. | |
- For now, the endpoint is gated. Set appropriate `hf_token` with READ permission to the organization. | |
- Input filtering | |
- The model performs unexpectedly for non-questions | |
- For this reason, a simple SVM-based filter is applied | |
- The filter is a `OneClassSVM` trained with question sections of na-llm | |
- The model, along with corresponding vectorizer, is saved in `question_undetector.pkl` as `(vectorizer, model)` object. | |
- The hosting machine should be powerful enough at least to run a simple SVM model | |