doberst113080 commited on
Commit
087687b
·
verified ·
1 Parent(s): 68c2936

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -11
README.md CHANGED
@@ -11,9 +11,9 @@ license: cc-by-sa-4.0
11
 
12
  This model combines two of the most popular traditional classifier capabilities (**sentiment analysis** and **named entity recognition**) and re-images them as function calls on a small specialized decoder LLM, generating output in the form of a python dictionary with keys corresponding to sentiment and NER identifiers.
13
 
14
- The intent of SLIMs is to forge a middle-ground between traditional encoder-based classifiers and open-ended API-based LLMs:
15
 
16
- The size of self-contained model binary is 1.71 GB, which is small enough to run locally on a CPU, and yet which comparables favorably with the use of two traditional FP32 versions of Roberta-Large for NER (1.42GB) and BERT for Sentiment Analysis (440 MB), while offering greater potential capacity depth with 2.7B parameters, and without the requirement of Pytorch and other external dependencies.
17
 
18
 
19
  [**slim-sa-ner-3b**](https://huggingface.co/llmware/slim-sa-ner-3b) is part of the SLIM ("**S**tructured **L**anguage **I**nstruction **M**odel") series, providing a set of small, specialized decoder-based LLMs, fine-tuned for function-calling.
@@ -36,15 +36,6 @@ Load in your favorite GGUF inference engine, or try with llmware as follows:
36
  ModelCatalog().tool_test_run("slim-sa-ner-3b-tool", verbose=True)
37
 
38
 
39
- Slim models can also be loaded even more simply as part of a multi-model, multi-step LLMfx calls:
40
-
41
- from llmware.agents import LLMfx
42
-
43
- llm_fx = LLMfx()
44
- llm_fx.load_tool("sa-ner")
45
- response = llm_fx.sa_ner(text)
46
-
47
-
48
  Note: please review [**config.json**](https://huggingface.co/llmware/slim-sa-ner-3b-tool/blob/main/config.json) in the repository for prompt wrapping information, details on the model, and full test set.
49
 
50
 
 
11
 
12
  This model combines two of the most popular traditional classifier capabilities (**sentiment analysis** and **named entity recognition**) and re-images them as function calls on a small specialized decoder LLM, generating output in the form of a python dictionary with keys corresponding to sentiment and NER identifiers.
13
 
14
+ The intent of SLIMs is to forge a middle-ground between traditional encoder-based classifiers and open-ended API-based LLMs.
15
 
16
+ The size of the self-contained GGUF model binary is 1.71 GB, which is small enough to run locally on a CPU, and yet which comparables favorably with the use of two traditional FP32 versions of Roberta-Large for NER (1.42GB) and BERT for Sentiment Analysis (440 MB), while offering greater potential capacity depth with 2.7B parameters, and without the requirement of Pytorch and other external dependencies.
17
 
18
 
19
  [**slim-sa-ner-3b**](https://huggingface.co/llmware/slim-sa-ner-3b) is part of the SLIM ("**S**tructured **L**anguage **I**nstruction **M**odel") series, providing a set of small, specialized decoder-based LLMs, fine-tuned for function-calling.
 
36
  ModelCatalog().tool_test_run("slim-sa-ner-3b-tool", verbose=True)
37
 
38
 
 
 
 
 
 
 
 
 
 
39
  Note: please review [**config.json**](https://huggingface.co/llmware/slim-sa-ner-3b-tool/blob/main/config.json) in the repository for prompt wrapping information, details on the model, and full test set.
40
 
41