doberst113080 commited on
Commit
68c2936
·
verified ·
1 Parent(s): f4dcf14

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -12
README.md CHANGED
@@ -9,21 +9,11 @@ license: cc-by-sa-4.0
9
 
10
  **slim-sa-ner-3b-tool** is a 4_K_M quantized GGUF version of slim-sa-ner-3b, providing a small, fast inference implementation, optimized for multi-model concurrent deployment.
11
 
12
- This model combines two of the most popular traditional classifier capabilities (sentiment analysis and named entity recognition) and re-images them as function calls on a small specialized decoder LLM, generating output in the form of a python dictionary with keys corresponding to sentiment and NER identifiers.
13
-
14
 
15
  The intent of SLIMs is to forge a middle-ground between traditional encoder-based classifiers and open-ended API-based LLMs:
16
 
17
- Compared to Encoder Classifiers
18
- --With 4-bit quantization, this model stand-alone is 1.71 GB binary, which compares favorably with the combination of using FP32 versions of Roberta-Large for NER (1.42GB) and BERT for Sentiment (439 MB), which also require Pytorch and other external dependencies.
19
- --2.7B parameters vs. Roberta-Large (336M) and BERT (110M)
20
- --Provides intuitive natural language based responses that fit more naturally in LLM-based agent processes
21
- --One model can provide multiple modalities of classification, and can be trained using 'union' of two separate datasets
22
-
23
- Compared to Very Large Language Models
24
- --Runs locally with no API charges
25
- --Does not require complex prompt instructions
26
- --Small & Specialized - more limited capability but trained to do only 1-2 things well
27
 
28
 
29
  [**slim-sa-ner-3b**](https://huggingface.co/llmware/slim-sa-ner-3b) is part of the SLIM ("**S**tructured **L**anguage **I**nstruction **M**odel") series, providing a set of small, specialized decoder-based LLMs, fine-tuned for function-calling.
 
9
 
10
  **slim-sa-ner-3b-tool** is a 4_K_M quantized GGUF version of slim-sa-ner-3b, providing a small, fast inference implementation, optimized for multi-model concurrent deployment.
11
 
12
+ This model combines two of the most popular traditional classifier capabilities (**sentiment analysis** and **named entity recognition**) and re-images them as function calls on a small specialized decoder LLM, generating output in the form of a python dictionary with keys corresponding to sentiment and NER identifiers.
 
13
 
14
  The intent of SLIMs is to forge a middle-ground between traditional encoder-based classifiers and open-ended API-based LLMs:
15
 
16
+ The size of self-contained model binary is 1.71 GB, which is small enough to run locally on a CPU, and yet which comparables favorably with the use of two traditional FP32 versions of Roberta-Large for NER (1.42GB) and BERT for Sentiment Analysis (440 MB), while offering greater potential capacity depth with 2.7B parameters, and without the requirement of Pytorch and other external dependencies.
 
 
 
 
 
 
 
 
 
17
 
18
 
19
  [**slim-sa-ner-3b**](https://huggingface.co/llmware/slim-sa-ner-3b) is part of the SLIM ("**S**tructured **L**anguage **I**nstruction **M**odel") series, providing a set of small, specialized decoder-based LLMs, fine-tuned for function-calling.