doberst commited on
Commit
8d1c670
·
verified ·
1 Parent(s): c1d6128

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +11 -10
README.md CHANGED
@@ -7,12 +7,15 @@ inference: false
7
 
8
  <!-- Provide a quick summary of what the model is/does. -->
9
 
10
- **slim-sentiment** is part of the SLIM ("Structured Language Instruction Model") model series, providing a set of small, specialized decoder-based LLMs, fine-tuned for function-calling.
11
 
12
  slim-sentiment has been fine-tuned for **sentiment analysis** function calls, generating output consisting of a python dictionary corresponding to specified keys, e.g.:
13
 
14
- {"sentiment": ["positive"]}
15
-
 
 
 
16
  Each slim model has a corresponding 'tool' in a separate repository, e.g., [**'slim-sentiment-tool'**](https://huggingface.co/llmware/slim-sentiment-tool), which a 4-bit quantized gguf version of the model that is intended to be used for inference.
17
 
18
 
@@ -26,11 +29,6 @@ Each slim model has a corresponding 'tool' in a separate repository, e.g., [**'
26
  - **License:** Apache 2.0
27
  - **Finetuned from model:** Tiny Llama 1B
28
 
29
- ## Uses
30
-
31
- <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
32
-
33
- The intended use of SLIM models is to re-imagine traditional 'hard-coded' classifiers through the use of function calls, and to provide a natural language flexible tool that can be used as decision gates and processing steps in a complex LLM-based automation workflow.
34
 
35
  ## Prompt format:
36
 
@@ -41,7 +39,7 @@ The intended use of SLIM models is to re-imagine traditional 'hard-coded' classi
41
  "/n<bot>:"
42
 
43
  <details>
44
- <summary><b>Getting Started Example: </b> </summary>
45
 
46
  model = AutoModelForCausalLM.from_pretrained("llmware/slim-sentiment")
47
  tokenizer = AutoTokenizer.from_pretrained("llmware/slim-sentiment")
@@ -79,7 +77,8 @@ The intended use of SLIM models is to re-imagine traditional 'hard-coded' classi
79
  </details>
80
 
81
 
82
- ## Using as Function Call in LLMWare
 
83
 
84
  We envision the slim models deployed in a pipeline/workflow/templating framework that handles the prompt packaging more elegantly.
85
 
@@ -91,6 +90,8 @@ Check out llmware for one such implementation:
91
 
92
  print("llmware - llm_response: ", response)
93
 
 
 
94
 
95
  ## Model Card Contact
96
 
 
7
 
8
  <!-- Provide a quick summary of what the model is/does. -->
9
 
10
+ **slim-sentiment** is part of the SLIM ("Structured Language Instruction Model") model series, consisting of small, specialized decoder-based models, fine-tuned for function-calling.
11
 
12
  slim-sentiment has been fine-tuned for **sentiment analysis** function calls, generating output consisting of a python dictionary corresponding to specified keys, e.g.:
13
 
14
+ `{"sentiment": ["positive"]}`
15
+
16
+
17
+ SLIM models re-imagine traditional 'hard-coded' classifiers through the use of function calls, and to provide a natural language flexible tool that can be used as decision gates and processing steps in a complex LLM-based automation workflow.
18
+
19
  Each slim model has a corresponding 'tool' in a separate repository, e.g., [**'slim-sentiment-tool'**](https://huggingface.co/llmware/slim-sentiment-tool), which a 4-bit quantized gguf version of the model that is intended to be used for inference.
20
 
21
 
 
29
  - **License:** Apache 2.0
30
  - **Finetuned from model:** Tiny Llama 1B
31
 
 
 
 
 
 
32
 
33
  ## Prompt format:
34
 
 
39
  "/n<bot>:"
40
 
41
  <details>
42
+ <summary><b>Getting Started with Transformers Script </b> </summary>
43
 
44
  model = AutoModelForCausalLM.from_pretrained("llmware/slim-sentiment")
45
  tokenizer = AutoTokenizer.from_pretrained("llmware/slim-sentiment")
 
77
  </details>
78
 
79
 
80
+ <details>
81
+ <summary><b>Using as Function Call in LLMWare</b></summary>
82
 
83
  We envision the slim models deployed in a pipeline/workflow/templating framework that handles the prompt packaging more elegantly.
84
 
 
90
 
91
  print("llmware - llm_response: ", response)
92
 
93
+ </details>
94
+
95
 
96
  ## Model Card Contact
97