File size: 1,604 Bytes
d2d4a3e
77eaf74
d2d4a3e
 
 
 
 
 
 
fa7c378
 
 
 
 
 
 
 
 
d2d4a3e
fa7c378
 
 
d2d4a3e
fa7c378
 
d2d4a3e
 
fa7c378
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
d2d4a3e
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
---
base_model: MaziyarPanahi/Llama-3-16B-Instruct-v0.1
library_name: transformers
tags:
- 4-bit
- AWQ
- text-generation
- autotrain_compatible
- endpoints_compatible
  - mergekit
  - merge
  - facebook
  - meta
  - pytorch
  - llama
  - llama-3
language:
  - en
pipeline_tag: text-generation
license: other
license_name: llama3
license_link: LICENSE
inference: false
model_creator: MaziyarPanahi
model_name: Llama-3-16B-Instruct-v0.1
quantized_by: Suparious
---
# MaziyarPanahi/Llama-3-16B-Instruct-v0.1 AWQ

- Model creator: [MaziyarPanahi](https://huggingface.co/MaziyarPanahi)
- Original model: [Llama-3-16B-Instruct-v0.1](https://huggingface.co/MaziyarPanahi/Llama-3-16B-Instruct-v0.1)

<img src="./llama-3-merges.webp" alt="Goku 8x22B v0.1 Logo" width="500" style="margin-left:'auto' margin-right:'auto' display:'block'"/>

## Model Summary

This model is a self-merge of `MaziyarPanahi/Llama-3-11B-Instruct-v0.1` model.

## Prompt template

```text
<|begin_of_text|><|start_header_id|>system<|end_header_id|>

You are a helpful assistant.<|eot_id|><|start_header_id|>user<|end_header_id|>

what's 25-4*2+3<|eot_id|><|start_header_id|>assistant<|end_header_id|>

To evaluate this expression, we need to follow the order of operations (PEMDAS):

1. First, multiply 4 and 2: 4*2 = 8
2. Then, subtract 8 from 25: 25 - 8 = 17
3. Finally, add 3: 17 + 3 = 20

So, 25-4*2+3 = 20!<|eot_id|>
To evaluate this expression, we need to follow the order of operations (PEMDAS):

1. First, multiply 4 and 2: 4*2 = 8
2. Then, subtract 8 from 25: 25 - 8 = 17
3. Finally, add 3: 17 + 3 = 20

So, 25-4*2+3 = 20!
```