|
--- |
|
license: llama2 |
|
library_name: transformers |
|
tags: |
|
- llama |
|
- merge |
|
datasets: |
|
- GBaker/MedQA-USMLE-4-options |
|
- cognitivecomputations/samantha-data |
|
- shibing624/medical |
|
--- |
|
|
|
# Dr. Samantha |
|
|
|
<p align="center"> |
|
<img src="https://huggingface.co/sethuiyer/Dr_Samantha-7b/resolve/main/dr_samantha_anime_style_reduced_quality.webp" height="256px" alt="SynthIQ"> |
|
</p> |
|
|
|
## Overview |
|
|
|
Dr. Samantha is a language model made by merging `Severus27/BeingWell_llama2_7b` and `ParthasarathyShanmugam/llama-2-7b-samantha` using [mergekit](https://github.com/cg123/mergekit). |
|
|
|
Has capabilities of a medical knowledge-focused model (trained on USMLE databases and doctor-patient interactions) with the philosophical, psychological, and relational understanding of the Samantha-7b model. |
|
|
|
As both a medical consultant and personal counselor, Dr.Samantha could effectively support both physical and mental wellbeing - important for whole-person care. |
|
|
|
|
|
# Yaml Config |
|
|
|
```yaml |
|
|
|
slices: |
|
- sources: |
|
- model: Severus27/BeingWell_llama2_7b |
|
layer_range: [0, 32] |
|
- model: ParthasarathyShanmugam/llama-2-7b-samantha |
|
layer_range: [0, 32] |
|
|
|
merge_method: slerp |
|
base_model: TinyPixel/Llama-2-7B-bf16-sharded |
|
|
|
parameters: |
|
t: |
|
- filter: self_attn |
|
value: [0, 0.5, 0.3, 0.7, 1] |
|
- filter: mlp |
|
value: [1, 0.5, 0.7, 0.3, 0] |
|
- value: 0.5 # fallback for rest of tensors |
|
tokenizer_source: union |
|
|
|
dtype: bfloat16 |
|
|
|
``` |