id
stringlengths
7
117
author
stringclasses
6 values
sha
null
created_at
unknown
last_modified
null
disabled
null
downloads
int64
0
18.6M
downloads_all_time
null
gated
bool
1 class
gguf
null
inference
null
likes
int64
0
4.77k
library_name
stringclasses
36 values
tags
sequencelengths
1
430
pipeline_tag
stringclasses
32 values
mask_token
null
model_index
null
trending_score
int64
0
132
architectures
sequencelengths
1
5
bos_token_id
int64
-1
256k
eos_token_id
int64
-1
256k
hidden_act
stringclasses
15 values
hidden_size
int64
1
20.5k
initializer_range
float64
0
1
intermediate_size
int64
1
98.3k
max_position_embeddings
int64
8
1.05M
model_type
stringclasses
530 values
num_attention_heads
int64
1
5k
num_hidden_layers
int64
-1
8.93k
num_key_value_heads
int64
1
160
rms_norm_eps
float64
0
7
rope_theta
float64
1k
1,000B
sliding_window
int64
0
262k
tie_word_embeddings
bool
2 classes
torch_dtype
stringclasses
8 values
transformers_version
stringclasses
207 values
use_cache
bool
2 classes
vocab_size
int64
-1
5.03M
attention_bias
bool
2 classes
attention_dropout
float64
0
0.5
head_dim
int64
2
256
mlp_bias
bool
2 classes
pretraining_tp
int64
0
8
rope_scaling
dict
tastypear/CausalLM-7B-DPO-alpha-GGUF
null
null
"2023-11-19T15:36:16Z"
null
null
681
null
null
null
null
26
transformers
[ "transformers", "gguf", "llama", "llama2", "qwen", "text-generation", "en", "zh", "dataset:JosephusCheung/GuanacoDataset", "dataset:Open-Orca/OpenOrca", "dataset:stingning/ultrachat", "dataset:meta-math/MetaMathQA", "dataset:liuhaotian/LLaVA-Instruct-150K", "dataset:jondurbin/airoboros-3.1", "dataset:WizardLM/WizardLM_evol_instruct_V2_196k", "dataset:RyokoAI/ShareGPT52K", "dataset:RyokoAI/Fandom23K", "dataset:milashkaarshif/MoeGirlPedia_wikitext_raw_archive", "dataset:wikipedia", "dataset:wiki_lingua", "dataset:fnlp/moss-003-sft-data", "dataset:garage-bAInd/Open-Platypus", "dataset:LDJnr/Puffin", "dataset:openbmb/llava_zh", "dataset:BAAI/COIG", "dataset:TigerResearch/tigerbot-zhihu-zh-10k", "dataset:liwu/MNBVC", "dataset:teknium/openhermes", "base_model:CausalLM/7B-DPO-alpha", "base_model:quantized:CausalLM/7B-DPO-alpha", "license:wtfpl", "region:us" ]
text-generation
null
null
1
null
null
null
null
null
null
null
null
llama
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
pansophic/rocket-3B
null
null
"2023-11-19T17:14:29Z"
null
null
275
null
null
null
null
82
transformers
[ "transformers", "pytorch", "safetensors", "stablelm", "text-generation", "en", "arxiv:2305.18290", "arxiv:2101.00027", "arxiv:2305.06161", "base_model:stabilityai/stablelm-3b-4e1t", "base_model:finetune:stabilityai/stablelm-3b-4e1t", "license:cc-by-sa-4.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "StableLmForCausalLM" ]
0
50,279
silu
2,560
0.02
6,912
4,096
stablelm
32
32
32
null
10,000
null
false
bfloat16
4.38.0
true
50,304
null
null
null
null
null
null
TinyLlama/TinyLlama-1.1B-Chat-v0.6
null
null
"2023-11-20T08:59:23Z"
null
null
356,751
null
null
null
null
89
transformers
[ "transformers", "safetensors", "gguf", "llama", "text-generation", "conversational", "en", "dataset:cerebras/SlimPajama-627B", "dataset:bigcode/starcoderdata", "dataset:OpenAssistant/oasst_top1_2023-08-25", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
1
2
silu
2,048
0.02
5,632
2,048
llama
32
22
4
0.00001
10,000
null
false
bfloat16
4.35.0
false
32,000
false
null
null
null
1
null
allenai/digital-socrates-13b
null
null
"2023-11-21T01:55:26Z"
null
null
845
null
null
null
null
10
transformers
[ "transformers", "pytorch", "llama", "text-generation", "en", "arxiv:2311.09613", "license:apache-2.0", "model-index", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
1
2
silu
5,120
0.02
13,824
4,096
llama
40
40
40
0.00001
10,000
null
false
float32
4.35.0.dev0
true
32,001
false
null
null
null
1
null
TheBloke/Qwen-7B-Chat-AWQ
null
null
"2023-11-21T09:54:47Z"
null
null
110
null
null
null
null
8
transformers
[ "transformers", "safetensors", "qwen", "text-generation", "custom_code", "zh", "en", "arxiv:2309.16609", "arxiv:2305.08322", "arxiv:2009.03300", "arxiv:2305.05280", "arxiv:2210.03629", "base_model:Qwen/Qwen-7B-Chat", "base_model:quantized:Qwen/Qwen-7B-Chat", "autotrain_compatible", "4-bit", "awq", "region:us" ]
text-generation
null
null
1
[ "QWenLMHeadModel" ]
null
null
null
4,096
0.02
22,016
8,192
qwen
32
32
null
null
null
null
false
bfloat16
4.35.0
true
151,936
null
null
null
null
1
null
VamsiPranav/language-training
null
null
"2023-11-21T13:25:15Z"
null
null
156
null
null
null
null
1
transformers
[ "transformers", "tensorboard", "safetensors", "gpt2", "text-generation", "generated_from_trainer", "base_model:openai-community/gpt2", "base_model:finetune:openai-community/gpt2", "license:mit", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "GPT2LMHeadModel" ]
0
0
null
null
0.02
null
null
gpt2
null
null
null
null
null
null
null
float32
4.35.2
true
4,037
null
null
null
null
null
null
togethercomputer/StripedHyena-Hessian-7B
null
null
"2023-11-21T15:43:25Z"
null
null
93
null
null
null
null
62
transformers
[ "transformers", "safetensors", "stripedhyena", "text-generation", "custom_code", "en", "arxiv:2302.10866", "arxiv:2310.18780", "arxiv:2311.05908", "license:apache-2.0", "autotrain_compatible", "region:us" ]
text-generation
null
null
1
[ "StripedHyenaModelForCausalLM" ]
null
null
null
4,096
null
null
null
stripedhyena
32
null
null
null
null
null
null
bfloat16
null
true
32,000
null
null
null
null
null
null
stabilityai/stablelm-zephyr-3b
null
null
"2023-11-21T16:25:10Z"
null
null
9,614
null
null
null
null
247
transformers
[ "transformers", "safetensors", "stablelm", "text-generation", "causal-lm", "conversational", "en", "dataset:HuggingFaceH4/ultrachat_200k", "dataset:HuggingFaceH4/ultrafeedback_binarized", "dataset:meta-math/MetaMathQA", "dataset:WizardLM/WizardLM_evol_instruct_V2_196k", "dataset:Intel/orca_dpo_pairs", "arxiv:2305.18290", "arxiv:2306.05685", "license:other", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "StableLmForCausalLM" ]
0
0
silu
2,560
0.02
6,912
4,096
stablelm
32
32
32
null
10,000
null
false
bfloat16
4.38.0
true
50,304
null
null
null
null
null
null
FPHam/Karen_TheEditor_V2_CREATIVE_Mistral_7B
null
null
"2023-11-21T20:52:15Z"
null
null
375
null
null
null
null
22
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "llm", "llama", "spellcheck", "grammar", "conversational", "license:llama2", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "MistralForCausalLM" ]
1
32,000
silu
4,096
0.02
14,336
32,768
mistral
32
32
8
0.00001
10,000
4,096
false
float16
4.34.1
true
32,002
null
null
null
null
null
null
echo840/Monkey
null
null
"2023-11-22T09:29:28Z"
null
null
904
null
null
null
null
30
transformers
[ "transformers", "pytorch", "monkey", "text-generation", "custom_code", "arxiv:2311.06607", "autotrain_compatible", "region:us" ]
text-generation
null
null
1
[ "MonkeyLMHeadModel" ]
null
null
null
4,096
0.02
22,016
8,192
monkey
32
32
null
null
null
null
false
bfloat16
4.32.0
false
151,936
null
null
null
null
null
null
TheBloke/Synatra-RP-Orca-2-7B-v0.1-GPTQ
null
null
"2023-11-23T17:11:49Z"
null
null
29
null
null
null
null
2
transformers
[ "transformers", "safetensors", "llama", "text-generation", "base_model:maywell/Synatra-RP-Orca-2-7b-v0.1", "base_model:quantized:maywell/Synatra-RP-Orca-2-7b-v0.1", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "4-bit", "gptq", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
1
2
silu
4,096
0.02
11,008
4,096
llama
32
32
32
0.00001
10,000
null
false
bfloat16
4.35.2
true
32,003
false
null
null
null
1
null
second-state/Yi-34B-Chat-GGUF
null
null
"2023-11-24T02:31:29Z"
null
null
392
null
null
null
null
2
transformers
[ "transformers", "gguf", "llama", "text-generation", "base_model:01-ai/Yi-34B-Chat", "base_model:quantized:01-ai/Yi-34B-Chat", "license:apache-2.0", "autotrain_compatible", "region:us", "conversational" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
1
2
silu
7,168
0.02
20,480
4,096
llama
56
60
8
0.00001
5,000,000
null
false
bfloat16
4.35.0
true
64,000
false
null
null
null
1
null
ostorc/Conversational_Spanish_GPT
null
null
"2023-11-24T16:37:04Z"
null
null
791
null
null
null
null
7
transformers
[ "transformers", "safetensors", "gpt2", "text-generation", "chatbot", "conversational", "es", "base_model:microsoft/DialoGPT-small", "base_model:finetune:microsoft/DialoGPT-small", "autotrain_compatible", "text-generation-inference", "region:us" ]
text-generation
null
null
1
[ "GPT2LMHeadModel" ]
50,256
50,256
null
null
0.02
null
null
gpt2
null
null
null
null
null
null
null
float32
4.35.2
true
50,257
null
null
null
null
null
null
VAGOsolutions/SauerkrautLM-7b-HerO
null
null
"2023-11-24T16:55:27Z"
null
null
882
null
null
null
null
32
transformers
[ "transformers", "pytorch", "safetensors", "mistral", "text-generation", "finetune", "chatml", "augmentation", "german", "merge", "mergekit", "conversational", "en", "de", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "MistralForCausalLM" ]
1
32,000
silu
4,096
0.02
14,336
32,768
mistral
32
32
8
0.00001
10,000
4,096
false
float16
4.35.1
false
32,002
null
null
null
null
null
null
TheBloke/Rose-20B-GGUF
null
null
"2023-11-24T20:34:27Z"
null
null
435
null
null
null
null
25
transformers
[ "transformers", "gguf", "llama", "text-generation-inference", "instruct", "text-generation", "en", "base_model:tavtav/Rose-20B", "base_model:quantized:tavtav/Rose-20B", "license:llama2", "region:us" ]
text-generation
null
null
1
null
null
null
null
null
null
null
null
llama
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
tokyotech-llm/Swallow-70b-hf
null
null
"2023-11-25T02:13:02Z"
null
null
1,451
null
null
null
null
9
transformers
[ "transformers", "pytorch", "llama", "text-generation", "en", "ja", "arxiv:2404.17790", "arxiv:2404.17733", "license:llama2", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
1
2
silu
8,192
0.02
28,672
4,096
llama
64
80
8
0.00001
null
null
false
bfloat16
4.33.2
true
43,176
null
null
null
null
1
null
berkeley-nest/Starling-LM-7B-alpha
null
null
"2023-11-25T17:42:15Z"
null
null
40,678
null
null
null
null
553
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "reward model", "RLHF", "RLAIF", "conversational", "en", "dataset:berkeley-nest/Nectar", "arxiv:2306.02231", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "MistralForCausalLM" ]
1
32,000
silu
4,096
0.02
14,336
8,192
mistral
32
32
8
0.00001
10,000
4,096
false
bfloat16
4.35.0
true
32,002
null
null
null
null
null
null
GeneZC/MiniChat-1.5-3B
null
null
"2023-11-26T03:37:49Z"
null
null
1,478
null
null
null
null
35
transformers
[ "transformers", "pytorch", "safetensors", "llama", "text-generation", "conversational", "en", "zh", "arxiv:2311.07052", "arxiv:2310.05914", "arxiv:2305.18290", "license:apache-2.0", "model-index", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
1
2
silu
3,072
0.02
8,192
4,096
llama
24
24
24
0.00001
10,000
null
false
float16
4.33.2
true
49,216
null
null
null
null
1
null
listen2you002/ChartLlama-13b
null
null
"2023-11-27T06:57:18Z"
null
null
132
null
null
null
null
15
transformers
[ "transformers", "llava", "text-generation", "en", "dataset:listen2you002/ChartLlama-Dataset", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "LlavaLlamaForCausalLM" ]
1
2
silu
5,120
0.02
13,824
4,096
llava
40
40
40
0.00001
null
null
false
float16
4.31.0
true
32,000
null
null
null
null
1
null
mesolitica/mallam-1.1B-4096
null
null
"2023-11-27T14:33:33Z"
null
null
486
null
null
null
null
5
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "ms", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "MistralForCausalLM" ]
1
2
silu
2,048
0.02
5,632
32,768
mistral
32
22
8
0.00001
10,000
4,096
false
bfloat16
4.36.0.dev0
true
32,000
null
0
null
null
null
null
google/madlad400-8b-lm
null
null
"2023-11-27T16:00:07Z"
null
null
253
null
null
null
null
4
transformers
[ "transformers", "safetensors", "t5", "text-generation", "text-generation-inference", "custom_code", "en", "ru", "es", "fr", "de", "it", "pt", "pl", "nl", "vi", "tr", "sv", "id", "ro", "cs", "zh", "hu", "ja", "th", "fi", "fa", "uk", "da", "el", "no", "bg", "sk", "ko", "ar", "lt", "ca", "sl", "he", "et", "lv", "hi", "sq", "ms", "az", "sr", "ta", "hr", "kk", "is", "ml", "mr", "te", "af", "gl", "fil", "be", "mk", "eu", "bn", "ka", "mn", "bs", "uz", "ur", "sw", "yue", "ne", "kn", "kaa", "gu", "si", "cy", "eo", "la", "hy", "ky", "tg", "ga", "mt", "my", "km", "tt", "so", "ku", "ps", "pa", "rw", "lo", "ha", "dv", "fy", "lb", "ckb", "mg", "gd", "am", "ug", "ht", "grc", "hmn", "sd", "jv", "mi", "tk", "ceb", "yi", "ba", "fo", "or", "xh", "su", "kl", "ny", "sm", "sn", "co", "zu", "ig", "yo", "pap", "st", "haw", "as", "oc", "cv", "lus", "tet", "gsw", "sah", "br", "rm", "sa", "bo", "om", "se", "ce", "cnh", "ilo", "hil", "udm", "os", "lg", "ti", "vec", "ts", "tyv", "kbd", "ee", "iba", "av", "kha", "to", "tn", "nso", "fj", "zza", "ak", "ada", "otq", "dz", "bua", "cfm", "ln", "chm", "gn", "krc", "wa", "hif", "yua", "srn", "war", "rom", "bik", "pam", "sg", "lu", "ady", "kbp", "syr", "ltg", "myv", "iso", "kac", "bho", "ay", "kum", "qu", "za", "pag", "ngu", "ve", "pck", "zap", "tyz", "hui", "bbc", "tzo", "tiv", "ksd", "gom", "min", "ang", "nhe", "bgp", "nzi", "nnb", "nv", "zxx", "bci", "kv", "new", "mps", "alt", "meu", "bew", "fon", "iu", "abt", "mgh", "mnw", "tvl", "dov", "tlh", "ho", "kw", "mrj", "meo", "crh", "mbt", "emp", "ace", "ium", "mam", "gym", "mai", "crs", "pon", "ubu", "fip", "quc", "gv", "kj", "btx", "ape", "chk", "rcf", "shn", "tzh", "mdf", "ppk", "ss", "gag", "cab", "kri", "seh", "ibb", "tbz", "bru", "enq", "ach", "cuk", "kmb", "wo", "kek", "qub", "tab", "bts", "kos", "rwo", "cak", "tuc", "bum", "cjk", "gil", "stq", "tsg", "quh", "mak", "arn", "ban", "jiv", "sja", "yap", "tcy", "toj", "twu", "xal", "amu", "rmc", "hus", "nia", "kjh", "bm", "guh", "mas", "acf", "dtp", "ksw", "bzj", "din", "zne", "mad", "msi", "mag", "mkn", "kg", "lhu", "ch", "qvi", "mh", "djk", "sus", "mfe", "srm", "dyu", "ctu", "gui", "pau", "inb", "bi", "mni", "guc", "jam", "wal", "jac", "bas", "gor", "skr", "nyu", "noa", "sda", "gub", "nog", "cni", "teo", "tdx", "sxn", "rki", "nr", "frp", "alz", "taj", "lrc", "cce", "rn", "jvn", "hvn", "nij", "dwr", "izz", "msm", "bus", "ktu", "chr", "maz", "tzj", "suz", "knj", "bim", "gvl", "bqc", "tca", "pis", "prk", "laj", "mel", "qxr", "niq", "ahk", "shp", "hne", "spp", "koi", "krj", "quf", "luz", "agr", "tsc", "mqy", "gof", "gbm", "miq", "dje", "awa", "bjj", "qvz", "sjp", "tll", "raj", "kjg", "bgz", "quy", "cbk", "akb", "oj", "ify", "mey", "ks", "cac", "brx", "qup", "syl", "jax", "ff", "ber", "tks", "trp", "mrw", "adh", "smt", "srr", "ffm", "qvc", "mtr", "ann", "aa", "noe", "nut", "gyn", "kwi", "xmm", "msb", "dataset:allenai/MADLAD-400", "arxiv:2204.02311", "arxiv:2309.04662", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "DecoderOnlyT5Model" ]
null
3
null
null
null
null
null
t5
null
null
null
null
null
null
true
null
4.23.1
true
256,512
null
null
null
null
null
null
ceadar-ie/FinanceConnect-13B
null
null
"2023-11-28T14:26:49Z"
null
null
751
null
null
null
null
14
transformers
[ "transformers", "safetensors", "llama", "text-generation", "summarization", "classification", "translation", "NLP", "finance", "domain specific llm", "conversational", "en", "dataset:FinTalk-19k", "doi:10.57967/hf/1405", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
1
2
silu
5,120
0.02
13,824
4,096
llama
40
40
40
0.00001
10,000
null
false
float16
4.36.0.dev0
true
32,000
false
0
null
null
1
null
NousResearch/Nous-Hermes-2-Vision-Alpha
null
null
"2023-11-28T23:18:55Z"
null
null
105
null
null
null
null
304
transformers
[ "transformers", "pytorch", "llava_mistral", "text-generation", "mistral", "instruct", "finetune", "chatml", "gpt4", "synthetic data", "distillation", "multimodal", "llava", "conversational", "en", "base_model:mistralai/Mistral-7B-v0.1", "base_model:finetune:mistralai/Mistral-7B-v0.1", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "LlavaMistralForCausalLM" ]
1
32,000
silu
4,096
0.02
14,336
32,768
llava_mistral
32
32
8
0.00001
10,000
4,096
false
bfloat16
4.34.1
true
32,002
null
null
null
null
null
null
deepseek-ai/deepseek-llm-7b-chat
null
null
"2023-11-29T03:19:32Z"
null
null
11,712
null
null
null
null
76
transformers
[ "transformers", "pytorch", "llama", "text-generation", "conversational", "license:other", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
100,000
100,001
silu
4,096
0.02
11,008
4,096
llama
32
30
32
0.000001
10,000
null
false
bfloat16
4.33.1
true
102,400
null
null
null
null
1
null
deepseek-ai/deepseek-llm-67b-base
null
null
"2023-11-29T03:23:47Z"
null
null
1,280
null
null
null
null
111
transformers
[ "transformers", "pytorch", "llama", "text-generation", "license:other", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
1
2
silu
8,192
0.02
22,016
4,096
llama
64
95
8
0.000001
10,000
null
false
bfloat16
4.33.1
true
102,400
null
null
null
null
1
null
deepseek-ai/deepseek-llm-67b-chat
null
null
"2023-11-29T03:30:00Z"
null
null
1,518
null
null
null
null
177
transformers
[ "transformers", "pytorch", "llama", "text-generation", "conversational", "license:other", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
100,000
100,001
silu
8,192
0.02
22,016
4,096
llama
64
95
8
0.000001
10,000
null
false
bfloat16
4.33.1
true
102,400
null
null
null
null
1
null
athirdpath/Iambe-20b-DARE
null
null
"2023-11-29T06:31:49Z"
null
null
19
null
null
null
null
11
transformers
[ "transformers", "safetensors", "llama", "text-generation", "license:cc-by-nc-4.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
1
2
silu
5,120
0.02
13,824
4,096
llama
40
62
40
0.00001
10,000
null
false
bfloat16
4.35.2
false
32,000
false
null
null
null
1
null
SUSTech/SUS-Chat-34B
null
null
"2023-11-29T09:07:53Z"
null
null
1,029
null
null
null
null
122
transformers
[ "transformers", "pytorch", "llama", "text-generation", "doi:10.57967/hf/1718", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
1
2
silu
7,168
0.02
20,480
8,192
llama
56
60
8
0.00001
5,000,000
null
false
bfloat16
4.35.0
true
64,000
false
null
null
null
1
null
Qwen/Qwen-72B-Chat
null
null
"2023-11-29T09:37:07Z"
null
null
2,157
null
null
null
null
148
transformers
[ "transformers", "safetensors", "qwen", "text-generation", "custom_code", "zh", "en", "arxiv:2309.16609", "arxiv:2305.08322", "arxiv:2009.03300", "arxiv:2307.11088", "base_model:Qwen/Qwen-72B", "base_model:finetune:Qwen/Qwen-72B", "license:other", "autotrain_compatible", "region:us" ]
text-generation
null
null
1
[ "QWenLMHeadModel" ]
null
null
null
8,192
0.02
49,152
32,768
qwen
64
80
null
null
1,000,000
null
false
null
4.32.0
true
152,064
null
null
null
null
null
null
mlabonne/NeuralHermes-2.5-Mistral-7B
null
null
"2023-11-29T12:23:31Z"
null
null
318
null
null
null
null
151
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "instruct", "finetune", "chatml", "gpt4", "synthetic data", "distillation", "dpo", "rlhf", "conversational", "en", "dataset:mlabonne/chatml_dpo_pairs", "base_model:teknium/OpenHermes-2-coder-Mistral-7B-v2-e4", "base_model:finetune:teknium/OpenHermes-2-coder-Mistral-7B-v2-e4", "license:apache-2.0", "model-index", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "MistralForCausalLM" ]
1
32,000
silu
4,096
0.02
14,336
32,768
mistral
32
32
8
0.00001
10,000
4,096
false
float16
4.35.2
false
32,002
null
null
null
null
null
null
LLM360/AmberChat
null
null
"2023-11-30T00:12:23Z"
null
null
351
null
null
null
null
23
transformers
[ "transformers", "safetensors", "llama", "text-generation", "nlp", "llm", "en", "dataset:WizardLM/WizardLM_evol_instruct_V2_196k", "dataset:icybee/share_gpt_90k_v1", "arxiv:2312.06550", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
1
2
silu
4,096
0.02
11,008
2,048
llama
32
32
32
0.000001
10,000
null
false
bfloat16
4.35.2
true
32,000
false
null
null
null
1
null
Qwen/Qwen-1_8B-Chat
null
null
"2023-11-30T02:56:11Z"
null
null
57,677
null
null
null
null
108
transformers
[ "transformers", "safetensors", "qwen", "text-generation", "custom_code", "zh", "en", "arxiv:2309.16609", "arxiv:2305.08322", "arxiv:2009.03300", "autotrain_compatible", "region:us" ]
text-generation
null
null
1
[ "QWenLMHeadModel" ]
null
null
null
2,048
0.02
11,008
8,192
qwen
16
24
null
null
null
null
false
null
4.32.0
true
151,936
null
null
null
null
null
null
m-a-p/ChatMusician
null
null
"2023-11-30T03:48:16Z"
null
null
283
null
null
null
null
116
transformers
[ "transformers", "pytorch", "safetensors", "llama", "text-generation", "en", "arxiv:2402.16153", "license:mit", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
1
2
silu
4,096
0.02
11,008
2,048
llama
32
32
32
0.00001
null
null
false
float16
4.32.0
true
32,000
null
null
null
null
1
null
maywell/Synatra-42dot-1.3B
null
null
"2023-11-30T05:11:41Z"
null
null
3,834
null
null
null
null
6
transformers
[ "transformers", "safetensors", "llama", "text-generation", "license:cc-by-nc-4.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
50,257
50,256
silu
2,048
0.01
5,632
8,192
llama
32
24
32
0.000001
10,000
null
false
bfloat16
4.35.2
false
50,304
false
null
null
null
1
null
Qwen/Qwen-Audio-Chat
null
null
"2023-11-30T09:38:13Z"
null
null
116,602
null
null
null
null
74
transformers
[ "transformers", "safetensors", "qwen", "text-generation", "custom_code", "zh", "en", "arxiv:2311.07919", "autotrain_compatible", "region:us" ]
text-generation
null
null
1
[ "QWenLMHeadModel" ]
null
null
null
4,096
0.02
22,016
2,048
qwen
32
32
null
null
null
null
false
bfloat16
4.32.0
true
155,947
null
null
null
null
null
null
mzyil/llama-2-13b-mzyil-3-merged
null
null
"2023-12-03T01:13:56Z"
null
null
13
null
null
null
null
1
transformers
[ "transformers", "pytorch", "safetensors", "llama", "text-generation", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
1
2
silu
5,120
0.02
13,824
4,096
llama
40
40
40
0.00001
null
null
false
float16
4.31.0
true
32,000
null
null
null
null
1
null
mzyil/llama-2-7b-mzyil-3-merged
null
null
"2023-12-03T01:48:10Z"
null
null
8
null
null
null
null
1
transformers
[ "transformers", "pytorch", "safetensors", "llama", "text-generation", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
1
2
silu
4,096
0.02
11,008
4,096
llama
32
32
32
0.00001
null
null
false
float16
4.31.0
true
32,000
null
null
null
null
1
null
GAIR/autoj-bilingual-6b
null
null
"2023-12-03T11:31:52Z"
null
null
16
null
null
null
null
5
transformers
[ "transformers", "pytorch", "llama", "text-generation", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
1
2
silu
4,096
0.02
11,008
4,096
llama
32
32
4
0.00001
5,000,000
null
false
float16
4.35.1
true
64,000
false
null
null
null
1
null
cmarkea/bloomz-3b-dpo-chat
null
null
"2023-12-04T08:30:58Z"
null
null
65
null
null
null
null
2
transformers
[ "transformers", "pytorch", "safetensors", "bloom", "text-generation", "en", "fr", "dataset:Anthropic/hh-rlhf", "license:bigscience-bloom-rail-1.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "BloomForCausalLM" ]
1
2
null
2,560
0.02
null
null
bloom
null
null
null
null
null
null
null
bfloat16
4.34.0
true
250,880
null
0
null
null
4
null
lamm-mit/SilkomeGPT
null
null
"2023-12-04T11:28:26Z"
null
null
394
null
null
null
null
2
transformers
[ "transformers", "pytorch", "gpt_neox", "text-generation", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "GPTNeoXForCausalLM" ]
0
2
gelu
1,024
0.02
4,096
2,048
gpt_neox
8
12
null
null
null
null
false
float32
4.31.0.dev0
false
50,000
null
null
null
null
null
null
Trelis/Llama-2-7b-chat-hf-function-calling-v3
null
null
"2023-12-04T14:42:08Z"
null
null
267
null
null
null
null
40
transformers
[ "transformers", "safetensors", "llama", "text-generation", "facebook", "meta", "pytorch", "llama-2", "gguf", "function-calling", "function calling", "conversational", "en", "arxiv:2307.09288", "autotrain_compatible", "text-generation-inference", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
1
2
silu
4,096
0.02
11,008
4,096
llama
32
32
32
0.00001
10,000
null
false
bfloat16
4.38.1
true
32,000
false
0
null
null
1
null
togethercomputer/StripedHyena-Nous-7B
null
null
"2023-12-04T19:56:49Z"
null
null
17,584
null
null
null
null
140
transformers
[ "transformers", "pytorch", "safetensors", "stripedhyena", "text-generation", "custom_code", "en", "arxiv:2302.10866", "arxiv:2310.18780", "arxiv:2311.05908", "doi:10.57967/hf/1595", "license:apache-2.0", "autotrain_compatible", "region:us" ]
text-generation
null
null
1
[ "StripedHyenaModelForCausalLM" ]
null
null
null
4,096
null
null
null
stripedhyena
32
null
null
null
null
null
null
bfloat16
null
true
32,000
null
null
null
null
null
null
Nexusflow/NexusRaven-V2-13B
null
null
"2023-12-04T22:06:57Z"
null
null
7,876
null
null
null
null
462
transformers
[ "transformers", "pytorch", "llama", "text-generation", "function calling", "arxiv:2308.12950", "base_model:codellama/CodeLlama-13b-Instruct-hf", "base_model:finetune:codellama/CodeLlama-13b-Instruct-hf", "license:other", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
1
2
silu
5,120
0.02
13,824
16,384
llama
40
40
40
0.00001
1,000,000
null
false
bfloat16
4.33.0
true
32,024
null
null
null
null
1
null
OrionZheng/openmoe-base
null
null
"2023-12-08T03:29:33Z"
null
null
593
null
null
null
null
4
transformers
[ "transformers", "pytorch", "llama", "text-generation", "custom_code", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "OpenMoeForCausalLM" ]
0
1
swiglu
768
0.02
2,048
2,048
llama
12
12
12
0.000001
10,000
null
false
float32
4.34.0
true
256,384
false
null
64
null
1
null
LoneStriker/dolphin-2.2-yi-34b-200k-4.0bpw-h6-exl2
null
null
"2023-12-08T09:05:45Z"
null
null
19
null
null
null
null
1
transformers
[ "transformers", "pytorch", "llama", "text-generation", "en", "dataset:ehartford/dolphin", "dataset:jondurbin/airoboros-2.2.1", "dataset:ehartford/samantha-data", "dataset:ehartford/WizardLM_evol_instruct_V2_196k_unfiltered_merged_split", "license:other", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
1
7
silu
7,168
0.02
20,480
200,000
llama
56
60
8
0.00001
5,000,000
null
false
bfloat16
4.35.2
true
64,000
false
null
null
null
1
null
Weyaxi/OpenHermes-2.5-neural-chat-v3-2-Slerp
null
null
"2023-12-08T13:19:27Z"
null
null
93
null
null
null
null
29
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "merge", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "MistralForCausalLM" ]
1
2
silu
4,096
0.02
14,336
32,768
mistral
32
32
8
0.00001
10,000
4,096
false
bfloat16
4.35.2
true
32,000
null
null
null
null
null
null
AdaptLLM/medicine-chat
null
null
"2023-12-09T01:49:56Z"
null
null
3,740
null
null
null
null
44
transformers
[ "transformers", "pytorch", "safetensors", "llama", "text-generation", "biology", "medical", "en", "dataset:EleutherAI/pile", "dataset:Open-Orca/OpenOrca", "dataset:GAIR/lima", "dataset:WizardLM/WizardLM_evol_instruct_V2_196k", "arxiv:2309.09530", "arxiv:2406.14491", "license:llama2", "model-index", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
1
2
silu
4,096
0.02
11,008
4,096
llama
32
32
32
0.00001
null
null
false
float16
4.31.0.dev0
true
32,001
null
null
null
null
1
null
AdaptLLM/law-chat
null
null
"2023-12-09T02:42:38Z"
null
null
1,030
null
null
null
null
33
transformers
[ "transformers", "pytorch", "llama", "text-generation", "legal", "en", "dataset:EleutherAI/pile", "dataset:Open-Orca/OpenOrca", "dataset:GAIR/lima", "dataset:WizardLM/WizardLM_evol_instruct_V2_196k", "arxiv:2309.09530", "arxiv:2406.14491", "license:llama2", "model-index", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
1
2
silu
4,096
0.02
11,008
4,096
llama
32
32
32
0.00001
null
null
false
float16
4.31.0.dev0
true
32,001
null
null
null
null
1
null
TheBloke/OrcaMaid-13B-GGUF
null
null
"2023-12-09T10:37:08Z"
null
null
522
null
null
null
null
8
transformers
[ "transformers", "gguf", "llama", "text-generation", "base_model:ddh0/OrcaMaid-13b", "base_model:quantized:ddh0/OrcaMaid-13b", "license:other", "region:us" ]
text-generation
null
null
1
null
null
null
null
null
null
null
null
llama
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
migtissera/Synthia-MoE-v3-CE
null
null
"2023-12-11T01:28:47Z"
null
null
35
null
null
null
null
16
transformers
[ "transformers", "pytorch", "mistral", "text-generation", "custom_code", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "MixtralForCausalLM" ]
1
2
silu
4,096
0.02
14,336
32,768
mistral
32
32
8
0.00001
1,000,000
null
false
float16
4.36.0.dev0
false
32,000
null
0
null
null
null
null
TinyLlama/TinyLlama-1.1B-intermediate-step-1195k-token-2.5T
null
null
"2023-12-11T06:13:09Z"
null
null
3,814
null
null
null
null
50
transformers
[ "transformers", "pytorch", "safetensors", "llama", "text-generation", "en", "dataset:cerebras/SlimPajama-627B", "dataset:bigcode/starcoderdata", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
1
2
silu
2,048
0.02
5,632
2,048
llama
32
22
4
0.00001
null
null
false
float32
4.31.0.dev0
true
32,000
null
null
null
null
1
null
jamesdborin/llama2-13b-chat-4bit-AWQ
null
null
"2023-12-11T09:12:09Z"
null
null
20
null
null
null
null
1
transformers
[ "transformers", "safetensors", "llama", "text-generation", "facebook", "meta", "pytorch", "llama-2", "en", "arxiv:2307.09288", "base_model:meta-llama/Llama-2-13b-chat-hf", "base_model:quantized:meta-llama/Llama-2-13b-chat-hf", "license:llama2", "autotrain_compatible", "text-generation-inference", "4-bit", "awq", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
1
2
silu
5,120
0.02
13,824
4,096
llama
40
40
40
0.00001
null
null
false
float16
4.31.0.dev0
true
32,000
null
null
null
null
1
null
TheBloke/Mixtral-8x7B-Instruct-v0.1-GPTQ
null
null
"2023-12-11T18:49:53Z"
null
null
57,771
null
null
null
null
134
transformers
[ "transformers", "safetensors", "mixtral", "text-generation", "conversational", "fr", "it", "de", "es", "en", "base_model:mistralai/Mixtral-8x7B-Instruct-v0.1", "base_model:quantized:mistralai/Mixtral-8x7B-Instruct-v0.1", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "4-bit", "gptq", "region:us" ]
text-generation
null
null
1
[ "MixtralForCausalLM" ]
1
2
silu
4,096
0.02
14,336
32,768
mixtral
32
32
8
0.00001
1,000,000
4,096
false
bfloat16
4.36.0
true
32,000
null
0
null
null
1
null
TheBloke/Mixtral-8x7B-Instruct-v0.1-AWQ
null
null
"2023-12-11T21:01:33Z"
null
null
383
null
null
null
null
58
transformers
[ "transformers", "safetensors", "mixtral", "text-generation", "conversational", "fr", "it", "de", "es", "en", "base_model:mistralai/Mixtral-8x7B-Instruct-v0.1", "base_model:quantized:mistralai/Mixtral-8x7B-Instruct-v0.1", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "4-bit", "awq", "region:us" ]
text-generation
null
null
1
[ "MixtralForCausalLM" ]
1
2
silu
4,096
0.02
14,336
32,768
mixtral
32
32
8
0.00001
1,000,000
4,096
false
float16
4.36.0.dev0
true
32,000
null
0
null
null
1
null
TheBloke/Mistral-7B-Instruct-v0.2-GPTQ
null
null
"2023-12-11T22:18:46Z"
null
null
456,687
null
null
null
null
50
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "finetuned", "conversational", "arxiv:2310.06825", "base_model:mistralai/Mistral-7B-Instruct-v0.2", "base_model:quantized:mistralai/Mistral-7B-Instruct-v0.2", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "4-bit", "gptq", "region:us" ]
text-generation
null
null
1
[ "MistralForCausalLM" ]
1
2
silu
4,096
0.02
14,336
32,768
mistral
32
32
8
0.00001
1,000,000
null
false
bfloat16
4.36.0.dev0
true
32,000
null
0
null
null
1
null
Open-Orca/Mixtral-SlimOrca-8x7B
null
null
"2023-12-12T02:30:31Z"
null
null
36
null
null
null
null
52
transformers
[ "transformers", "safetensors", "mixtral", "text-generation", "dataset:Open-Orca/SlimOrca", "base_model:mistralai/Mixtral-8x7B-v0.1", "base_model:finetune:mistralai/Mixtral-8x7B-v0.1", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "MixtralForCausalLM" ]
1
32,000
silu
4,096
0.02
14,336
32,768
mixtral
32
32
8
0.00001
1,000,000
4,096
false
bfloat16
4.36.0.dev0
true
32,002
null
0
null
null
null
null
wenge-research/yayi2-30b
null
null
"2023-12-12T08:40:29Z"
null
null
80
null
null
null
null
75
transformers
[ "transformers", "pytorch", "yayi", "text-generation", "custom_code", "arxiv:2312.14862", "arxiv:2307.09288", "license:other", "autotrain_compatible", "region:us" ]
text-generation
null
null
1
[ "YayiForCausalLM" ]
1
2
silu
7,168
0.02
16,384
4,096
yayi
64
64
null
0.000001
null
null
false
bfloat16
4.29.1
true
81,920
null
null
null
null
null
null
vilm/vinallama-7b-chat
null
null
"2023-12-12T16:58:23Z"
null
null
1,011
null
null
null
null
21
transformers
[ "transformers", "pytorch", "safetensors", "llama", "text-generation", "conversational", "vi", "arxiv:2312.11011", "license:llama2", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
1
46,303
silu
4,096
0.02
11,008
4,096
llama
32
32
32
0.00001
10,000
null
false
bfloat16
4.38.1
false
46,305
false
0
null
null
1
null
ybelkada/Mixtral-8x7B-Instruct-v0.1-AWQ
null
null
"2023-12-12T22:22:51Z"
null
null
389
null
null
null
null
15
transformers
[ "transformers", "safetensors", "mixtral", "text-generation", "conversational", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "4-bit", "awq", "region:us" ]
text-generation
null
null
1
[ "MixtralForCausalLM" ]
1
2
silu
4,096
0.02
14,336
32,768
mixtral
32
32
8
0.00001
1,000,000
4,096
false
float16
4.36.0.dev0
true
32,000
null
0
null
null
null
null
sarvamai/OpenHathi-7B-Hi-v0.1-Base
null
null
"2023-12-13T13:41:11Z"
null
null
1,692
null
null
null
null
105
transformers
[ "transformers", "safetensors", "gguf", "llama", "text-generation", "hi", "license:llama2", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
1
2
silu
4,096
0.02
11,008
4,096
llama
32
32
32
0.00001
10,000
null
false
bfloat16
4.35.0
true
48,064
false
null
null
null
1
null
w4r10ck/SOLAR-10.7B-Instruct-v1.0-uncensored
null
null
"2023-12-14T01:33:36Z"
null
null
167
null
null
null
null
30
transformers
[ "transformers", "safetensors", "llama", "text-generation", "license:apache-2.0", "model-index", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
1
2
silu
4,096
0.02
14,336
4,096
llama
32
48
8
0.00001
10,000
null
false
float16
4.35.2
false
32,000
false
null
null
null
1
null
Lin-Chen/ShareGPT4V-13B
null
null
"2023-12-14T05:55:54Z"
null
null
425
null
null
null
null
33
transformers
[ "transformers", "pytorch", "share4v", "text-generation", "arxiv:2311.12793", "autotrain_compatible", "region:us" ]
text-generation
null
null
1
[ "Share4VLlamaForCausalLM" ]
1
2
silu
5,120
0.02
13,824
4,096
share4v
40
40
40
0.00001
null
null
false
bfloat16
4.31.0
true
32,000
null
null
null
null
1
null
joey00072/ToxicHermes-2.5-Mistral-7B
null
null
"2023-12-14T11:19:11Z"
null
null
19
null
null
null
null
17
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "instruct", "finetune", "chatml", "gpt4", "synthetic data", "distillation", "dpo", "rlhf", "conversational", "en", "dataset:unalignment/toxic-dpo-v0.1", "base_model:teknium/OpenHermes-2.5-Mistral-7B", "base_model:finetune:teknium/OpenHermes-2.5-Mistral-7B", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "MistralForCausalLM" ]
1
32,000
silu
4,096
0.02
14,336
32,768
mistral
32
32
8
0.00001
10,000
4,096
false
float16
4.36.1
false
32,002
null
0
null
null
null
null
THUDM/cogagent-chat-hf
null
null
"2023-12-15T06:12:34Z"
null
null
10,428
null
null
null
null
64
transformers
[ "transformers", "safetensors", "text-generation", "custom_code", "en", "arxiv:2312.08914", "arxiv:2311.03079", "license:apache-2.0", "autotrain_compatible", "region:us" ]
text-generation
null
null
1
[ "CogAgentForCausalLM" ]
1
2
silu
4,096
0.02
11,008
2,048
null
32
32
null
0.00001
null
null
false
bfloat16
4.36.0.dev0
true
32,000
null
null
null
null
null
null
chargoddard/SmolLlamix-8x101M
null
null
"2023-12-15T06:25:58Z"
null
null
970
null
null
null
null
12
transformers
[ "transformers", "safetensors", "mixtral", "text-generation", "llama", "en", "dataset:togethercomputer/RedPajama-Data-1T-Sample", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "MixtralForCausalLM" ]
1
2
silu
768
0.02
3,072
32,768
mixtral
24
6
8
0.00001
10,000
1,024
false
bfloat16
4.37.0.dev0
false
32,128
false
0
null
null
1
null
VAGOsolutions/SauerkrautLM-Mixtral-8x7B-Instruct
null
null
"2023-12-15T16:01:09Z"
null
null
77,291
null
null
null
null
22
transformers
[ "transformers", "safetensors", "mixtral", "text-generation", "mistral", "finetune", "dpo", "Instruct", "augmentation", "german", "moe", "conversational", "en", "de", "fr", "it", "es", "dataset:argilla/distilabel-math-preference-dpo", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "MixtralForCausalLM" ]
1
2
silu
4,096
0.02
14,336
32,768
mixtral
32
32
8
0.00001
1,000,000
null
false
bfloat16
4.36.0.dev0
true
32,000
null
0
null
null
null
null
WhiteRabbitNeo/WhiteRabbitNeo-13B-v1
null
null
"2023-12-17T16:26:55Z"
null
null
1,869
null
null
null
null
392
transformers
[ "transformers", "pytorch", "llama", "text-generation", "custom_code", "license:llama2", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
1
2
silu
5,120
0.02
13,824
16,384
llama
40
40
40
0.00001
1,000,000
null
false
bfloat16
4.36.0.dev0
false
32,016
false
0
null
null
1
null
SciPhi/Sensei-7B-V1
null
null
"2023-12-18T17:43:53Z"
null
null
71
null
null
null
null
90
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "MistralForCausalLM" ]
1
2
silu
4,096
0.02
14,336
32,768
mistral
32
32
8
0.00001
10,000
4,096
false
float32
4.37.0.dev0
false
32,000
null
0
null
null
null
null
ignos/Mistral-T5-7B-v1
null
null
"2023-12-18T18:09:44Z"
null
null
784
null
null
null
null
8
transformers
[ "transformers", "pytorch", "mistral", "text-generation", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "MistralForCausalLM" ]
1
2
silu
4,096
0.02
14,336
32,768
mistral
32
32
8
0.00001
10,000
4,096
false
bfloat16
4.37.0.dev0
false
32,000
null
0
null
null
null
null
danish-foundation-models/munin-7b-alpha
null
null
"2023-12-18T18:42:32Z"
null
null
93
null
null
null
null
29
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "pretrained", "da", "dataset:DDSC/partial-danish-gigaword-no-twitter", "base_model:mistralai/Mistral-7B-v0.1", "base_model:finetune:mistralai/Mistral-7B-v0.1", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "MistralForCausalLM" ]
1
2
silu
4,096
0.02
14,336
32,768
mistral
32
32
8
0.00001
10,000
4,096
false
bfloat16
4.36.2
false
32,000
null
0
null
null
null
null
TheBloke/phi-2-GGUF
null
null
"2023-12-18T20:22:56Z"
null
null
5,205
null
null
null
null
192
transformers
[ "transformers", "gguf", "phi-msft", "nlp", "code", "text-generation", "en", "base_model:microsoft/phi-2", "base_model:quantized:microsoft/phi-2", "license:other", "region:us" ]
text-generation
null
null
1
null
null
null
null
null
null
null
null
phi-msft
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
TKDKid1000/phi-1_5-GGUF
null
null
"2023-12-19T02:28:32Z"
null
null
224
null
null
null
null
6
transformers
[ "transformers", "pytorch", "gguf", "phi-msft", "text-generation", "nlp", "code", "custom_code", "en", "arxiv:2309.05463", "license:other", "autotrain_compatible", "region:us" ]
text-generation
null
null
1
[ "PhiForCausalLM" ]
null
null
null
null
0.02
null
null
phi-msft
null
null
null
null
null
null
false
float16
4.34.1
null
51,200
null
null
null
null
null
null
rinna/nekomata-14b-instruction
null
null
"2023-12-19T08:10:24Z"
null
null
1,268
null
null
null
null
22
transformers
[ "transformers", "pytorch", "safetensors", "qwen", "text-generation", "custom_code", "ja", "en", "dataset:databricks/databricks-dolly-15k", "dataset:kunishou/databricks-dolly-15k-ja", "dataset:izumi-lab/llm-japanese-dataset", "arxiv:2309.16609", "arxiv:2404.01657", "base_model:rinna/nekomata-14b", "base_model:finetune:rinna/nekomata-14b", "license:other", "autotrain_compatible", "region:us" ]
text-generation
null
null
1
[ "QWenLMHeadModel" ]
null
null
null
5,120
0.02
27,392
8,192
qwen
40
40
null
null
null
null
false
null
4.32.0
true
152,064
null
null
null
null
null
null
ThisIs-Developer/Llama-2-GGML-Medical-Chatbot
null
null
"2023-12-19T14:51:37Z"
null
null
241
null
null
null
null
19
transformers
[ "transformers", "llama", "medical", "conversational", "text-generation", "question-answering", "en", "base_model:TheBloke/Llama-2-7B-Chat-GGML", "base_model:finetune:TheBloke/Llama-2-7B-Chat-GGML", "license:mit", "endpoints_compatible", "region:us" ]
question-answering
null
null
1
null
null
null
null
null
null
null
null
llama
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
mlx-community/phi-2
null
null
"2023-12-19T16:45:41Z"
null
null
23
null
null
null
null
52
mlx
[ "mlx", "phi-msft", "nlp", "code", "text-generation", "en", "base_model:microsoft/phi-2", "base_model:finetune:microsoft/phi-2", "license:other", "region:us" ]
text-generation
null
null
1
null
null
null
null
null
null
null
null
phi-msft
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
AdaptLLM/law-LLM-13B
null
null
"2023-12-20T01:54:43Z"
null
null
417
null
null
null
null
31
transformers
[ "transformers", "pytorch", "safetensors", "llama", "text-generation", "legal", "en", "dataset:Open-Orca/OpenOrca", "dataset:GAIR/lima", "dataset:WizardLM/WizardLM_evol_instruct_V2_196k", "dataset:EleutherAI/pile", "arxiv:2309.09530", "arxiv:2406.14491", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
1
2
silu
5,120
0.02
13,824
2,048
llama
40
40
null
0.000001
null
null
false
float16
4.28.0.dev0
true
32,001
null
null
null
null
null
null
scb10x/typhoon-7b
null
null
"2023-12-20T14:07:35Z"
null
null
1,462
null
null
null
null
100
transformers
[ "transformers", "pytorch", "safetensors", "mistral", "text-generation", "pretrained", "th", "arxiv:2312.13951", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "MistralForCausalLM" ]
1
2
silu
4,096
0.02
14,336
32,768
mistral
32
32
8
0.00001
10,000
4,096
false
bfloat16
4.34.0.dev0
true
35,219
null
null
null
null
null
null
beowolx/CodeNinja-1.0-OpenChat-7B
null
null
"2023-12-20T20:28:01Z"
null
null
4,502
null
null
null
null
105
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "code", "text-generation-inference", "conversational", "en", "dataset:glaiveai/glaive-code-assistant-v2", "dataset:TokenBender/code_instructions_122k_alpaca_style", "doi:10.57967/hf/1535", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "MistralForCausalLM" ]
1
2
silu
4,096
0.02
14,336
8,192
mistral
32
32
8
0.00001
10,000
4,096
false
bfloat16
4.37.0.dev0
false
32,002
null
0
null
null
null
null
TheBloke/DaringMaid-20B-GGUF
null
null
"2023-12-20T21:04:04Z"
null
null
1,040
null
null
null
null
18
transformers
[ "transformers", "gguf", "llama", "text-generation", "en", "base_model:Kooten/DaringMaid-20B", "base_model:quantized:Kooten/DaringMaid-20B", "license:cc-by-nc-4.0", "region:us" ]
text-generation
null
null
1
null
null
null
null
null
null
null
null
llama
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
SanjiWatsuki/Loyal-Toppy-Bruins-Maid-7B-DARE
null
null
"2023-12-21T01:50:18Z"
null
null
825
null
null
null
null
11
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "merge", "license:cc-by-nc-4.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "MistralForCausalLM" ]
1
2
silu
4,096
0.02
14,336
32,768
mistral
32
32
8
0.00001
10,000
4,096
false
bfloat16
4.36.2
true
32,000
null
0
null
null
null
null
HillZhang/untruthful_llama2_7b
null
null
"2023-12-21T07:40:22Z"
null
null
27
null
null
null
null
2
transformers
[ "transformers", "pytorch", "llama", "text-generation", "text-generation-inference", "English", "en", "dataset:pminervini/HaluEval", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
1
2
silu
4,096
0.02
11,008
2,048
llama
32
32
32
0.00001
10,000
null
false
float16
4.34.1
true
32,000
false
null
null
null
1
null
TheBloke/OrcaMaid-v2-FIX-13B-32k-GGUF
null
null
"2023-12-21T14:10:10Z"
null
null
525
null
null
null
null
5
transformers
[ "transformers", "gguf", "llama", "text-generation", "base_model:ddh0/OrcaMaid-v2-FIX-13b-32k", "base_model:quantized:ddh0/OrcaMaid-v2-FIX-13b-32k", "license:other", "region:us" ]
text-generation
null
null
1
null
null
null
null
null
null
null
null
llama
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
TheBloke/Llama-2-7B-ft-instruct-es-GPTQ
null
null
"2023-12-21T15:23:09Z"
null
null
112
null
null
null
null
2
transformers
[ "transformers", "safetensors", "llama", "text-generation", "es", "base_model:clibrain/Llama-2-7b-ft-instruct-es", "base_model:quantized:clibrain/Llama-2-7b-ft-instruct-es", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "4-bit", "gptq", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
1
2
silu
4,096
0.02
11,008
4,096
llama
32
32
32
0.00001
10,000
null
false
float16
4.35.2
true
32,000
false
null
null
null
1
null
BEE-spoke-data/smol_llama-220M-GQA
null
null
"2023-12-22T10:48:33Z"
null
null
2,810
null
null
null
null
12
transformers
[ "transformers", "safetensors", "llama", "text-generation", "smol_llama", "llama2", "en", "dataset:JeanKaddour/minipile", "dataset:pszemraj/simple_wikipedia_LM", "dataset:mattymchen/refinedweb-3m", "dataset:BEE-spoke-data/knowledge-inoc-concat-v1", "license:apache-2.0", "model-index", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
1
2
silu
1,024
0.02
4,096
2,048
llama
32
10
8
0.000001
10,000
null
false
bfloat16
4.37.0.dev0
true
32,128
false
0
null
null
1
null
casperhansen/mixtral-instruct-awq
null
null
"2023-12-22T12:15:06Z"
null
null
30,699
null
null
null
null
43
transformers
[ "transformers", "safetensors", "mixtral", "text-generation", "conversational", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "4-bit", "awq", "region:us" ]
text-generation
null
null
1
[ "MixtralForCausalLM" ]
1
2
silu
4,096
0.02
14,336
32,768
mixtral
32
32
8
0.00001
1,000,000
null
false
float16
4.36.2
true
32,000
null
0
null
null
null
null
cognitivecomputations/dolphin-2_6-phi-2
null
null
"2023-12-23T07:51:30Z"
null
null
264
null
null
null
null
192
transformers
[ "transformers", "pytorch", "safetensors", "phi-msft", "text-generation", "conversational", "custom_code", "en", "dataset:ehartford/dolphin", "dataset:jondurbin/airoboros-2.2.1", "dataset:ehartford/dolphin-coder", "dataset:teknium/openhermes", "dataset:ise-uiuc/Magicoder-OSS-Instruct-75K", "dataset:ise-uiuc/Magicoder-Evol-Instruct-110K", "dataset:LDJnr/Capybara", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "PhiForCausalLM" ]
null
null
null
null
0.02
null
null
phi-msft
null
null
null
null
null
null
false
float16
4.37.0.dev0
false
51,200
null
null
null
null
null
null
NousResearch/Nous-Hermes-2-Yi-34B
null
null
"2023-12-23T19:47:48Z"
null
null
17,622
null
null
null
null
247
transformers
[ "transformers", "safetensors", "llama", "text-generation", "yi", "instruct", "finetune", "chatml", "gpt4", "synthetic data", "distillation", "conversational", "en", "dataset:teknium/OpenHermes-2.5", "base_model:01-ai/Yi-34B", "base_model:finetune:01-ai/Yi-34B", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
1
7
silu
7,168
0.02
20,480
4,096
llama
56
60
8
0.00001
5,000,000
null
false
bfloat16
4.37.0.dev0
false
64,000
false
0
null
null
1
null
cloudyu/Mixtral_13Bx2_MOE_22B
null
null
"2023-12-24T03:05:32Z"
null
null
41
null
null
null
null
12
transformers
[ "transformers", "safetensors", "mixtral", "text-generation", "license:cc-by-nc-4.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "MixtralForCausalLM" ]
1
2
silu
5,120
0.02
13,824
4,096
mixtral
40
40
40
0.00001
10,000
null
false
float16
4.36.2
false
32,000
false
0
null
null
1
null
cygu/llama-2-7b-logit-watermark-distill-kgw-k1-gamma0.25-delta2
null
null
"2023-12-24T07:21:40Z"
null
null
56
null
null
null
null
1
transformers
[ "transformers", "pytorch", "llama", "text-generation", "generated_from_trainer", "dataset:openwebtext", "arxiv:2312.04469", "license:llama2", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
1
2
silu
4,096
0.02
11,008
4,096
llama
32
32
32
0.00001
null
null
false
float32
4.29.2
true
32,000
null
null
null
null
1
null
NeverSleep/Noromaid-v0.1-mixtral-8x7b-v3
null
null
"2023-12-24T10:43:51Z"
null
null
50
null
null
null
null
15
transformers
[ "transformers", "pytorch", "mixtral", "text-generation", "license:cc-by-nc-4.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "MixtralForCausalLM" ]
1
2
silu
4,096
0.02
14,336
32,768
mixtral
32
32
8
0.00001
1,000,000
null
false
bfloat16
4.36.2
false
32,000
null
0
null
null
null
null
TheBloke/typhoon-7B-GGUF
null
null
"2023-12-24T16:25:16Z"
null
null
480
null
null
null
null
8
transformers
[ "transformers", "gguf", "mistral", "pretrained", "text-generation", "th", "arxiv:2312.13951", "base_model:scb10x/typhoon-7b", "base_model:quantized:scb10x/typhoon-7b", "license:apache-2.0", "region:us" ]
text-generation
null
null
1
null
null
null
null
null
null
null
null
mistral
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
ybelkada/Mixtral-8x7B-Instruct-v0.1-bnb-4bit
null
null
"2023-12-25T11:09:56Z"
null
null
125
null
null
null
null
59
transformers
[ "transformers", "safetensors", "mixtral", "text-generation", "mistral", "moe", "conversational", "en", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "4-bit", "bitsandbytes", "region:us" ]
text-generation
null
null
1
[ "MixtralForCausalLM" ]
1
2
silu
4,096
0.02
14,336
32,768
mixtral
32
32
8
0.00001
1,000,000
null
false
float16
4.37.0.dev0
true
32,000
null
0
null
null
null
null
smelborp/MixtralOrochi8x7B
null
null
"2023-12-25T13:53:28Z"
null
null
795
null
null
null
null
17
transformers
[ "transformers", "safetensors", "mixtral", "text-generation", "uncensored", "high-intelligence", "en", "license:cc-by-nc-4.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "MixtralForCausalLM" ]
1
2
silu
4,096
0.02
14,336
32,768
mixtral
32
32
8
0.00001
1,000,000
null
false
bfloat16
4.36.2
true
32,000
null
0
null
null
null
null
NeverSleep/Noromaid-v0.1-mixtral-8x7b-Instruct-v3
null
null
"2023-12-25T17:04:07Z"
null
null
1,059
null
null
null
null
45
transformers
[ "transformers", "pytorch", "mixtral", "text-generation", "conversational", "license:cc-by-nc-4.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "MixtralForCausalLM" ]
1
2
silu
4,096
0.02
14,336
32,768
mixtral
32
32
8
0.00001
1,000,000
null
false
bfloat16
4.36.2
false
32,000
null
0
null
null
null
null
unsloth/mistral-7b-bnb-4bit
null
null
"2023-12-25T17:15:38Z"
null
null
10,055
null
null
null
null
24
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "unsloth", "mistral-7b", "bnb", "en", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "4-bit", "bitsandbytes", "region:us" ]
text-generation
null
null
1
[ "MistralForCausalLM" ]
1
2
silu
4,096
0.02
14,336
32,768
mistral
32
32
8
0.00001
10,000
4,096
false
bfloat16
4.44.2
true
32,000
null
0
128
null
null
null
Azure99/blossom-v4-qwen-14b
null
null
"2023-12-26T03:20:12Z"
null
null
13
null
null
null
null
3
transformers
[ "transformers", "pytorch", "qwen", "feature-extraction", "text-generation", "custom_code", "zh", "en", "dataset:Azure99/blossom-chat-v2", "dataset:Azure99/blossom-math-v3", "dataset:Azure99/blossom-wizard-v2", "dataset:Azure99/blossom-orca-v2", "license:apache-2.0", "region:us" ]
text-generation
null
null
1
[ "QWenLMHeadModel" ]
null
null
null
5,120
0.02
27,392
8,192
qwen
40
40
null
null
null
null
false
float16
4.33.2
true
152,064
null
null
null
null
null
null
NTQAI/chatntq-ja-7b-v1.0
null
null
"2023-12-26T06:22:59Z"
null
null
120
null
null
null
null
12
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "text-generation-inference", "ja", "en", "arxiv:2310.06825", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "MistralForCausalLM" ]
1
2
silu
4,096
0.02
14,336
4,096
mistral
32
32
8
0.00001
10,000
4,096
false
bfloat16
4.36.2
true
32,000
null
0
null
null
null
null
TheBloke/Nous-Hermes-2-Yi-34B-AWQ
null
null
"2023-12-26T07:55:32Z"
null
null
43
null
null
null
null
8
transformers
[ "transformers", "safetensors", "llama", "text-generation", "yi", "instruct", "finetune", "chatml", "gpt4", "synthetic data", "distillation", "conversational", "en", "base_model:NousResearch/Nous-Hermes-2-Yi-34B", "base_model:quantized:NousResearch/Nous-Hermes-2-Yi-34B", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "4-bit", "awq", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
1
7
silu
7,168
0.02
20,480
4,096
llama
56
60
8
0.00001
5,000,000
null
false
float16
4.36.2
true
64,000
false
0
null
null
1
null
Suru/Bhagvad-Gita-LLM
null
null
"2023-12-27T07:06:53Z"
null
null
67
null
null
null
null
3
transformers
[ "transformers", "pytorch", "llama", "text-generation", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
1
2
silu
4,096
0.02
11,008
4,096
llama
32
32
32
0.00001
null
null
false
float16
4.31.0
true
32,000
null
null
null
null
1
null
OpenPipe/mistral-ft-optimized-1227
null
null
"2023-12-27T14:51:48Z"
null
null
44,196
null
null
null
null
81
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "base_model:Intel/neural-chat-7b-v3-3", "base_model:finetune:Intel/neural-chat-7b-v3-3", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "MistralForCausalLM" ]
1
2
silu
4,096
0.02
14,336
32,768
mistral
32
32
8
0.00001
10,000
4,096
false
bfloat16
4.36.0
true
32,000
null
0
null
null
null
null