id
stringlengths 7
117
| author
stringclasses 6
values | sha
null | created_at
unknown | last_modified
null | disabled
null | downloads
int64 0
18.6M
| downloads_all_time
null | gated
bool 1
class | gguf
null | inference
null | likes
int64 0
4.77k
| library_name
stringclasses 36
values | tags
sequencelengths 1
430
| pipeline_tag
stringclasses 32
values | mask_token
null | model_index
null | trending_score
int64 0
132
| architectures
sequencelengths 1
5
⌀ | bos_token_id
int64 -1
256k
⌀ | eos_token_id
int64 -1
256k
⌀ | hidden_act
stringclasses 15
values | hidden_size
int64 1
20.5k
⌀ | initializer_range
float64 0
1
⌀ | intermediate_size
int64 1
98.3k
⌀ | max_position_embeddings
int64 8
1.05M
⌀ | model_type
stringclasses 530
values | num_attention_heads
int64 1
5k
⌀ | num_hidden_layers
int64 -1
8.93k
⌀ | num_key_value_heads
int64 1
160
⌀ | rms_norm_eps
float64 0
7
⌀ | rope_theta
float64 1k
1,000B
⌀ | sliding_window
int64 0
262k
⌀ | tie_word_embeddings
bool 2
classes | torch_dtype
stringclasses 8
values | transformers_version
stringclasses 207
values | use_cache
bool 2
classes | vocab_size
int64 -1
5.03M
⌀ | attention_bias
bool 2
classes | attention_dropout
float64 0
0.5
⌀ | head_dim
int64 2
256
⌀ | mlp_bias
bool 2
classes | pretraining_tp
int64 0
8
⌀ | rope_scaling
dict |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
tastypear/CausalLM-7B-DPO-alpha-GGUF | null | null | "2023-11-19T15:36:16Z" | null | null | 681 | null | null | null | null | 26 | transformers | [
"transformers",
"gguf",
"llama",
"llama2",
"qwen",
"text-generation",
"en",
"zh",
"dataset:JosephusCheung/GuanacoDataset",
"dataset:Open-Orca/OpenOrca",
"dataset:stingning/ultrachat",
"dataset:meta-math/MetaMathQA",
"dataset:liuhaotian/LLaVA-Instruct-150K",
"dataset:jondurbin/airoboros-3.1",
"dataset:WizardLM/WizardLM_evol_instruct_V2_196k",
"dataset:RyokoAI/ShareGPT52K",
"dataset:RyokoAI/Fandom23K",
"dataset:milashkaarshif/MoeGirlPedia_wikitext_raw_archive",
"dataset:wikipedia",
"dataset:wiki_lingua",
"dataset:fnlp/moss-003-sft-data",
"dataset:garage-bAInd/Open-Platypus",
"dataset:LDJnr/Puffin",
"dataset:openbmb/llava_zh",
"dataset:BAAI/COIG",
"dataset:TigerResearch/tigerbot-zhihu-zh-10k",
"dataset:liwu/MNBVC",
"dataset:teknium/openhermes",
"base_model:CausalLM/7B-DPO-alpha",
"base_model:quantized:CausalLM/7B-DPO-alpha",
"license:wtfpl",
"region:us"
] | text-generation | null | null | 1 | null | null | null | null | null | null | null | null | llama | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
pansophic/rocket-3B | null | null | "2023-11-19T17:14:29Z" | null | null | 275 | null | null | null | null | 82 | transformers | [
"transformers",
"pytorch",
"safetensors",
"stablelm",
"text-generation",
"en",
"arxiv:2305.18290",
"arxiv:2101.00027",
"arxiv:2305.06161",
"base_model:stabilityai/stablelm-3b-4e1t",
"base_model:finetune:stabilityai/stablelm-3b-4e1t",
"license:cc-by-sa-4.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"StableLmForCausalLM"
] | 0 | 50,279 | silu | 2,560 | 0.02 | 6,912 | 4,096 | stablelm | 32 | 32 | 32 | null | 10,000 | null | false | bfloat16 | 4.38.0 | true | 50,304 | null | null | null | null | null | null |
TinyLlama/TinyLlama-1.1B-Chat-v0.6 | null | null | "2023-11-20T08:59:23Z" | null | null | 356,751 | null | null | null | null | 89 | transformers | [
"transformers",
"safetensors",
"gguf",
"llama",
"text-generation",
"conversational",
"en",
"dataset:cerebras/SlimPajama-627B",
"dataset:bigcode/starcoderdata",
"dataset:OpenAssistant/oasst_top1_2023-08-25",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 2,048 | 0.02 | 5,632 | 2,048 | llama | 32 | 22 | 4 | 0.00001 | 10,000 | null | false | bfloat16 | 4.35.0 | false | 32,000 | false | null | null | null | 1 | null |
allenai/digital-socrates-13b | null | null | "2023-11-21T01:55:26Z" | null | null | 845 | null | null | null | null | 10 | transformers | [
"transformers",
"pytorch",
"llama",
"text-generation",
"en",
"arxiv:2311.09613",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 5,120 | 0.02 | 13,824 | 4,096 | llama | 40 | 40 | 40 | 0.00001 | 10,000 | null | false | float32 | 4.35.0.dev0 | true | 32,001 | false | null | null | null | 1 | null |
TheBloke/Qwen-7B-Chat-AWQ | null | null | "2023-11-21T09:54:47Z" | null | null | 110 | null | null | null | null | 8 | transformers | [
"transformers",
"safetensors",
"qwen",
"text-generation",
"custom_code",
"zh",
"en",
"arxiv:2309.16609",
"arxiv:2305.08322",
"arxiv:2009.03300",
"arxiv:2305.05280",
"arxiv:2210.03629",
"base_model:Qwen/Qwen-7B-Chat",
"base_model:quantized:Qwen/Qwen-7B-Chat",
"autotrain_compatible",
"4-bit",
"awq",
"region:us"
] | text-generation | null | null | 1 | [
"QWenLMHeadModel"
] | null | null | null | 4,096 | 0.02 | 22,016 | 8,192 | qwen | 32 | 32 | null | null | null | null | false | bfloat16 | 4.35.0 | true | 151,936 | null | null | null | null | 1 | null |
VamsiPranav/language-training | null | null | "2023-11-21T13:25:15Z" | null | null | 156 | null | null | null | null | 1 | transformers | [
"transformers",
"tensorboard",
"safetensors",
"gpt2",
"text-generation",
"generated_from_trainer",
"base_model:openai-community/gpt2",
"base_model:finetune:openai-community/gpt2",
"license:mit",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"GPT2LMHeadModel"
] | 0 | 0 | null | null | 0.02 | null | null | gpt2 | null | null | null | null | null | null | null | float32 | 4.35.2 | true | 4,037 | null | null | null | null | null | null |
togethercomputer/StripedHyena-Hessian-7B | null | null | "2023-11-21T15:43:25Z" | null | null | 93 | null | null | null | null | 62 | transformers | [
"transformers",
"safetensors",
"stripedhyena",
"text-generation",
"custom_code",
"en",
"arxiv:2302.10866",
"arxiv:2310.18780",
"arxiv:2311.05908",
"license:apache-2.0",
"autotrain_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"StripedHyenaModelForCausalLM"
] | null | null | null | 4,096 | null | null | null | stripedhyena | 32 | null | null | null | null | null | null | bfloat16 | null | true | 32,000 | null | null | null | null | null | null |
stabilityai/stablelm-zephyr-3b | null | null | "2023-11-21T16:25:10Z" | null | null | 9,614 | null | null | null | null | 247 | transformers | [
"transformers",
"safetensors",
"stablelm",
"text-generation",
"causal-lm",
"conversational",
"en",
"dataset:HuggingFaceH4/ultrachat_200k",
"dataset:HuggingFaceH4/ultrafeedback_binarized",
"dataset:meta-math/MetaMathQA",
"dataset:WizardLM/WizardLM_evol_instruct_V2_196k",
"dataset:Intel/orca_dpo_pairs",
"arxiv:2305.18290",
"arxiv:2306.05685",
"license:other",
"model-index",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"StableLmForCausalLM"
] | 0 | 0 | silu | 2,560 | 0.02 | 6,912 | 4,096 | stablelm | 32 | 32 | 32 | null | 10,000 | null | false | bfloat16 | 4.38.0 | true | 50,304 | null | null | null | null | null | null |
FPHam/Karen_TheEditor_V2_CREATIVE_Mistral_7B | null | null | "2023-11-21T20:52:15Z" | null | null | 375 | null | null | null | null | 22 | transformers | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"llm",
"llama",
"spellcheck",
"grammar",
"conversational",
"license:llama2",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"MistralForCausalLM"
] | 1 | 32,000 | silu | 4,096 | 0.02 | 14,336 | 32,768 | mistral | 32 | 32 | 8 | 0.00001 | 10,000 | 4,096 | false | float16 | 4.34.1 | true | 32,002 | null | null | null | null | null | null |
echo840/Monkey | null | null | "2023-11-22T09:29:28Z" | null | null | 904 | null | null | null | null | 30 | transformers | [
"transformers",
"pytorch",
"monkey",
"text-generation",
"custom_code",
"arxiv:2311.06607",
"autotrain_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"MonkeyLMHeadModel"
] | null | null | null | 4,096 | 0.02 | 22,016 | 8,192 | monkey | 32 | 32 | null | null | null | null | false | bfloat16 | 4.32.0 | false | 151,936 | null | null | null | null | null | null |
TheBloke/Synatra-RP-Orca-2-7B-v0.1-GPTQ | null | null | "2023-11-23T17:11:49Z" | null | null | 29 | null | null | null | null | 2 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"base_model:maywell/Synatra-RP-Orca-2-7b-v0.1",
"base_model:quantized:maywell/Synatra-RP-Orca-2-7b-v0.1",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"4-bit",
"gptq",
"region:us"
] | text-generation | null | null | 1 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 11,008 | 4,096 | llama | 32 | 32 | 32 | 0.00001 | 10,000 | null | false | bfloat16 | 4.35.2 | true | 32,003 | false | null | null | null | 1 | null |
second-state/Yi-34B-Chat-GGUF | null | null | "2023-11-24T02:31:29Z" | null | null | 392 | null | null | null | null | 2 | transformers | [
"transformers",
"gguf",
"llama",
"text-generation",
"base_model:01-ai/Yi-34B-Chat",
"base_model:quantized:01-ai/Yi-34B-Chat",
"license:apache-2.0",
"autotrain_compatible",
"region:us",
"conversational"
] | text-generation | null | null | 1 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 7,168 | 0.02 | 20,480 | 4,096 | llama | 56 | 60 | 8 | 0.00001 | 5,000,000 | null | false | bfloat16 | 4.35.0 | true | 64,000 | false | null | null | null | 1 | null |
ostorc/Conversational_Spanish_GPT | null | null | "2023-11-24T16:37:04Z" | null | null | 791 | null | null | null | null | 7 | transformers | [
"transformers",
"safetensors",
"gpt2",
"text-generation",
"chatbot",
"conversational",
"es",
"base_model:microsoft/DialoGPT-small",
"base_model:finetune:microsoft/DialoGPT-small",
"autotrain_compatible",
"text-generation-inference",
"region:us"
] | text-generation | null | null | 1 | [
"GPT2LMHeadModel"
] | 50,256 | 50,256 | null | null | 0.02 | null | null | gpt2 | null | null | null | null | null | null | null | float32 | 4.35.2 | true | 50,257 | null | null | null | null | null | null |
VAGOsolutions/SauerkrautLM-7b-HerO | null | null | "2023-11-24T16:55:27Z" | null | null | 882 | null | null | null | null | 32 | transformers | [
"transformers",
"pytorch",
"safetensors",
"mistral",
"text-generation",
"finetune",
"chatml",
"augmentation",
"german",
"merge",
"mergekit",
"conversational",
"en",
"de",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"MistralForCausalLM"
] | 1 | 32,000 | silu | 4,096 | 0.02 | 14,336 | 32,768 | mistral | 32 | 32 | 8 | 0.00001 | 10,000 | 4,096 | false | float16 | 4.35.1 | false | 32,002 | null | null | null | null | null | null |
TheBloke/Rose-20B-GGUF | null | null | "2023-11-24T20:34:27Z" | null | null | 435 | null | null | null | null | 25 | transformers | [
"transformers",
"gguf",
"llama",
"text-generation-inference",
"instruct",
"text-generation",
"en",
"base_model:tavtav/Rose-20B",
"base_model:quantized:tavtav/Rose-20B",
"license:llama2",
"region:us"
] | text-generation | null | null | 1 | null | null | null | null | null | null | null | null | llama | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
tokyotech-llm/Swallow-70b-hf | null | null | "2023-11-25T02:13:02Z" | null | null | 1,451 | null | null | null | null | 9 | transformers | [
"transformers",
"pytorch",
"llama",
"text-generation",
"en",
"ja",
"arxiv:2404.17790",
"arxiv:2404.17733",
"license:llama2",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 8,192 | 0.02 | 28,672 | 4,096 | llama | 64 | 80 | 8 | 0.00001 | null | null | false | bfloat16 | 4.33.2 | true | 43,176 | null | null | null | null | 1 | null |
berkeley-nest/Starling-LM-7B-alpha | null | null | "2023-11-25T17:42:15Z" | null | null | 40,678 | null | null | null | null | 553 | transformers | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"reward model",
"RLHF",
"RLAIF",
"conversational",
"en",
"dataset:berkeley-nest/Nectar",
"arxiv:2306.02231",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"MistralForCausalLM"
] | 1 | 32,000 | silu | 4,096 | 0.02 | 14,336 | 8,192 | mistral | 32 | 32 | 8 | 0.00001 | 10,000 | 4,096 | false | bfloat16 | 4.35.0 | true | 32,002 | null | null | null | null | null | null |
GeneZC/MiniChat-1.5-3B | null | null | "2023-11-26T03:37:49Z" | null | null | 1,478 | null | null | null | null | 35 | transformers | [
"transformers",
"pytorch",
"safetensors",
"llama",
"text-generation",
"conversational",
"en",
"zh",
"arxiv:2311.07052",
"arxiv:2310.05914",
"arxiv:2305.18290",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 3,072 | 0.02 | 8,192 | 4,096 | llama | 24 | 24 | 24 | 0.00001 | 10,000 | null | false | float16 | 4.33.2 | true | 49,216 | null | null | null | null | 1 | null |
listen2you002/ChartLlama-13b | null | null | "2023-11-27T06:57:18Z" | null | null | 132 | null | null | null | null | 15 | transformers | [
"transformers",
"llava",
"text-generation",
"en",
"dataset:listen2you002/ChartLlama-Dataset",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"LlavaLlamaForCausalLM"
] | 1 | 2 | silu | 5,120 | 0.02 | 13,824 | 4,096 | llava | 40 | 40 | 40 | 0.00001 | null | null | false | float16 | 4.31.0 | true | 32,000 | null | null | null | null | 1 | null |
mesolitica/mallam-1.1B-4096 | null | null | "2023-11-27T14:33:33Z" | null | null | 486 | null | null | null | null | 5 | transformers | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"ms",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"MistralForCausalLM"
] | 1 | 2 | silu | 2,048 | 0.02 | 5,632 | 32,768 | mistral | 32 | 22 | 8 | 0.00001 | 10,000 | 4,096 | false | bfloat16 | 4.36.0.dev0 | true | 32,000 | null | 0 | null | null | null | null |
google/madlad400-8b-lm | null | null | "2023-11-27T16:00:07Z" | null | null | 253 | null | null | null | null | 4 | transformers | [
"transformers",
"safetensors",
"t5",
"text-generation",
"text-generation-inference",
"custom_code",
"en",
"ru",
"es",
"fr",
"de",
"it",
"pt",
"pl",
"nl",
"vi",
"tr",
"sv",
"id",
"ro",
"cs",
"zh",
"hu",
"ja",
"th",
"fi",
"fa",
"uk",
"da",
"el",
"no",
"bg",
"sk",
"ko",
"ar",
"lt",
"ca",
"sl",
"he",
"et",
"lv",
"hi",
"sq",
"ms",
"az",
"sr",
"ta",
"hr",
"kk",
"is",
"ml",
"mr",
"te",
"af",
"gl",
"fil",
"be",
"mk",
"eu",
"bn",
"ka",
"mn",
"bs",
"uz",
"ur",
"sw",
"yue",
"ne",
"kn",
"kaa",
"gu",
"si",
"cy",
"eo",
"la",
"hy",
"ky",
"tg",
"ga",
"mt",
"my",
"km",
"tt",
"so",
"ku",
"ps",
"pa",
"rw",
"lo",
"ha",
"dv",
"fy",
"lb",
"ckb",
"mg",
"gd",
"am",
"ug",
"ht",
"grc",
"hmn",
"sd",
"jv",
"mi",
"tk",
"ceb",
"yi",
"ba",
"fo",
"or",
"xh",
"su",
"kl",
"ny",
"sm",
"sn",
"co",
"zu",
"ig",
"yo",
"pap",
"st",
"haw",
"as",
"oc",
"cv",
"lus",
"tet",
"gsw",
"sah",
"br",
"rm",
"sa",
"bo",
"om",
"se",
"ce",
"cnh",
"ilo",
"hil",
"udm",
"os",
"lg",
"ti",
"vec",
"ts",
"tyv",
"kbd",
"ee",
"iba",
"av",
"kha",
"to",
"tn",
"nso",
"fj",
"zza",
"ak",
"ada",
"otq",
"dz",
"bua",
"cfm",
"ln",
"chm",
"gn",
"krc",
"wa",
"hif",
"yua",
"srn",
"war",
"rom",
"bik",
"pam",
"sg",
"lu",
"ady",
"kbp",
"syr",
"ltg",
"myv",
"iso",
"kac",
"bho",
"ay",
"kum",
"qu",
"za",
"pag",
"ngu",
"ve",
"pck",
"zap",
"tyz",
"hui",
"bbc",
"tzo",
"tiv",
"ksd",
"gom",
"min",
"ang",
"nhe",
"bgp",
"nzi",
"nnb",
"nv",
"zxx",
"bci",
"kv",
"new",
"mps",
"alt",
"meu",
"bew",
"fon",
"iu",
"abt",
"mgh",
"mnw",
"tvl",
"dov",
"tlh",
"ho",
"kw",
"mrj",
"meo",
"crh",
"mbt",
"emp",
"ace",
"ium",
"mam",
"gym",
"mai",
"crs",
"pon",
"ubu",
"fip",
"quc",
"gv",
"kj",
"btx",
"ape",
"chk",
"rcf",
"shn",
"tzh",
"mdf",
"ppk",
"ss",
"gag",
"cab",
"kri",
"seh",
"ibb",
"tbz",
"bru",
"enq",
"ach",
"cuk",
"kmb",
"wo",
"kek",
"qub",
"tab",
"bts",
"kos",
"rwo",
"cak",
"tuc",
"bum",
"cjk",
"gil",
"stq",
"tsg",
"quh",
"mak",
"arn",
"ban",
"jiv",
"sja",
"yap",
"tcy",
"toj",
"twu",
"xal",
"amu",
"rmc",
"hus",
"nia",
"kjh",
"bm",
"guh",
"mas",
"acf",
"dtp",
"ksw",
"bzj",
"din",
"zne",
"mad",
"msi",
"mag",
"mkn",
"kg",
"lhu",
"ch",
"qvi",
"mh",
"djk",
"sus",
"mfe",
"srm",
"dyu",
"ctu",
"gui",
"pau",
"inb",
"bi",
"mni",
"guc",
"jam",
"wal",
"jac",
"bas",
"gor",
"skr",
"nyu",
"noa",
"sda",
"gub",
"nog",
"cni",
"teo",
"tdx",
"sxn",
"rki",
"nr",
"frp",
"alz",
"taj",
"lrc",
"cce",
"rn",
"jvn",
"hvn",
"nij",
"dwr",
"izz",
"msm",
"bus",
"ktu",
"chr",
"maz",
"tzj",
"suz",
"knj",
"bim",
"gvl",
"bqc",
"tca",
"pis",
"prk",
"laj",
"mel",
"qxr",
"niq",
"ahk",
"shp",
"hne",
"spp",
"koi",
"krj",
"quf",
"luz",
"agr",
"tsc",
"mqy",
"gof",
"gbm",
"miq",
"dje",
"awa",
"bjj",
"qvz",
"sjp",
"tll",
"raj",
"kjg",
"bgz",
"quy",
"cbk",
"akb",
"oj",
"ify",
"mey",
"ks",
"cac",
"brx",
"qup",
"syl",
"jax",
"ff",
"ber",
"tks",
"trp",
"mrw",
"adh",
"smt",
"srr",
"ffm",
"qvc",
"mtr",
"ann",
"aa",
"noe",
"nut",
"gyn",
"kwi",
"xmm",
"msb",
"dataset:allenai/MADLAD-400",
"arxiv:2204.02311",
"arxiv:2309.04662",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"DecoderOnlyT5Model"
] | null | 3 | null | null | null | null | null | t5 | null | null | null | null | null | null | true | null | 4.23.1 | true | 256,512 | null | null | null | null | null | null |
ceadar-ie/FinanceConnect-13B | null | null | "2023-11-28T14:26:49Z" | null | null | 751 | null | null | null | null | 14 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"summarization",
"classification",
"translation",
"NLP",
"finance",
"domain specific llm",
"conversational",
"en",
"dataset:FinTalk-19k",
"doi:10.57967/hf/1405",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 5,120 | 0.02 | 13,824 | 4,096 | llama | 40 | 40 | 40 | 0.00001 | 10,000 | null | false | float16 | 4.36.0.dev0 | true | 32,000 | false | 0 | null | null | 1 | null |
NousResearch/Nous-Hermes-2-Vision-Alpha | null | null | "2023-11-28T23:18:55Z" | null | null | 105 | null | null | null | null | 304 | transformers | [
"transformers",
"pytorch",
"llava_mistral",
"text-generation",
"mistral",
"instruct",
"finetune",
"chatml",
"gpt4",
"synthetic data",
"distillation",
"multimodal",
"llava",
"conversational",
"en",
"base_model:mistralai/Mistral-7B-v0.1",
"base_model:finetune:mistralai/Mistral-7B-v0.1",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"LlavaMistralForCausalLM"
] | 1 | 32,000 | silu | 4,096 | 0.02 | 14,336 | 32,768 | llava_mistral | 32 | 32 | 8 | 0.00001 | 10,000 | 4,096 | false | bfloat16 | 4.34.1 | true | 32,002 | null | null | null | null | null | null |
deepseek-ai/deepseek-llm-7b-chat | null | null | "2023-11-29T03:19:32Z" | null | null | 11,712 | null | null | null | null | 76 | transformers | [
"transformers",
"pytorch",
"llama",
"text-generation",
"conversational",
"license:other",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"LlamaForCausalLM"
] | 100,000 | 100,001 | silu | 4,096 | 0.02 | 11,008 | 4,096 | llama | 32 | 30 | 32 | 0.000001 | 10,000 | null | false | bfloat16 | 4.33.1 | true | 102,400 | null | null | null | null | 1 | null |
deepseek-ai/deepseek-llm-67b-base | null | null | "2023-11-29T03:23:47Z" | null | null | 1,280 | null | null | null | null | 111 | transformers | [
"transformers",
"pytorch",
"llama",
"text-generation",
"license:other",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 8,192 | 0.02 | 22,016 | 4,096 | llama | 64 | 95 | 8 | 0.000001 | 10,000 | null | false | bfloat16 | 4.33.1 | true | 102,400 | null | null | null | null | 1 | null |
deepseek-ai/deepseek-llm-67b-chat | null | null | "2023-11-29T03:30:00Z" | null | null | 1,518 | null | null | null | null | 177 | transformers | [
"transformers",
"pytorch",
"llama",
"text-generation",
"conversational",
"license:other",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"LlamaForCausalLM"
] | 100,000 | 100,001 | silu | 8,192 | 0.02 | 22,016 | 4,096 | llama | 64 | 95 | 8 | 0.000001 | 10,000 | null | false | bfloat16 | 4.33.1 | true | 102,400 | null | null | null | null | 1 | null |
athirdpath/Iambe-20b-DARE | null | null | "2023-11-29T06:31:49Z" | null | null | 19 | null | null | null | null | 11 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"license:cc-by-nc-4.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 5,120 | 0.02 | 13,824 | 4,096 | llama | 40 | 62 | 40 | 0.00001 | 10,000 | null | false | bfloat16 | 4.35.2 | false | 32,000 | false | null | null | null | 1 | null |
SUSTech/SUS-Chat-34B | null | null | "2023-11-29T09:07:53Z" | null | null | 1,029 | null | null | null | null | 122 | transformers | [
"transformers",
"pytorch",
"llama",
"text-generation",
"doi:10.57967/hf/1718",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 7,168 | 0.02 | 20,480 | 8,192 | llama | 56 | 60 | 8 | 0.00001 | 5,000,000 | null | false | bfloat16 | 4.35.0 | true | 64,000 | false | null | null | null | 1 | null |
Qwen/Qwen-72B-Chat | null | null | "2023-11-29T09:37:07Z" | null | null | 2,157 | null | null | null | null | 148 | transformers | [
"transformers",
"safetensors",
"qwen",
"text-generation",
"custom_code",
"zh",
"en",
"arxiv:2309.16609",
"arxiv:2305.08322",
"arxiv:2009.03300",
"arxiv:2307.11088",
"base_model:Qwen/Qwen-72B",
"base_model:finetune:Qwen/Qwen-72B",
"license:other",
"autotrain_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"QWenLMHeadModel"
] | null | null | null | 8,192 | 0.02 | 49,152 | 32,768 | qwen | 64 | 80 | null | null | 1,000,000 | null | false | null | 4.32.0 | true | 152,064 | null | null | null | null | null | null |
mlabonne/NeuralHermes-2.5-Mistral-7B | null | null | "2023-11-29T12:23:31Z" | null | null | 318 | null | null | null | null | 151 | transformers | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"instruct",
"finetune",
"chatml",
"gpt4",
"synthetic data",
"distillation",
"dpo",
"rlhf",
"conversational",
"en",
"dataset:mlabonne/chatml_dpo_pairs",
"base_model:teknium/OpenHermes-2-coder-Mistral-7B-v2-e4",
"base_model:finetune:teknium/OpenHermes-2-coder-Mistral-7B-v2-e4",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"MistralForCausalLM"
] | 1 | 32,000 | silu | 4,096 | 0.02 | 14,336 | 32,768 | mistral | 32 | 32 | 8 | 0.00001 | 10,000 | 4,096 | false | float16 | 4.35.2 | false | 32,002 | null | null | null | null | null | null |
LLM360/AmberChat | null | null | "2023-11-30T00:12:23Z" | null | null | 351 | null | null | null | null | 23 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"nlp",
"llm",
"en",
"dataset:WizardLM/WizardLM_evol_instruct_V2_196k",
"dataset:icybee/share_gpt_90k_v1",
"arxiv:2312.06550",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 11,008 | 2,048 | llama | 32 | 32 | 32 | 0.000001 | 10,000 | null | false | bfloat16 | 4.35.2 | true | 32,000 | false | null | null | null | 1 | null |
Qwen/Qwen-1_8B-Chat | null | null | "2023-11-30T02:56:11Z" | null | null | 57,677 | null | null | null | null | 108 | transformers | [
"transformers",
"safetensors",
"qwen",
"text-generation",
"custom_code",
"zh",
"en",
"arxiv:2309.16609",
"arxiv:2305.08322",
"arxiv:2009.03300",
"autotrain_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"QWenLMHeadModel"
] | null | null | null | 2,048 | 0.02 | 11,008 | 8,192 | qwen | 16 | 24 | null | null | null | null | false | null | 4.32.0 | true | 151,936 | null | null | null | null | null | null |
m-a-p/ChatMusician | null | null | "2023-11-30T03:48:16Z" | null | null | 283 | null | null | null | null | 116 | transformers | [
"transformers",
"pytorch",
"safetensors",
"llama",
"text-generation",
"en",
"arxiv:2402.16153",
"license:mit",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 11,008 | 2,048 | llama | 32 | 32 | 32 | 0.00001 | null | null | false | float16 | 4.32.0 | true | 32,000 | null | null | null | null | 1 | null |
maywell/Synatra-42dot-1.3B | null | null | "2023-11-30T05:11:41Z" | null | null | 3,834 | null | null | null | null | 6 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"license:cc-by-nc-4.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"LlamaForCausalLM"
] | 50,257 | 50,256 | silu | 2,048 | 0.01 | 5,632 | 8,192 | llama | 32 | 24 | 32 | 0.000001 | 10,000 | null | false | bfloat16 | 4.35.2 | false | 50,304 | false | null | null | null | 1 | null |
Qwen/Qwen-Audio-Chat | null | null | "2023-11-30T09:38:13Z" | null | null | 116,602 | null | null | null | null | 74 | transformers | [
"transformers",
"safetensors",
"qwen",
"text-generation",
"custom_code",
"zh",
"en",
"arxiv:2311.07919",
"autotrain_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"QWenLMHeadModel"
] | null | null | null | 4,096 | 0.02 | 22,016 | 2,048 | qwen | 32 | 32 | null | null | null | null | false | bfloat16 | 4.32.0 | true | 155,947 | null | null | null | null | null | null |
mzyil/llama-2-13b-mzyil-3-merged | null | null | "2023-12-03T01:13:56Z" | null | null | 13 | null | null | null | null | 1 | transformers | [
"transformers",
"pytorch",
"safetensors",
"llama",
"text-generation",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 5,120 | 0.02 | 13,824 | 4,096 | llama | 40 | 40 | 40 | 0.00001 | null | null | false | float16 | 4.31.0 | true | 32,000 | null | null | null | null | 1 | null |
mzyil/llama-2-7b-mzyil-3-merged | null | null | "2023-12-03T01:48:10Z" | null | null | 8 | null | null | null | null | 1 | transformers | [
"transformers",
"pytorch",
"safetensors",
"llama",
"text-generation",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 11,008 | 4,096 | llama | 32 | 32 | 32 | 0.00001 | null | null | false | float16 | 4.31.0 | true | 32,000 | null | null | null | null | 1 | null |
GAIR/autoj-bilingual-6b | null | null | "2023-12-03T11:31:52Z" | null | null | 16 | null | null | null | null | 5 | transformers | [
"transformers",
"pytorch",
"llama",
"text-generation",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 11,008 | 4,096 | llama | 32 | 32 | 4 | 0.00001 | 5,000,000 | null | false | float16 | 4.35.1 | true | 64,000 | false | null | null | null | 1 | null |
cmarkea/bloomz-3b-dpo-chat | null | null | "2023-12-04T08:30:58Z" | null | null | 65 | null | null | null | null | 2 | transformers | [
"transformers",
"pytorch",
"safetensors",
"bloom",
"text-generation",
"en",
"fr",
"dataset:Anthropic/hh-rlhf",
"license:bigscience-bloom-rail-1.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"BloomForCausalLM"
] | 1 | 2 | null | 2,560 | 0.02 | null | null | bloom | null | null | null | null | null | null | null | bfloat16 | 4.34.0 | true | 250,880 | null | 0 | null | null | 4 | null |
lamm-mit/SilkomeGPT | null | null | "2023-12-04T11:28:26Z" | null | null | 394 | null | null | null | null | 2 | transformers | [
"transformers",
"pytorch",
"gpt_neox",
"text-generation",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"GPTNeoXForCausalLM"
] | 0 | 2 | gelu | 1,024 | 0.02 | 4,096 | 2,048 | gpt_neox | 8 | 12 | null | null | null | null | false | float32 | 4.31.0.dev0 | false | 50,000 | null | null | null | null | null | null |
Trelis/Llama-2-7b-chat-hf-function-calling-v3 | null | null | "2023-12-04T14:42:08Z" | null | null | 267 | null | null | null | null | 40 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"facebook",
"meta",
"pytorch",
"llama-2",
"gguf",
"function-calling",
"function calling",
"conversational",
"en",
"arxiv:2307.09288",
"autotrain_compatible",
"text-generation-inference",
"region:us"
] | text-generation | null | null | 1 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 11,008 | 4,096 | llama | 32 | 32 | 32 | 0.00001 | 10,000 | null | false | bfloat16 | 4.38.1 | true | 32,000 | false | 0 | null | null | 1 | null |
togethercomputer/StripedHyena-Nous-7B | null | null | "2023-12-04T19:56:49Z" | null | null | 17,584 | null | null | null | null | 140 | transformers | [
"transformers",
"pytorch",
"safetensors",
"stripedhyena",
"text-generation",
"custom_code",
"en",
"arxiv:2302.10866",
"arxiv:2310.18780",
"arxiv:2311.05908",
"doi:10.57967/hf/1595",
"license:apache-2.0",
"autotrain_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"StripedHyenaModelForCausalLM"
] | null | null | null | 4,096 | null | null | null | stripedhyena | 32 | null | null | null | null | null | null | bfloat16 | null | true | 32,000 | null | null | null | null | null | null |
Nexusflow/NexusRaven-V2-13B | null | null | "2023-12-04T22:06:57Z" | null | null | 7,876 | null | null | null | null | 462 | transformers | [
"transformers",
"pytorch",
"llama",
"text-generation",
"function calling",
"arxiv:2308.12950",
"base_model:codellama/CodeLlama-13b-Instruct-hf",
"base_model:finetune:codellama/CodeLlama-13b-Instruct-hf",
"license:other",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 5,120 | 0.02 | 13,824 | 16,384 | llama | 40 | 40 | 40 | 0.00001 | 1,000,000 | null | false | bfloat16 | 4.33.0 | true | 32,024 | null | null | null | null | 1 | null |
OrionZheng/openmoe-base | null | null | "2023-12-08T03:29:33Z" | null | null | 593 | null | null | null | null | 4 | transformers | [
"transformers",
"pytorch",
"llama",
"text-generation",
"custom_code",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"OpenMoeForCausalLM"
] | 0 | 1 | swiglu | 768 | 0.02 | 2,048 | 2,048 | llama | 12 | 12 | 12 | 0.000001 | 10,000 | null | false | float32 | 4.34.0 | true | 256,384 | false | null | 64 | null | 1 | null |
LoneStriker/dolphin-2.2-yi-34b-200k-4.0bpw-h6-exl2 | null | null | "2023-12-08T09:05:45Z" | null | null | 19 | null | null | null | null | 1 | transformers | [
"transformers",
"pytorch",
"llama",
"text-generation",
"en",
"dataset:ehartford/dolphin",
"dataset:jondurbin/airoboros-2.2.1",
"dataset:ehartford/samantha-data",
"dataset:ehartford/WizardLM_evol_instruct_V2_196k_unfiltered_merged_split",
"license:other",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"LlamaForCausalLM"
] | 1 | 7 | silu | 7,168 | 0.02 | 20,480 | 200,000 | llama | 56 | 60 | 8 | 0.00001 | 5,000,000 | null | false | bfloat16 | 4.35.2 | true | 64,000 | false | null | null | null | 1 | null |
Weyaxi/OpenHermes-2.5-neural-chat-v3-2-Slerp | null | null | "2023-12-08T13:19:27Z" | null | null | 93 | null | null | null | null | 29 | transformers | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"merge",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"MistralForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 14,336 | 32,768 | mistral | 32 | 32 | 8 | 0.00001 | 10,000 | 4,096 | false | bfloat16 | 4.35.2 | true | 32,000 | null | null | null | null | null | null |
AdaptLLM/medicine-chat | null | null | "2023-12-09T01:49:56Z" | null | null | 3,740 | null | null | null | null | 44 | transformers | [
"transformers",
"pytorch",
"safetensors",
"llama",
"text-generation",
"biology",
"medical",
"en",
"dataset:EleutherAI/pile",
"dataset:Open-Orca/OpenOrca",
"dataset:GAIR/lima",
"dataset:WizardLM/WizardLM_evol_instruct_V2_196k",
"arxiv:2309.09530",
"arxiv:2406.14491",
"license:llama2",
"model-index",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 11,008 | 4,096 | llama | 32 | 32 | 32 | 0.00001 | null | null | false | float16 | 4.31.0.dev0 | true | 32,001 | null | null | null | null | 1 | null |
AdaptLLM/law-chat | null | null | "2023-12-09T02:42:38Z" | null | null | 1,030 | null | null | null | null | 33 | transformers | [
"transformers",
"pytorch",
"llama",
"text-generation",
"legal",
"en",
"dataset:EleutherAI/pile",
"dataset:Open-Orca/OpenOrca",
"dataset:GAIR/lima",
"dataset:WizardLM/WizardLM_evol_instruct_V2_196k",
"arxiv:2309.09530",
"arxiv:2406.14491",
"license:llama2",
"model-index",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 11,008 | 4,096 | llama | 32 | 32 | 32 | 0.00001 | null | null | false | float16 | 4.31.0.dev0 | true | 32,001 | null | null | null | null | 1 | null |
TheBloke/OrcaMaid-13B-GGUF | null | null | "2023-12-09T10:37:08Z" | null | null | 522 | null | null | null | null | 8 | transformers | [
"transformers",
"gguf",
"llama",
"text-generation",
"base_model:ddh0/OrcaMaid-13b",
"base_model:quantized:ddh0/OrcaMaid-13b",
"license:other",
"region:us"
] | text-generation | null | null | 1 | null | null | null | null | null | null | null | null | llama | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
migtissera/Synthia-MoE-v3-CE | null | null | "2023-12-11T01:28:47Z" | null | null | 35 | null | null | null | null | 16 | transformers | [
"transformers",
"pytorch",
"mistral",
"text-generation",
"custom_code",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"MixtralForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 14,336 | 32,768 | mistral | 32 | 32 | 8 | 0.00001 | 1,000,000 | null | false | float16 | 4.36.0.dev0 | false | 32,000 | null | 0 | null | null | null | null |
TinyLlama/TinyLlama-1.1B-intermediate-step-1195k-token-2.5T | null | null | "2023-12-11T06:13:09Z" | null | null | 3,814 | null | null | null | null | 50 | transformers | [
"transformers",
"pytorch",
"safetensors",
"llama",
"text-generation",
"en",
"dataset:cerebras/SlimPajama-627B",
"dataset:bigcode/starcoderdata",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 2,048 | 0.02 | 5,632 | 2,048 | llama | 32 | 22 | 4 | 0.00001 | null | null | false | float32 | 4.31.0.dev0 | true | 32,000 | null | null | null | null | 1 | null |
jamesdborin/llama2-13b-chat-4bit-AWQ | null | null | "2023-12-11T09:12:09Z" | null | null | 20 | null | null | null | null | 1 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"facebook",
"meta",
"pytorch",
"llama-2",
"en",
"arxiv:2307.09288",
"base_model:meta-llama/Llama-2-13b-chat-hf",
"base_model:quantized:meta-llama/Llama-2-13b-chat-hf",
"license:llama2",
"autotrain_compatible",
"text-generation-inference",
"4-bit",
"awq",
"region:us"
] | text-generation | null | null | 1 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 5,120 | 0.02 | 13,824 | 4,096 | llama | 40 | 40 | 40 | 0.00001 | null | null | false | float16 | 4.31.0.dev0 | true | 32,000 | null | null | null | null | 1 | null |
TheBloke/Mixtral-8x7B-Instruct-v0.1-GPTQ | null | null | "2023-12-11T18:49:53Z" | null | null | 57,771 | null | null | null | null | 134 | transformers | [
"transformers",
"safetensors",
"mixtral",
"text-generation",
"conversational",
"fr",
"it",
"de",
"es",
"en",
"base_model:mistralai/Mixtral-8x7B-Instruct-v0.1",
"base_model:quantized:mistralai/Mixtral-8x7B-Instruct-v0.1",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"4-bit",
"gptq",
"region:us"
] | text-generation | null | null | 1 | [
"MixtralForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 14,336 | 32,768 | mixtral | 32 | 32 | 8 | 0.00001 | 1,000,000 | 4,096 | false | bfloat16 | 4.36.0 | true | 32,000 | null | 0 | null | null | 1 | null |
TheBloke/Mixtral-8x7B-Instruct-v0.1-AWQ | null | null | "2023-12-11T21:01:33Z" | null | null | 383 | null | null | null | null | 58 | transformers | [
"transformers",
"safetensors",
"mixtral",
"text-generation",
"conversational",
"fr",
"it",
"de",
"es",
"en",
"base_model:mistralai/Mixtral-8x7B-Instruct-v0.1",
"base_model:quantized:mistralai/Mixtral-8x7B-Instruct-v0.1",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"4-bit",
"awq",
"region:us"
] | text-generation | null | null | 1 | [
"MixtralForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 14,336 | 32,768 | mixtral | 32 | 32 | 8 | 0.00001 | 1,000,000 | 4,096 | false | float16 | 4.36.0.dev0 | true | 32,000 | null | 0 | null | null | 1 | null |
TheBloke/Mistral-7B-Instruct-v0.2-GPTQ | null | null | "2023-12-11T22:18:46Z" | null | null | 456,687 | null | null | null | null | 50 | transformers | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"finetuned",
"conversational",
"arxiv:2310.06825",
"base_model:mistralai/Mistral-7B-Instruct-v0.2",
"base_model:quantized:mistralai/Mistral-7B-Instruct-v0.2",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"4-bit",
"gptq",
"region:us"
] | text-generation | null | null | 1 | [
"MistralForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 14,336 | 32,768 | mistral | 32 | 32 | 8 | 0.00001 | 1,000,000 | null | false | bfloat16 | 4.36.0.dev0 | true | 32,000 | null | 0 | null | null | 1 | null |
Open-Orca/Mixtral-SlimOrca-8x7B | null | null | "2023-12-12T02:30:31Z" | null | null | 36 | null | null | null | null | 52 | transformers | [
"transformers",
"safetensors",
"mixtral",
"text-generation",
"dataset:Open-Orca/SlimOrca",
"base_model:mistralai/Mixtral-8x7B-v0.1",
"base_model:finetune:mistralai/Mixtral-8x7B-v0.1",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"MixtralForCausalLM"
] | 1 | 32,000 | silu | 4,096 | 0.02 | 14,336 | 32,768 | mixtral | 32 | 32 | 8 | 0.00001 | 1,000,000 | 4,096 | false | bfloat16 | 4.36.0.dev0 | true | 32,002 | null | 0 | null | null | null | null |
wenge-research/yayi2-30b | null | null | "2023-12-12T08:40:29Z" | null | null | 80 | null | null | null | null | 75 | transformers | [
"transformers",
"pytorch",
"yayi",
"text-generation",
"custom_code",
"arxiv:2312.14862",
"arxiv:2307.09288",
"license:other",
"autotrain_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"YayiForCausalLM"
] | 1 | 2 | silu | 7,168 | 0.02 | 16,384 | 4,096 | yayi | 64 | 64 | null | 0.000001 | null | null | false | bfloat16 | 4.29.1 | true | 81,920 | null | null | null | null | null | null |
vilm/vinallama-7b-chat | null | null | "2023-12-12T16:58:23Z" | null | null | 1,011 | null | null | null | null | 21 | transformers | [
"transformers",
"pytorch",
"safetensors",
"llama",
"text-generation",
"conversational",
"vi",
"arxiv:2312.11011",
"license:llama2",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"LlamaForCausalLM"
] | 1 | 46,303 | silu | 4,096 | 0.02 | 11,008 | 4,096 | llama | 32 | 32 | 32 | 0.00001 | 10,000 | null | false | bfloat16 | 4.38.1 | false | 46,305 | false | 0 | null | null | 1 | null |
ybelkada/Mixtral-8x7B-Instruct-v0.1-AWQ | null | null | "2023-12-12T22:22:51Z" | null | null | 389 | null | null | null | null | 15 | transformers | [
"transformers",
"safetensors",
"mixtral",
"text-generation",
"conversational",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"4-bit",
"awq",
"region:us"
] | text-generation | null | null | 1 | [
"MixtralForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 14,336 | 32,768 | mixtral | 32 | 32 | 8 | 0.00001 | 1,000,000 | 4,096 | false | float16 | 4.36.0.dev0 | true | 32,000 | null | 0 | null | null | null | null |
sarvamai/OpenHathi-7B-Hi-v0.1-Base | null | null | "2023-12-13T13:41:11Z" | null | null | 1,692 | null | null | null | null | 105 | transformers | [
"transformers",
"safetensors",
"gguf",
"llama",
"text-generation",
"hi",
"license:llama2",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 11,008 | 4,096 | llama | 32 | 32 | 32 | 0.00001 | 10,000 | null | false | bfloat16 | 4.35.0 | true | 48,064 | false | null | null | null | 1 | null |
w4r10ck/SOLAR-10.7B-Instruct-v1.0-uncensored | null | null | "2023-12-14T01:33:36Z" | null | null | 167 | null | null | null | null | 30 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 14,336 | 4,096 | llama | 32 | 48 | 8 | 0.00001 | 10,000 | null | false | float16 | 4.35.2 | false | 32,000 | false | null | null | null | 1 | null |
Lin-Chen/ShareGPT4V-13B | null | null | "2023-12-14T05:55:54Z" | null | null | 425 | null | null | null | null | 33 | transformers | [
"transformers",
"pytorch",
"share4v",
"text-generation",
"arxiv:2311.12793",
"autotrain_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"Share4VLlamaForCausalLM"
] | 1 | 2 | silu | 5,120 | 0.02 | 13,824 | 4,096 | share4v | 40 | 40 | 40 | 0.00001 | null | null | false | bfloat16 | 4.31.0 | true | 32,000 | null | null | null | null | 1 | null |
joey00072/ToxicHermes-2.5-Mistral-7B | null | null | "2023-12-14T11:19:11Z" | null | null | 19 | null | null | null | null | 17 | transformers | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"instruct",
"finetune",
"chatml",
"gpt4",
"synthetic data",
"distillation",
"dpo",
"rlhf",
"conversational",
"en",
"dataset:unalignment/toxic-dpo-v0.1",
"base_model:teknium/OpenHermes-2.5-Mistral-7B",
"base_model:finetune:teknium/OpenHermes-2.5-Mistral-7B",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"MistralForCausalLM"
] | 1 | 32,000 | silu | 4,096 | 0.02 | 14,336 | 32,768 | mistral | 32 | 32 | 8 | 0.00001 | 10,000 | 4,096 | false | float16 | 4.36.1 | false | 32,002 | null | 0 | null | null | null | null |
THUDM/cogagent-chat-hf | null | null | "2023-12-15T06:12:34Z" | null | null | 10,428 | null | null | null | null | 64 | transformers | [
"transformers",
"safetensors",
"text-generation",
"custom_code",
"en",
"arxiv:2312.08914",
"arxiv:2311.03079",
"license:apache-2.0",
"autotrain_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"CogAgentForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 11,008 | 2,048 | null | 32 | 32 | null | 0.00001 | null | null | false | bfloat16 | 4.36.0.dev0 | true | 32,000 | null | null | null | null | null | null |
chargoddard/SmolLlamix-8x101M | null | null | "2023-12-15T06:25:58Z" | null | null | 970 | null | null | null | null | 12 | transformers | [
"transformers",
"safetensors",
"mixtral",
"text-generation",
"llama",
"en",
"dataset:togethercomputer/RedPajama-Data-1T-Sample",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"MixtralForCausalLM"
] | 1 | 2 | silu | 768 | 0.02 | 3,072 | 32,768 | mixtral | 24 | 6 | 8 | 0.00001 | 10,000 | 1,024 | false | bfloat16 | 4.37.0.dev0 | false | 32,128 | false | 0 | null | null | 1 | null |
VAGOsolutions/SauerkrautLM-Mixtral-8x7B-Instruct | null | null | "2023-12-15T16:01:09Z" | null | null | 77,291 | null | null | null | null | 22 | transformers | [
"transformers",
"safetensors",
"mixtral",
"text-generation",
"mistral",
"finetune",
"dpo",
"Instruct",
"augmentation",
"german",
"moe",
"conversational",
"en",
"de",
"fr",
"it",
"es",
"dataset:argilla/distilabel-math-preference-dpo",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"MixtralForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 14,336 | 32,768 | mixtral | 32 | 32 | 8 | 0.00001 | 1,000,000 | null | false | bfloat16 | 4.36.0.dev0 | true | 32,000 | null | 0 | null | null | null | null |
WhiteRabbitNeo/WhiteRabbitNeo-13B-v1 | null | null | "2023-12-17T16:26:55Z" | null | null | 1,869 | null | null | null | null | 392 | transformers | [
"transformers",
"pytorch",
"llama",
"text-generation",
"custom_code",
"license:llama2",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 5,120 | 0.02 | 13,824 | 16,384 | llama | 40 | 40 | 40 | 0.00001 | 1,000,000 | null | false | bfloat16 | 4.36.0.dev0 | false | 32,016 | false | 0 | null | null | 1 | null |
SciPhi/Sensei-7B-V1 | null | null | "2023-12-18T17:43:53Z" | null | null | 71 | null | null | null | null | 90 | transformers | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"MistralForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 14,336 | 32,768 | mistral | 32 | 32 | 8 | 0.00001 | 10,000 | 4,096 | false | float32 | 4.37.0.dev0 | false | 32,000 | null | 0 | null | null | null | null |
ignos/Mistral-T5-7B-v1 | null | null | "2023-12-18T18:09:44Z" | null | null | 784 | null | null | null | null | 8 | transformers | [
"transformers",
"pytorch",
"mistral",
"text-generation",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"MistralForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 14,336 | 32,768 | mistral | 32 | 32 | 8 | 0.00001 | 10,000 | 4,096 | false | bfloat16 | 4.37.0.dev0 | false | 32,000 | null | 0 | null | null | null | null |
danish-foundation-models/munin-7b-alpha | null | null | "2023-12-18T18:42:32Z" | null | null | 93 | null | null | null | null | 29 | transformers | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"pretrained",
"da",
"dataset:DDSC/partial-danish-gigaword-no-twitter",
"base_model:mistralai/Mistral-7B-v0.1",
"base_model:finetune:mistralai/Mistral-7B-v0.1",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"MistralForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 14,336 | 32,768 | mistral | 32 | 32 | 8 | 0.00001 | 10,000 | 4,096 | false | bfloat16 | 4.36.2 | false | 32,000 | null | 0 | null | null | null | null |
TheBloke/phi-2-GGUF | null | null | "2023-12-18T20:22:56Z" | null | null | 5,205 | null | null | null | null | 192 | transformers | [
"transformers",
"gguf",
"phi-msft",
"nlp",
"code",
"text-generation",
"en",
"base_model:microsoft/phi-2",
"base_model:quantized:microsoft/phi-2",
"license:other",
"region:us"
] | text-generation | null | null | 1 | null | null | null | null | null | null | null | null | phi-msft | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
TKDKid1000/phi-1_5-GGUF | null | null | "2023-12-19T02:28:32Z" | null | null | 224 | null | null | null | null | 6 | transformers | [
"transformers",
"pytorch",
"gguf",
"phi-msft",
"text-generation",
"nlp",
"code",
"custom_code",
"en",
"arxiv:2309.05463",
"license:other",
"autotrain_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"PhiForCausalLM"
] | null | null | null | null | 0.02 | null | null | phi-msft | null | null | null | null | null | null | false | float16 | 4.34.1 | null | 51,200 | null | null | null | null | null | null |
rinna/nekomata-14b-instruction | null | null | "2023-12-19T08:10:24Z" | null | null | 1,268 | null | null | null | null | 22 | transformers | [
"transformers",
"pytorch",
"safetensors",
"qwen",
"text-generation",
"custom_code",
"ja",
"en",
"dataset:databricks/databricks-dolly-15k",
"dataset:kunishou/databricks-dolly-15k-ja",
"dataset:izumi-lab/llm-japanese-dataset",
"arxiv:2309.16609",
"arxiv:2404.01657",
"base_model:rinna/nekomata-14b",
"base_model:finetune:rinna/nekomata-14b",
"license:other",
"autotrain_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"QWenLMHeadModel"
] | null | null | null | 5,120 | 0.02 | 27,392 | 8,192 | qwen | 40 | 40 | null | null | null | null | false | null | 4.32.0 | true | 152,064 | null | null | null | null | null | null |
ThisIs-Developer/Llama-2-GGML-Medical-Chatbot | null | null | "2023-12-19T14:51:37Z" | null | null | 241 | null | null | null | null | 19 | transformers | [
"transformers",
"llama",
"medical",
"conversational",
"text-generation",
"question-answering",
"en",
"base_model:TheBloke/Llama-2-7B-Chat-GGML",
"base_model:finetune:TheBloke/Llama-2-7B-Chat-GGML",
"license:mit",
"endpoints_compatible",
"region:us"
] | question-answering | null | null | 1 | null | null | null | null | null | null | null | null | llama | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
mlx-community/phi-2 | null | null | "2023-12-19T16:45:41Z" | null | null | 23 | null | null | null | null | 52 | mlx | [
"mlx",
"phi-msft",
"nlp",
"code",
"text-generation",
"en",
"base_model:microsoft/phi-2",
"base_model:finetune:microsoft/phi-2",
"license:other",
"region:us"
] | text-generation | null | null | 1 | null | null | null | null | null | null | null | null | phi-msft | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
AdaptLLM/law-LLM-13B | null | null | "2023-12-20T01:54:43Z" | null | null | 417 | null | null | null | null | 31 | transformers | [
"transformers",
"pytorch",
"safetensors",
"llama",
"text-generation",
"legal",
"en",
"dataset:Open-Orca/OpenOrca",
"dataset:GAIR/lima",
"dataset:WizardLM/WizardLM_evol_instruct_V2_196k",
"dataset:EleutherAI/pile",
"arxiv:2309.09530",
"arxiv:2406.14491",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 5,120 | 0.02 | 13,824 | 2,048 | llama | 40 | 40 | null | 0.000001 | null | null | false | float16 | 4.28.0.dev0 | true | 32,001 | null | null | null | null | null | null |
scb10x/typhoon-7b | null | null | "2023-12-20T14:07:35Z" | null | null | 1,462 | null | null | null | null | 100 | transformers | [
"transformers",
"pytorch",
"safetensors",
"mistral",
"text-generation",
"pretrained",
"th",
"arxiv:2312.13951",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"MistralForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 14,336 | 32,768 | mistral | 32 | 32 | 8 | 0.00001 | 10,000 | 4,096 | false | bfloat16 | 4.34.0.dev0 | true | 35,219 | null | null | null | null | null | null |
beowolx/CodeNinja-1.0-OpenChat-7B | null | null | "2023-12-20T20:28:01Z" | null | null | 4,502 | null | null | null | null | 105 | transformers | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"code",
"text-generation-inference",
"conversational",
"en",
"dataset:glaiveai/glaive-code-assistant-v2",
"dataset:TokenBender/code_instructions_122k_alpaca_style",
"doi:10.57967/hf/1535",
"license:mit",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"MistralForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 14,336 | 8,192 | mistral | 32 | 32 | 8 | 0.00001 | 10,000 | 4,096 | false | bfloat16 | 4.37.0.dev0 | false | 32,002 | null | 0 | null | null | null | null |
TheBloke/DaringMaid-20B-GGUF | null | null | "2023-12-20T21:04:04Z" | null | null | 1,040 | null | null | null | null | 18 | transformers | [
"transformers",
"gguf",
"llama",
"text-generation",
"en",
"base_model:Kooten/DaringMaid-20B",
"base_model:quantized:Kooten/DaringMaid-20B",
"license:cc-by-nc-4.0",
"region:us"
] | text-generation | null | null | 1 | null | null | null | null | null | null | null | null | llama | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
SanjiWatsuki/Loyal-Toppy-Bruins-Maid-7B-DARE | null | null | "2023-12-21T01:50:18Z" | null | null | 825 | null | null | null | null | 11 | transformers | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"merge",
"license:cc-by-nc-4.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"MistralForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 14,336 | 32,768 | mistral | 32 | 32 | 8 | 0.00001 | 10,000 | 4,096 | false | bfloat16 | 4.36.2 | true | 32,000 | null | 0 | null | null | null | null |
HillZhang/untruthful_llama2_7b | null | null | "2023-12-21T07:40:22Z" | null | null | 27 | null | null | null | null | 2 | transformers | [
"transformers",
"pytorch",
"llama",
"text-generation",
"text-generation-inference",
"English",
"en",
"dataset:pminervini/HaluEval",
"license:mit",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 11,008 | 2,048 | llama | 32 | 32 | 32 | 0.00001 | 10,000 | null | false | float16 | 4.34.1 | true | 32,000 | false | null | null | null | 1 | null |
TheBloke/OrcaMaid-v2-FIX-13B-32k-GGUF | null | null | "2023-12-21T14:10:10Z" | null | null | 525 | null | null | null | null | 5 | transformers | [
"transformers",
"gguf",
"llama",
"text-generation",
"base_model:ddh0/OrcaMaid-v2-FIX-13b-32k",
"base_model:quantized:ddh0/OrcaMaid-v2-FIX-13b-32k",
"license:other",
"region:us"
] | text-generation | null | null | 1 | null | null | null | null | null | null | null | null | llama | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
TheBloke/Llama-2-7B-ft-instruct-es-GPTQ | null | null | "2023-12-21T15:23:09Z" | null | null | 112 | null | null | null | null | 2 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"es",
"base_model:clibrain/Llama-2-7b-ft-instruct-es",
"base_model:quantized:clibrain/Llama-2-7b-ft-instruct-es",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"4-bit",
"gptq",
"region:us"
] | text-generation | null | null | 1 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 11,008 | 4,096 | llama | 32 | 32 | 32 | 0.00001 | 10,000 | null | false | float16 | 4.35.2 | true | 32,000 | false | null | null | null | 1 | null |
BEE-spoke-data/smol_llama-220M-GQA | null | null | "2023-12-22T10:48:33Z" | null | null | 2,810 | null | null | null | null | 12 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"smol_llama",
"llama2",
"en",
"dataset:JeanKaddour/minipile",
"dataset:pszemraj/simple_wikipedia_LM",
"dataset:mattymchen/refinedweb-3m",
"dataset:BEE-spoke-data/knowledge-inoc-concat-v1",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 1,024 | 0.02 | 4,096 | 2,048 | llama | 32 | 10 | 8 | 0.000001 | 10,000 | null | false | bfloat16 | 4.37.0.dev0 | true | 32,128 | false | 0 | null | null | 1 | null |
casperhansen/mixtral-instruct-awq | null | null | "2023-12-22T12:15:06Z" | null | null | 30,699 | null | null | null | null | 43 | transformers | [
"transformers",
"safetensors",
"mixtral",
"text-generation",
"conversational",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"4-bit",
"awq",
"region:us"
] | text-generation | null | null | 1 | [
"MixtralForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 14,336 | 32,768 | mixtral | 32 | 32 | 8 | 0.00001 | 1,000,000 | null | false | float16 | 4.36.2 | true | 32,000 | null | 0 | null | null | null | null |
cognitivecomputations/dolphin-2_6-phi-2 | null | null | "2023-12-23T07:51:30Z" | null | null | 264 | null | null | null | null | 192 | transformers | [
"transformers",
"pytorch",
"safetensors",
"phi-msft",
"text-generation",
"conversational",
"custom_code",
"en",
"dataset:ehartford/dolphin",
"dataset:jondurbin/airoboros-2.2.1",
"dataset:ehartford/dolphin-coder",
"dataset:teknium/openhermes",
"dataset:ise-uiuc/Magicoder-OSS-Instruct-75K",
"dataset:ise-uiuc/Magicoder-Evol-Instruct-110K",
"dataset:LDJnr/Capybara",
"license:mit",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"PhiForCausalLM"
] | null | null | null | null | 0.02 | null | null | phi-msft | null | null | null | null | null | null | false | float16 | 4.37.0.dev0 | false | 51,200 | null | null | null | null | null | null |
NousResearch/Nous-Hermes-2-Yi-34B | null | null | "2023-12-23T19:47:48Z" | null | null | 17,622 | null | null | null | null | 247 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"yi",
"instruct",
"finetune",
"chatml",
"gpt4",
"synthetic data",
"distillation",
"conversational",
"en",
"dataset:teknium/OpenHermes-2.5",
"base_model:01-ai/Yi-34B",
"base_model:finetune:01-ai/Yi-34B",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"LlamaForCausalLM"
] | 1 | 7 | silu | 7,168 | 0.02 | 20,480 | 4,096 | llama | 56 | 60 | 8 | 0.00001 | 5,000,000 | null | false | bfloat16 | 4.37.0.dev0 | false | 64,000 | false | 0 | null | null | 1 | null |
cloudyu/Mixtral_13Bx2_MOE_22B | null | null | "2023-12-24T03:05:32Z" | null | null | 41 | null | null | null | null | 12 | transformers | [
"transformers",
"safetensors",
"mixtral",
"text-generation",
"license:cc-by-nc-4.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"MixtralForCausalLM"
] | 1 | 2 | silu | 5,120 | 0.02 | 13,824 | 4,096 | mixtral | 40 | 40 | 40 | 0.00001 | 10,000 | null | false | float16 | 4.36.2 | false | 32,000 | false | 0 | null | null | 1 | null |
cygu/llama-2-7b-logit-watermark-distill-kgw-k1-gamma0.25-delta2 | null | null | "2023-12-24T07:21:40Z" | null | null | 56 | null | null | null | null | 1 | transformers | [
"transformers",
"pytorch",
"llama",
"text-generation",
"generated_from_trainer",
"dataset:openwebtext",
"arxiv:2312.04469",
"license:llama2",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 11,008 | 4,096 | llama | 32 | 32 | 32 | 0.00001 | null | null | false | float32 | 4.29.2 | true | 32,000 | null | null | null | null | 1 | null |
NeverSleep/Noromaid-v0.1-mixtral-8x7b-v3 | null | null | "2023-12-24T10:43:51Z" | null | null | 50 | null | null | null | null | 15 | transformers | [
"transformers",
"pytorch",
"mixtral",
"text-generation",
"license:cc-by-nc-4.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"MixtralForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 14,336 | 32,768 | mixtral | 32 | 32 | 8 | 0.00001 | 1,000,000 | null | false | bfloat16 | 4.36.2 | false | 32,000 | null | 0 | null | null | null | null |
TheBloke/typhoon-7B-GGUF | null | null | "2023-12-24T16:25:16Z" | null | null | 480 | null | null | null | null | 8 | transformers | [
"transformers",
"gguf",
"mistral",
"pretrained",
"text-generation",
"th",
"arxiv:2312.13951",
"base_model:scb10x/typhoon-7b",
"base_model:quantized:scb10x/typhoon-7b",
"license:apache-2.0",
"region:us"
] | text-generation | null | null | 1 | null | null | null | null | null | null | null | null | mistral | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
ybelkada/Mixtral-8x7B-Instruct-v0.1-bnb-4bit | null | null | "2023-12-25T11:09:56Z" | null | null | 125 | null | null | null | null | 59 | transformers | [
"transformers",
"safetensors",
"mixtral",
"text-generation",
"mistral",
"moe",
"conversational",
"en",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"4-bit",
"bitsandbytes",
"region:us"
] | text-generation | null | null | 1 | [
"MixtralForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 14,336 | 32,768 | mixtral | 32 | 32 | 8 | 0.00001 | 1,000,000 | null | false | float16 | 4.37.0.dev0 | true | 32,000 | null | 0 | null | null | null | null |
smelborp/MixtralOrochi8x7B | null | null | "2023-12-25T13:53:28Z" | null | null | 795 | null | null | null | null | 17 | transformers | [
"transformers",
"safetensors",
"mixtral",
"text-generation",
"uncensored",
"high-intelligence",
"en",
"license:cc-by-nc-4.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"MixtralForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 14,336 | 32,768 | mixtral | 32 | 32 | 8 | 0.00001 | 1,000,000 | null | false | bfloat16 | 4.36.2 | true | 32,000 | null | 0 | null | null | null | null |
NeverSleep/Noromaid-v0.1-mixtral-8x7b-Instruct-v3 | null | null | "2023-12-25T17:04:07Z" | null | null | 1,059 | null | null | null | null | 45 | transformers | [
"transformers",
"pytorch",
"mixtral",
"text-generation",
"conversational",
"license:cc-by-nc-4.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"MixtralForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 14,336 | 32,768 | mixtral | 32 | 32 | 8 | 0.00001 | 1,000,000 | null | false | bfloat16 | 4.36.2 | false | 32,000 | null | 0 | null | null | null | null |
unsloth/mistral-7b-bnb-4bit | null | null | "2023-12-25T17:15:38Z" | null | null | 10,055 | null | null | null | null | 24 | transformers | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"unsloth",
"mistral-7b",
"bnb",
"en",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"4-bit",
"bitsandbytes",
"region:us"
] | text-generation | null | null | 1 | [
"MistralForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 14,336 | 32,768 | mistral | 32 | 32 | 8 | 0.00001 | 10,000 | 4,096 | false | bfloat16 | 4.44.2 | true | 32,000 | null | 0 | 128 | null | null | null |
Azure99/blossom-v4-qwen-14b | null | null | "2023-12-26T03:20:12Z" | null | null | 13 | null | null | null | null | 3 | transformers | [
"transformers",
"pytorch",
"qwen",
"feature-extraction",
"text-generation",
"custom_code",
"zh",
"en",
"dataset:Azure99/blossom-chat-v2",
"dataset:Azure99/blossom-math-v3",
"dataset:Azure99/blossom-wizard-v2",
"dataset:Azure99/blossom-orca-v2",
"license:apache-2.0",
"region:us"
] | text-generation | null | null | 1 | [
"QWenLMHeadModel"
] | null | null | null | 5,120 | 0.02 | 27,392 | 8,192 | qwen | 40 | 40 | null | null | null | null | false | float16 | 4.33.2 | true | 152,064 | null | null | null | null | null | null |
NTQAI/chatntq-ja-7b-v1.0 | null | null | "2023-12-26T06:22:59Z" | null | null | 120 | null | null | null | null | 12 | transformers | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"text-generation-inference",
"ja",
"en",
"arxiv:2310.06825",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"MistralForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 14,336 | 4,096 | mistral | 32 | 32 | 8 | 0.00001 | 10,000 | 4,096 | false | bfloat16 | 4.36.2 | true | 32,000 | null | 0 | null | null | null | null |
TheBloke/Nous-Hermes-2-Yi-34B-AWQ | null | null | "2023-12-26T07:55:32Z" | null | null | 43 | null | null | null | null | 8 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"yi",
"instruct",
"finetune",
"chatml",
"gpt4",
"synthetic data",
"distillation",
"conversational",
"en",
"base_model:NousResearch/Nous-Hermes-2-Yi-34B",
"base_model:quantized:NousResearch/Nous-Hermes-2-Yi-34B",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"4-bit",
"awq",
"region:us"
] | text-generation | null | null | 1 | [
"LlamaForCausalLM"
] | 1 | 7 | silu | 7,168 | 0.02 | 20,480 | 4,096 | llama | 56 | 60 | 8 | 0.00001 | 5,000,000 | null | false | float16 | 4.36.2 | true | 64,000 | false | 0 | null | null | 1 | null |
Suru/Bhagvad-Gita-LLM | null | null | "2023-12-27T07:06:53Z" | null | null | 67 | null | null | null | null | 3 | transformers | [
"transformers",
"pytorch",
"llama",
"text-generation",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 11,008 | 4,096 | llama | 32 | 32 | 32 | 0.00001 | null | null | false | float16 | 4.31.0 | true | 32,000 | null | null | null | null | 1 | null |
OpenPipe/mistral-ft-optimized-1227 | null | null | "2023-12-27T14:51:48Z" | null | null | 44,196 | null | null | null | null | 81 | transformers | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"base_model:Intel/neural-chat-7b-v3-3",
"base_model:finetune:Intel/neural-chat-7b-v3-3",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 1 | [
"MistralForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 14,336 | 32,768 | mistral | 32 | 32 | 8 | 0.00001 | 10,000 | 4,096 | false | bfloat16 | 4.36.0 | true | 32,000 | null | 0 | null | null | null | null |