model_id
stringlengths 12
50
| downloads
int64 167
17M
| likes
int64 201
4.86k
| tags
stringlengths 116
738
| created_at
unknown | architectures
stringclasses 31
values | model_type
stringclasses 29
values | num_attention_heads
float64 12
128
⌀ | vocab_size
float64 32k
251k
⌀ | pad_token_id
float64 0
200k
⌀ | hidden_size
float64 896
16.4k
⌀ | intermediate_size
float64 4.1k
53.2k
⌀ | num_hidden_layers
float64 16
126
⌀ | hidden_act
stringclasses 5
values | layer_norm_eps
float64 0
0
⌀ | max_position_embeddings
float64 1.02k
1.05M
⌀ | activation_function
stringclasses 3
values | rms_norm_eps
float64 0
0
⌀ | attention_probs_dropout_prob
float64 0
0.1
⌀ | hidden_dropout_prob
float64 0
0.1
⌀ | year
int32 2.02k
2.03k
| month
int32 1
12
| day
int32 1
31
|
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
bigscience/bloom | 113,650 | 4,862 | ['transformers', 'pytorch', 'tensorboard', 'safetensors', 'bloom', 'text-generation', 'ak', 'ar', 'as', 'bm', 'bn', 'ca', 'code', 'en', 'es', 'eu', 'fon', 'fr', 'gu', 'hi', 'id', 'ig', 'ki', 'kn', 'lg', 'ln', 'ml', 'mr', 'ne', 'nso', 'ny', 'or', 'pa', 'pt', 'rn', 'rw', 'sn', 'st', 'sw', 'ta', 'te', 'tn', 'ts', 'tum', 'tw', 'ur', 'vi', 'wo', 'xh', 'yo', 'zh', 'zu', 'arxiv:2211.05100', 'arxiv:1909.08053', 'arxiv:2110.02861', 'arxiv:2108.12409', 'doi:10.57967/hf/0003', 'license:bigscience-bloom-rail-1.0', 'model-index', 'co2_eq_emissions', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2022-05-19T11:53:33Z" | ['BloomForCausalLM'] | bloom | 112 | 250,880 | 3 | null | null | null | null | null | null | null | null | null | null | 2,022 | 5 | 19 |
microsoft/phi-2 | 372,483 | 3,291 | ['transformers', 'safetensors', 'phi', 'text-generation', 'nlp', 'code', 'en', 'license:mit', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2023-12-13T21:19:59Z" | ['PhiForCausalLM'] | phi | 32 | 51,200 | null | 2,560 | 10,240 | 32 | gelu_new | 0.00001 | 2,048 | null | null | null | null | 2,023 | 12 | 13 |
openai-community/gpt2 | 16,958,734 | 2,620 | ['transformers', 'pytorch', 'tf', 'jax', 'tflite', 'rust', 'onnx', 'safetensors', 'gpt2', 'text-generation', 'exbert', 'en', 'doi:10.57967/hf/0039', 'license:mit', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2022-03-02T23:29:04Z" | ['GPT2LMHeadModel'] | gpt2 | null | 50,257 | null | null | null | null | null | null | null | gelu_new | null | null | null | 2,022 | 3 | 2 |
tiiuae/falcon-40b | 142,754 | 2,422 | ['transformers', 'pytorch', 'safetensors', 'falcon', 'text-generation', 'custom_code', 'en', 'de', 'es', 'fr', 'dataset:tiiuae/falcon-refinedweb', 'arxiv:2205.14135', 'arxiv:1911.02150', 'arxiv:2101.00027', 'arxiv:2005.14165', 'arxiv:2104.09864', 'arxiv:2306.01116', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'region:us'] | "2023-05-24T12:08:30Z" | ['FalconForCausalLM'] | falcon | 128 | 65,024 | null | 8,192 | null | 60 | null | null | null | null | null | null | null | 2,023 | 5 | 24 |
Qwen/QwQ-32B | 369,581 | 2,265 | ['transformers', 'safetensors', 'qwen2', 'text-generation', 'chat', 'conversational', 'en', 'arxiv:2309.00071', 'arxiv:2412.15115', 'base_model:Qwen/Qwen2.5-32B', 'base_model:finetune:Qwen/Qwen2.5-32B', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2025-03-05T14:16:59Z" | ['Qwen2ForCausalLM'] | qwen2 | 40 | 152,064 | null | 5,120 | 27,648 | 64 | silu | null | 40,960 | null | 0.00001 | null | null | 2,025 | 3 | 5 |
nvidia/Llama-3.1-Nemotron-70B-Instruct-HF | 262,081 | 2,028 | ['transformers', 'safetensors', 'llama', 'text-generation', 'nvidia', 'llama3.1', 'conversational', 'en', 'dataset:nvidia/HelpSteer2', 'arxiv:2410.01257', 'arxiv:2405.01481', 'arxiv:2406.08673', 'base_model:meta-llama/Llama-3.1-70B-Instruct', 'base_model:finetune:meta-llama/Llama-3.1-70B-Instruct', 'license:llama3.1', 'autotrain_compatible', 'text-generation-inference', 'region:us'] | "2024-10-12T02:37:13Z" | ['LlamaForCausalLM'] | llama | 64 | 128,256 | null | 8,192 | 28,672 | 80 | silu | null | 131,072 | null | 0.00001 | null | null | 2,024 | 10 | 12 |
microsoft/phi-4 | 493,023 | 1,905 | ['transformers', 'safetensors', 'phi3', 'text-generation', 'phi', 'nlp', 'math', 'code', 'chat', 'conversational', 'en', 'arxiv:2412.08905', 'license:mit', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2024-12-11T11:47:29Z" | ['Phi3ForCausalLM'] | phi3 | 40 | 100,352 | 100,349 | 5,120 | 17,920 | 40 | silu | null | 16,384 | null | 0.00001 | null | null | 2,024 | 12 | 11 |
Qwen/QwQ-32B-Preview | 225,986 | 1,720 | ['transformers', 'safetensors', 'qwen2', 'text-generation', 'chat', 'conversational', 'en', 'arxiv:2407.10671', 'base_model:Qwen/Qwen2.5-32B-Instruct', 'base_model:finetune:Qwen/Qwen2.5-32B-Instruct', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2024-11-27T15:50:55Z" | ['Qwen2ForCausalLM'] | qwen2 | 40 | 152,064 | null | 5,120 | 27,648 | 64 | silu | null | 32,768 | null | 0.00001 | null | null | 2,024 | 11 | 27 |
mattshumer/Reflection-Llama-3.1-70B | 691 | 1,715 | ['transformers', 'safetensors', 'llama', 'text-generation', 'conversational', 'base_model:meta-llama/Llama-3.1-70B-Instruct', 'base_model:finetune:meta-llama/Llama-3.1-70B-Instruct', 'license:llama3.1', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2024-09-05T18:29:50Z" | ['LlamaForCausalLM'] | llama | 64 | 128,262 | null | 8,192 | 28,672 | 80 | silu | null | 8,192 | null | 0.00001 | null | null | 2,024 | 9 | 5 |
Qwen/Qwen2.5-Coder-32B-Instruct | 259,126 | 1,714 | ['transformers', 'safetensors', 'qwen2', 'text-generation', 'code', 'codeqwen', 'chat', 'qwen', 'qwen-coder', 'conversational', 'en', 'arxiv:2409.12186', 'arxiv:2309.00071', 'arxiv:2407.10671', 'base_model:Qwen/Qwen2.5-Coder-32B', 'base_model:finetune:Qwen/Qwen2.5-Coder-32B', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2024-11-06T07:49:50Z" | ['Qwen2ForCausalLM'] | qwen2 | 40 | 152,064 | null | 5,120 | 27,648 | 64 | silu | null | 32,768 | null | 0.000001 | null | null | 2,024 | 11 | 6 |
microsoft/Phi-3-mini-128k-instruct | 116,893 | 1,637 | ['transformers', 'safetensors', 'phi3', 'text-generation', 'nlp', 'code', 'conversational', 'custom_code', 'en', 'license:mit', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2024-04-22T16:26:23Z" | ['Phi3ForCausalLM'] | phi3 | 32 | 32,064 | 32,000 | 3,072 | 8,192 | 32 | silu | null | 131,072 | null | 0.00001 | null | null | 2,024 | 4 | 22 |
EleutherAI/gpt-j-6b | 275,845 | 1,486 | ['transformers', 'pytorch', 'tf', 'jax', 'gptj', 'text-generation', 'causal-lm', 'en', 'dataset:EleutherAI/pile', 'arxiv:2104.09864', 'arxiv:2101.00027', 'license:apache-2.0', 'autotrain_compatible', 'endpoints_compatible', 'region:us'] | "2022-03-02T23:29:04Z" | ['GPTJForCausalLM'] | gptj | null | 50,400 | null | null | null | null | null | null | null | gelu_new | null | null | null | 2,022 | 3 | 2 |
microsoft/phi-1_5 | 108,020 | 1,329 | ['transformers', 'safetensors', 'phi', 'text-generation', 'nlp', 'code', 'en', 'arxiv:2309.05463', 'license:mit', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2023-09-10T04:03:46Z" | ['PhiForCausalLM'] | phi | 32 | 51,200 | null | 2,048 | 8,192 | 24 | gelu_new | 0.00001 | 2,048 | null | null | null | null | 2,023 | 9 | 10 |
deepseek-ai/DeepSeek-R1-Distill-Qwen-32B | 1,585,576 | 1,264 | ['transformers', 'safetensors', 'qwen2', 'text-generation', 'conversational', 'arxiv:2501.12948', 'license:mit', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2025-01-20T09:19:00Z" | ['Qwen2ForCausalLM'] | qwen2 | 40 | 152,064 | null | 5,120 | 27,648 | 64 | silu | null | 131,072 | null | 0.00001 | null | null | 2,025 | 1 | 20 |
cognitivecomputations/dolphin-2.5-mixtral-8x7b | 6,214 | 1,230 | ['transformers', 'pytorch', 'safetensors', 'mixtral', 'text-generation', 'conversational', 'en', 'dataset:ehartford/dolphin', 'dataset:jondurbin/airoboros-2.2.1', 'dataset:ehartford/dolphin-coder', 'dataset:migtissera/Synthia-v1.3', 'dataset:teknium/openhermes', 'dataset:ise-uiuc/Magicoder-OSS-Instruct-75K', 'dataset:ise-uiuc/Magicoder-Evol-Instruct-110K', 'dataset:LDJnr/Pure-Dove', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2023-12-14T00:45:47Z" | ['MixtralForCausalLM'] | mixtral | 32 | 32,002 | null | 4,096 | 14,336 | 32 | silu | null | 32,768 | null | 0.00001 | null | null | 2,023 | 12 | 14 |
TinyLlama/TinyLlama-1.1B-Chat-v1.0 | 1,077,704 | 1,187 | ['transformers', 'safetensors', 'llama', 'text-generation', 'conversational', 'en', 'dataset:cerebras/SlimPajama-627B', 'dataset:bigcode/starcoderdata', 'dataset:HuggingFaceH4/ultrachat_200k', 'dataset:HuggingFaceH4/ultrafeedback_binarized', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2023-12-30T06:27:30Z" | ['LlamaForCausalLM'] | llama | 32 | 32,000 | null | 2,048 | 5,632 | 22 | silu | null | 2,048 | null | 0.00001 | null | null | 2,023 | 12 | 30 |
tiiuae/falcon-40b-instruct | 90,022 | 1,175 | ['transformers', 'pytorch', 'falcon', 'text-generation', 'custom_code', 'en', 'dataset:tiiuae/falcon-refinedweb', 'arxiv:2205.14135', 'arxiv:1911.02150', 'arxiv:2005.14165', 'arxiv:2104.09864', 'arxiv:2306.01116', 'arxiv:2304.01196', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2023-05-25T10:14:36Z" | ['FalconForCausalLM'] | falcon | 128 | 65,024 | null | 8,192 | null | 60 | null | null | null | null | null | null | null | 2,023 | 5 | 25 |
mosaicml/mpt-7b | 28,190 | 1,168 | ['transformers', 'pytorch', 'mpt', 'text-generation', 'Composer', 'MosaicML', 'llm-foundry', 'StreamingDatasets', 'custom_code', 'dataset:mc4', 'dataset:c4', 'dataset:togethercomputer/RedPajama-Data-1T', 'dataset:bigcode/the-stack', 'dataset:allenai/s2orc', 'arxiv:2108.12409', 'arxiv:2302.13971', 'arxiv:2205.14135', 'arxiv:2010.04245', 'arxiv:1909.08053', 'arxiv:2302.06675', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'region:us'] | "2023-05-05T00:48:02Z" | ['MPTForCausalLM'] | mpt | null | 50,432 | null | null | null | null | null | null | null | null | null | null | null | 2,023 | 5 | 5 |
Qwen/Qwen2-VL-7B-Instruct | 1,255,774 | 1,156 | ['transformers', 'safetensors', 'qwen2_vl', 'image-text-to-text', 'multimodal', 'conversational', 'en', 'arxiv:2409.12191', 'arxiv:2308.12966', 'base_model:Qwen/Qwen2-VL-7B', 'base_model:finetune:Qwen/Qwen2-VL-7B', 'license:apache-2.0', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2024-08-28T09:03:13Z" | ['Qwen2VLForConditionalGeneration'] | qwen2_vl | 28 | 152,064 | null | 3,584 | 18,944 | 28 | silu | null | 32,768 | null | 0.000001 | null | null | 2,024 | 8 | 28 |
microsoft/Phi-4-multimodal-instruct | 586,127 | 1,155 | ['transformers', 'safetensors', 'phi4mm', 'text-generation', 'nlp', 'code', 'audio', 'automatic-speech-recognition', 'speech-summarization', 'speech-translation', 'visual-question-answering', 'phi-4-multimodal', 'phi', 'phi-4-mini', 'custom_code', 'multilingual', 'ar', 'zh', 'cs', 'da', 'nl', 'en', 'fi', 'fr', 'de', 'he', 'hu', 'it', 'ja', 'ko', 'no', 'pl', 'pt', 'ru', 'es', 'sv', 'th', 'tr', 'uk', 'arxiv:2407.13833', 'license:mit', 'autotrain_compatible', 'region:us'] | "2025-02-24T22:33:32Z" | ['Phi4MMForCausalLM'] | phi4mm | 24 | 200,064 | 199,999 | 3,072 | 8,192 | 32 | silu | null | 131,072 | null | 0.00001 | null | null | 2,025 | 2 | 24 |
microsoft/Phi-3-mini-4k-instruct | 902,600 | 1,153 | ['transformers', 'safetensors', 'phi3', 'text-generation', 'nlp', 'code', 'conversational', 'custom_code', 'en', 'fr', 'license:mit', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2024-04-22T16:18:17Z" | ['Phi3ForCausalLM'] | phi3 | 32 | 32,064 | 32,000 | 3,072 | 8,192 | 32 | silu | null | 4,096 | null | 0.00001 | null | null | 2,024 | 4 | 22 |
tiiuae/falcon-7b | 64,936 | 1,082 | ['transformers', 'pytorch', 'safetensors', 'falcon', 'text-generation', 'custom_code', 'en', 'dataset:tiiuae/falcon-refinedweb', 'arxiv:2205.14135', 'arxiv:1911.02150', 'arxiv:2101.00027', 'arxiv:2005.14165', 'arxiv:2104.09864', 'arxiv:2306.01116', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'region:us'] | "2023-04-24T16:36:24Z" | ['FalconForCausalLM'] | falcon | 71 | 65,024 | null | 4,544 | null | 32 | null | null | null | null | null | null | null | 2,023 | 4 | 24 |
deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B | 1,624,456 | 1,041 | ['transformers', 'safetensors', 'qwen2', 'text-generation', 'conversational', 'arxiv:2501.12948', 'license:mit', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2025-01-20T09:04:18Z" | ['Qwen2ForCausalLM'] | qwen2 | 12 | 151,936 | null | 1,536 | 8,960 | 28 | silu | null | 131,072 | null | 0.000001 | null | null | 2,025 | 1 | 20 |
microsoft/Phi-3-vision-128k-instruct | 83,521 | 957 | ['transformers', 'safetensors', 'phi3_v', 'text-generation', 'nlp', 'code', 'vision', 'conversational', 'custom_code', 'multilingual', 'license:mit', 'autotrain_compatible', 'region:us'] | "2024-05-19T15:07:39Z" | ['Phi3VForCausalLM'] | phi3_v | 32 | 32,064 | null | 3,072 | 8,192 | 32 | silu | null | 131,072 | null | 0.00001 | null | null | 2,024 | 5 | 19 |
tiiuae/falcon-7b-instruct | 171,457 | 950 | ['transformers', 'pytorch', 'coreml', 'safetensors', 'falcon', 'text-generation', 'conversational', 'custom_code', 'en', 'dataset:tiiuae/falcon-refinedweb', 'arxiv:2205.14135', 'arxiv:1911.02150', 'arxiv:2005.14165', 'arxiv:2104.09864', 'arxiv:2306.01116', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2023-04-25T06:21:01Z" | ['FalconForCausalLM'] | falcon | 71 | 65,024 | null | 4,544 | null | 32 | null | null | null | null | null | null | null | 2,023 | 4 | 25 |
TheBloke/Llama-2-7B-Chat-GGML | 3,307 | 871 | ['transformers', 'llama', 'facebook', 'meta', 'pytorch', 'llama-2', 'text-generation', 'en', 'arxiv:2307.09288', 'base_model:meta-llama/Llama-2-7b-chat-hf', 'base_model:finetune:meta-llama/Llama-2-7b-chat-hf', 'license:other', 'region:us'] | "2023-07-18T17:38:15Z" | null | llama | null | null | null | null | null | null | null | null | null | null | null | null | null | 2,023 | 7 | 18 |
microsoft/Phi-3.5-mini-instruct | 260,084 | 841 | ['transformers', 'safetensors', 'phi3', 'text-generation', 'nlp', 'code', 'conversational', 'custom_code', 'multilingual', 'arxiv:2404.14219', 'arxiv:2407.13833', 'arxiv:2403.06412', 'license:mit', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2024-08-16T20:48:26Z" | ['Phi3ForCausalLM'] | phi3 | 32 | 32,064 | 32,000 | 3,072 | 8,192 | 32 | silu | null | 131,072 | null | 0.00001 | null | null | 2,024 | 8 | 16 |
Phind/Phind-CodeLlama-34B-v2 | 11,528 | 832 | ['transformers', 'pytorch', 'llama', 'text-generation', 'code llama', 'license:llama2', 'model-index', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2023-08-28T21:29:09Z" | ['LlamaForCausalLM'] | llama | 64 | 32,000 | null | 8,192 | 22,016 | 48 | silu | null | 16,384 | null | 0.00001 | null | null | 2,023 | 8 | 28 |
mosaicml/mpt-7b-storywriter | 3,471 | 831 | ['transformers', 'pytorch', 'mpt', 'text-generation', 'Composer', 'MosaicML', 'llm-foundry', 'custom_code', 'dataset:the_pile_books3', 'arxiv:2108.12409', 'arxiv:2205.14135', 'arxiv:2302.06675', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'region:us'] | "2023-05-04T22:55:02Z" | ['MPTForCausalLM'] | mpt | null | 50,432 | null | null | null | null | null | null | null | null | null | null | null | 2,023 | 5 | 4 |
teknium/OpenHermes-2.5-Mistral-7B | 111,465 | 830 | ['transformers', 'pytorch', 'safetensors', 'mistral', 'text-generation', 'instruct', 'finetune', 'chatml', 'gpt4', 'synthetic data', 'distillation', 'conversational', 'en', 'dataset:teknium/OpenHermes-2.5', 'base_model:mistralai/Mistral-7B-v0.1', 'base_model:finetune:mistralai/Mistral-7B-v0.1', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2023-10-29T20:36:39Z" | ['MistralForCausalLM'] | mistral | 32 | 32,002 | null | 4,096 | 14,336 | 32 | silu | null | 32,768 | null | 0.00001 | null | null | 2,023 | 10 | 29 |
Qwen/Qwen2.5-72B-Instruct | 321,847 | 767 | ['transformers', 'safetensors', 'qwen2', 'text-generation', 'chat', 'conversational', 'en', 'arxiv:2309.00071', 'arxiv:2407.10671', 'base_model:Qwen/Qwen2.5-72B', 'base_model:finetune:Qwen/Qwen2.5-72B', 'license:other', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2024-09-16T11:56:31Z" | ['Qwen2ForCausalLM'] | qwen2 | 64 | 152,064 | null | 8,192 | 29,568 | 80 | silu | null | 32,768 | null | 0.000001 | null | null | 2,024 | 9 | 16 |
Qwen/Qwen-7B-Chat | 96,367 | 765 | ['transformers', 'safetensors', 'qwen', 'text-generation', 'custom_code', 'zh', 'en', 'arxiv:2309.16609', 'arxiv:2305.08322', 'arxiv:2009.03300', 'arxiv:2305.05280', 'arxiv:2210.03629', 'license:other', 'autotrain_compatible', 'region:us'] | "2023-08-03T03:01:31Z" | ['QWenLMHeadModel'] | qwen | 32 | 151,936 | null | 4,096 | 22,016 | 32 | null | null | 32,768 | null | null | null | null | 2,023 | 8 | 3 |
anon8231489123/gpt4-x-alpaca-13b-native-4bit-128g | 2,422 | 735 | ['transformers', 'pytorch', 'llama', 'text-generation', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2023-04-01T01:04:45Z" | ['LLaMAForCausalLM'] | llama | 40 | 32,001 | null | 5,120 | 13,824 | 40 | silu | null | null | null | 0.000001 | null | null | 2,023 | 4 | 1 |
Gustavosta/MagicPrompt-Stable-Diffusion | 18,878 | 723 | ['transformers', 'pytorch', 'coreml', 'safetensors', 'gpt2', 'text-generation', 'license:mit', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2022-09-17T22:34:07Z" | ['GPT2LMHeadModel'] | gpt2 | null | 50,257 | null | null | null | null | null | null | null | gelu_new | null | null | null | 2,022 | 9 | 17 |
Qwen/Qwen2-72B-Instruct | 36,650 | 713 | ['transformers', 'safetensors', 'qwen2', 'text-generation', 'chat', 'conversational', 'en', 'arxiv:2309.00071', 'base_model:Qwen/Qwen2-72B', 'base_model:finetune:Qwen/Qwen2-72B', 'license:other', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2024-05-28T03:48:49Z" | ['Qwen2ForCausalLM'] | qwen2 | 64 | 152,064 | null | 8,192 | 29,568 | 80 | silu | null | 32,768 | null | 0.000001 | null | null | 2,024 | 5 | 28 |
TheBloke/Llama-2-13B-chat-GGML | 647 | 697 | ['transformers', 'llama', 'facebook', 'meta', 'pytorch', 'llama-2', 'text-generation', 'en', 'arxiv:2307.09288', 'base_model:meta-llama/Llama-2-13b-chat-hf', 'base_model:finetune:meta-llama/Llama-2-13b-chat-hf', 'license:other', 'region:us'] | "2023-07-18T18:03:26Z" | null | llama | null | null | null | null | null | null | null | null | null | null | null | null | null | 2,023 | 7 | 18 |
togethercomputer/GPT-NeoXT-Chat-Base-20B | 5,207 | 696 | ['transformers', 'pytorch', 'gpt_neox', 'text-generation', 'en', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2023-03-03T00:24:29Z" | ['GPTNeoXForCausalLM'] | gpt_neox | 64 | 50,432 | null | 6,144 | 24,576 | 44 | gelu_fast | 0.00001 | 2,048 | null | null | 0 | 0 | 2,023 | 3 | 3 |
gradientai/Llama-3-8B-Instruct-Gradient-1048k | 5,103 | 682 | ['transformers', 'safetensors', 'llama', 'text-generation', 'meta', 'llama-3', 'conversational', 'en', 'arxiv:2309.00071', 'arxiv:2402.08268', 'arxiv:2305.14233', 'doi:10.57967/hf/3372', 'license:llama3', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2024-04-29T10:36:56Z" | ['LlamaForCausalLM'] | llama | 32 | 128,256 | null | 4,096 | 14,336 | 32 | silu | null | 1,048,576 | null | 0.00001 | null | null | 2,024 | 4 | 29 |
Open-Orca/Mistral-7B-OpenOrca | 22,652 | 681 | ['transformers', 'pytorch', 'mistral', 'text-generation', 'conversational', 'en', 'dataset:Open-Orca/OpenOrca', 'arxiv:2306.02707', 'arxiv:2301.13688', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2023-09-29T19:18:38Z" | ['MistralForCausalLM'] | mistral | 32 | 32,002 | null | 4,096 | 14,336 | 32 | silu | null | 32,768 | null | 0.00001 | null | null | 2,023 | 9 | 29 |
Qwen/Qwen2.5-VL-7B-Instruct | 3,334,765 | 678 | ['transformers', 'safetensors', 'qwen2_5_vl', 'image-text-to-text', 'multimodal', 'conversational', 'en', 'arxiv:2309.00071', 'arxiv:2409.12191', 'arxiv:2308.12966', 'license:apache-2.0', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2025-01-26T09:26:37Z" | ['Qwen2_5_VLForConditionalGeneration'] | qwen2_5_vl | 28 | 152,064 | null | 3,584 | 18,944 | 28 | silu | null | 128,000 | null | 0.000001 | null | null | 2,025 | 1 | 26 |
shenzhi-wang/Llama3-8B-Chinese-Chat | 40,553 | 677 | ['transformers', 'safetensors', 'llama', 'text-generation', 'llama-factory', 'orpo', 'conversational', 'en', 'zh', 'base_model:meta-llama/Meta-Llama-3-8B-Instruct', 'base_model:finetune:meta-llama/Meta-Llama-3-8B-Instruct', 'doi:10.57967/hf/2316', 'license:llama3', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2024-04-21T15:11:45Z" | ['LlamaForCausalLM'] | llama | 32 | 128,256 | null | 4,096 | 14,336 | 32 | silu | null | 8,192 | null | 0.00001 | null | null | 2,024 | 4 | 21 |
mistral-community/Mixtral-8x22B-v0.1 | 3,939 | 674 | ['transformers', 'safetensors', 'mixtral', 'text-generation', 'moe', 'fr', 'it', 'de', 'es', 'en', 'license:apache-2.0', 'model-index', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2024-04-10T08:43:36Z" | ['MixtralForCausalLM'] | mixtral | 48 | 32,000 | null | 6,144 | 16,384 | 56 | silu | null | 65,536 | null | 0.00001 | null | null | 2,024 | 4 | 10 |
microsoft/Phi-3.5-vision-instruct | 370,023 | 672 | ['transformers', 'safetensors', 'phi3_v', 'text-generation', 'nlp', 'code', 'vision', 'image-text-to-text', 'conversational', 'custom_code', 'multilingual', 'arxiv:2404.14219', 'license:mit', 'autotrain_compatible', 'region:us'] | "2024-08-16T23:48:22Z" | ['Phi3VForCausalLM'] | phi3_v | 32 | 32,064 | 32,000 | 3,072 | 8,192 | 32 | silu | null | 131,072 | null | 0.00001 | null | null | 2,024 | 8 | 16 |
anon8231489123/vicuna-13b-GPTQ-4bit-128g | 2,339 | 666 | ['transformers', 'pytorch', 'llama', 'text-generation', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2023-04-03T20:52:35Z" | ['LlamaForCausalLM'] | llama | 40 | 32,001 | 0 | 5,120 | 13,824 | 40 | silu | null | 2,048 | null | 0.000001 | null | null | 2,023 | 4 | 3 |
cerebras/Cerebras-GPT-13B | 2,504 | 647 | ['transformers', 'pytorch', 'gpt2', 'feature-extraction', 'causal-lm', 'text-generation', 'en', 'dataset:the_pile', 'arxiv:2304.03208', 'arxiv:2203.15556', 'arxiv:2101.00027', 'license:apache-2.0', 'text-generation-inference', 'region:us'] | "2023-03-20T20:45:54Z" | ['GPT2Model'] | gpt2 | null | 50,257 | null | null | null | null | null | null | null | gelu | null | null | null | 2,023 | 3 | 20 |
deepseek-ai/DeepSeek-R1-Distill-Llama-8B | 1,522,558 | 647 | ['transformers', 'safetensors', 'llama', 'text-generation', 'conversational', 'arxiv:2501.12948', 'license:mit', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2025-01-20T09:09:42Z" | ['LlamaForCausalLM'] | llama | 32 | 128,256 | null | 4,096 | 14,336 | 32 | silu | null | 131,072 | null | 0.00001 | null | null | 2,025 | 1 | 20 |
deepseek-ai/DeepSeek-R1-Distill-Llama-70B | 346,363 | 631 | ['transformers', 'safetensors', 'llama', 'text-generation', 'conversational', 'arxiv:2501.12948', 'license:mit', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2025-01-20T09:13:33Z" | ['LlamaForCausalLM'] | llama | 64 | 128,256 | null | 8,192 | 28,672 | 80 | silu | null | 131,072 | null | 0.00001 | null | null | 2,025 | 1 | 20 |
Qwen/Qwen2-7B-Instruct | 262,488 | 621 | ['transformers', 'safetensors', 'qwen2', 'text-generation', 'chat', 'conversational', 'en', 'arxiv:2309.00071', 'base_model:Qwen/Qwen2-7B', 'base_model:finetune:Qwen/Qwen2-7B', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2024-06-04T10:07:03Z" | ['Qwen2ForCausalLM'] | qwen2 | 28 | 152,064 | null | 3,584 | 18,944 | 28 | silu | null | 32,768 | null | 0.000001 | null | null | 2,024 | 6 | 4 |
TheBloke/Mixtral-8x7B-Instruct-v0.1-GGUF | 23,684 | 611 | ['transformers', 'gguf', 'mixtral', 'fr', 'it', 'de', 'es', 'en', 'base_model:mistralai/Mixtral-8x7B-Instruct-v0.1', 'base_model:quantized:mistralai/Mixtral-8x7B-Instruct-v0.1', 'license:apache-2.0', 'region:us', 'conversational'] | "2023-12-11T18:08:33Z" | null | mixtral | null | null | null | null | null | null | null | null | null | null | null | null | null | 2,023 | 12 | 11 |
bigcode/starcoder2-15b | 15,845 | 593 | ['transformers', 'safetensors', 'starcoder2', 'text-generation', 'code', 'dataset:bigcode/the-stack-v2-train', 'arxiv:2305.13245', 'arxiv:2205.14135', 'arxiv:2004.05150', 'arxiv:2207.14255', 'arxiv:2402.19173', 'license:bigcode-openrail-m', 'model-index', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2024-02-20T17:58:19Z" | ['Starcoder2ForCausalLM'] | starcoder2 | 48 | 49,152 | null | 6,144 | 24,576 | 40 | gelu_pytorch_tanh | null | 16,384 | null | null | null | null | 2,024 | 2 | 20 |
TheBloke/Wizard-Vicuna-30B-Uncensored-GPTQ | 1,037 | 577 | ['transformers', 'safetensors', 'llama', 'text-generation', 'uncensored', 'en', 'dataset:ehartford/wizard_vicuna_70k_unfiltered', 'base_model:cognitivecomputations/Wizard-Vicuna-30B-Uncensored', 'base_model:quantized:cognitivecomputations/Wizard-Vicuna-30B-Uncensored', 'license:other', 'autotrain_compatible', 'text-generation-inference', '4-bit', 'gptq', 'region:us'] | "2023-05-30T03:11:00Z" | ['LlamaForCausalLM'] | llama | 52 | 32,000 | 0 | 6,656 | 17,920 | 60 | silu | null | 2,048 | null | 0.000001 | null | null | 2,023 | 5 | 30 |
NousResearch/Yarn-Mistral-7b-128k | 15,415 | 572 | ['transformers', 'pytorch', 'mistral', 'text-generation', 'custom_code', 'en', 'dataset:emozilla/yarn-train-tokenized-16k-mistral', 'arxiv:2309.00071', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2023-10-31T13:15:14Z" | ['MistralForCausalLM'] | mistral | 32 | 32,000 | null | 4,096 | 14,336 | 32 | silu | null | 32,768 | null | 0.00001 | null | null | 2,023 | 10 | 31 |
Qwen/QVQ-72B-Preview | 172,407 | 563 | ['transformers', 'safetensors', 'qwen2_vl', 'image-text-to-text', 'chat', 'conversational', 'en', 'arxiv:2409.12191', 'base_model:Qwen/Qwen2-VL-72B', 'base_model:finetune:Qwen/Qwen2-VL-72B', 'license:other', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2024-12-24T10:13:22Z" | ['Qwen2VLForConditionalGeneration'] | qwen2_vl | 64 | 152,064 | null | 8,192 | 29,568 | 80 | silu | null | 128,000 | null | 0.000001 | null | null | 2,024 | 12 | 24 |
Qwen/Qwen2.5-7B-Instruct | 1,539,863 | 561 | ['transformers', 'safetensors', 'qwen2', 'text-generation', 'chat', 'conversational', 'en', 'arxiv:2309.00071', 'arxiv:2407.10671', 'base_model:Qwen/Qwen2.5-7B', 'base_model:finetune:Qwen/Qwen2.5-7B', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2024-09-16T11:55:40Z" | ['Qwen2ForCausalLM'] | qwen2 | 28 | 152,064 | null | 3,584 | 18,944 | 28 | silu | null | 32,768 | null | 0.000001 | null | null | 2,024 | 9 | 16 |
microsoft/Phi-3.5-MoE-instruct | 42,190 | 556 | ['transformers', 'safetensors', 'phimoe', 'text-generation', 'nlp', 'code', 'conversational', 'custom_code', 'multilingual', 'arxiv:2404.14219', 'arxiv:2407.13833', 'arxiv:2403.06412', 'license:mit', 'autotrain_compatible', 'region:us'] | "2024-08-17T21:34:19Z" | ['PhiMoEForCausalLM'] | phimoe | 32 | 32,064 | null | 4,096 | 6,400 | 32 | silu | null | 131,072 | null | 0.00001 | null | null | 2,024 | 8 | 17 |
EleutherAI/gpt-neox-20b | 19,438 | 555 | ['transformers', 'pytorch', 'safetensors', 'gpt_neox', 'text-generation', 'causal-lm', 'en', 'dataset:EleutherAI/pile', 'arxiv:2204.06745', 'arxiv:2101.00027', 'arxiv:2201.07311', 'arxiv:2104.09864', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2022-04-07T20:28:29Z" | ['GPTNeoXForCausalLM'] | gpt_neox | 64 | 50,432 | null | 6,144 | 24,576 | 44 | gelu_fast | 0.00001 | 2,048 | null | null | 0 | 0 | 2,022 | 4 | 7 |
nvidia/Llama3-ChatQA-1.5-8B | 6,334 | 555 | ['transformers', 'safetensors', 'llama', 'text-generation', 'nvidia', 'chatqa-1.5', 'chatqa', 'llama-3', 'pytorch', 'conversational', 'en', 'arxiv:2401.10225', 'license:llama3', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2024-04-28T21:32:50Z" | ['LlamaForCausalLM'] | llama | 32 | 128,256 | null | 4,096 | 14,336 | 32 | silu | null | 8,192 | null | 0.00001 | null | null | 2,024 | 4 | 28 |
deepseek-ai/DeepSeek-R1-Distill-Qwen-7B | 1,253,806 | 550 | ['transformers', 'safetensors', 'qwen2', 'text-generation', 'conversational', 'arxiv:2501.12948', 'license:mit', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2025-01-20T09:16:14Z" | ['Qwen2ForCausalLM'] | qwen2 | 28 | 152,064 | null | 3,584 | 18,944 | 28 | silu | null | 131,072 | null | 0.000001 | null | null | 2,025 | 1 | 20 |
togethercomputer/LLaMA-2-7B-32K | 21,533 | 538 | ['transformers', 'pytorch', 'llama', 'text-generation', 'en', 'dataset:togethercomputer/RedPajama-Data-1T', 'dataset:togethercomputer/RedPajama-Data-Instruct', 'dataset:EleutherAI/pile', 'dataset:togethercomputer/Long-Data-Collections', 'license:llama2', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2023-07-26T02:19:41Z" | ['LlamaForCausalLM'] | llama | 32 | 32,000 | 0 | 4,096 | 11,008 | 32 | silu | null | 32,768 | null | 0.00001 | null | null | 2,023 | 7 | 26 |
TheBloke/Mistral-7B-Instruct-v0.1-GGUF | 193,280 | 536 | ['transformers', 'gguf', 'mistral', 'finetuned', 'text-generation', 'base_model:mistralai/Mistral-7B-Instruct-v0.1', 'base_model:quantized:mistralai/Mistral-7B-Instruct-v0.1', 'license:apache-2.0', 'region:us'] | "2023-09-27T17:49:54Z" | null | mistral | null | null | null | null | null | null | null | null | null | null | null | null | null | 2,023 | 9 | 27 |
mosaicml/mpt-7b-chat | 90,975 | 513 | ['transformers', 'pytorch', 'mpt', 'text-generation', 'Composer', 'MosaicML', 'llm-foundry', 'custom_code', 'dataset:jeffwan/sharegpt_vicuna', 'dataset:Hello-SimpleAI/HC3', 'dataset:tatsu-lab/alpaca', 'dataset:Anthropic/hh-rlhf', 'dataset:victor123/evol_instruct_70k', 'arxiv:2205.14135', 'arxiv:2108.12409', 'arxiv:2010.04245', 'license:cc-by-nc-sa-4.0', 'autotrain_compatible', 'text-generation-inference', 'region:us'] | "2023-05-04T23:56:17Z" | ['MPTForCausalLM'] | mpt | null | 50,432 | null | null | null | null | null | null | null | null | null | null | null | 2,023 | 5 | 4 |
bigscience/bloomz | 3,398 | 511 | ['transformers', 'pytorch', 'tensorboard', 'safetensors', 'bloom', 'text-generation', 'ak', 'ar', 'as', 'bm', 'bn', 'ca', 'code', 'en', 'es', 'eu', 'fon', 'fr', 'gu', 'hi', 'id', 'ig', 'ki', 'kn', 'lg', 'ln', 'ml', 'mr', 'ne', 'nso', 'ny', 'or', 'pa', 'pt', 'rn', 'rw', 'sn', 'st', 'sw', 'ta', 'te', 'tn', 'ts', 'tum', 'tw', 'ur', 'vi', 'wo', 'xh', 'yo', 'zh', 'zu', 'dataset:bigscience/xP3', 'arxiv:2211.01786', 'license:bigscience-bloom-rail-1.0', 'model-index', 'autotrain_compatible', 'text-generation-inference', 'region:us'] | "2022-09-17T10:48:39Z" | ['BloomForCausalLM'] | bloom | 112 | 250,880 | 3 | null | null | null | null | null | null | null | null | null | null | 2,022 | 9 | 17 |
intfloat/e5-mistral-7b-instruct | 176,181 | 500 | ['sentence-transformers', 'pytorch', 'safetensors', 'mistral', 'feature-extraction', 'mteb', 'transformers', 'en', 'arxiv:2401.00368', 'arxiv:2104.08663', 'arxiv:2210.07316', 'license:mit', 'model-index', 'autotrain_compatible', 'text-generation-inference', 'text-embeddings-inference', 'endpoints_compatible', 'region:us'] | "2023-12-20T10:17:02Z" | ['MistralModel'] | mistral | 32 | 32,000 | 2 | 4,096 | 14,336 | 32 | silu | null | 32,768 | null | 0.00001 | null | null | 2,023 | 12 | 20 |
distilbert/distilgpt2 | 2,362,144 | 499 | ['transformers', 'pytorch', 'tf', 'jax', 'tflite', 'rust', 'coreml', 'safetensors', 'gpt2', 'text-generation', 'exbert', 'en', 'dataset:openwebtext', 'arxiv:1910.01108', 'arxiv:2201.08542', 'arxiv:2203.12574', 'arxiv:1910.09700', 'arxiv:1503.02531', 'license:apache-2.0', 'model-index', 'co2_eq_emissions', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2022-03-02T23:29:04Z" | ['GPT2LMHeadModel'] | gpt2 | null | 50,257 | null | null | null | null | null | null | null | gelu_new | null | null | null | 2,022 | 3 | 2 |
NousResearch/Hermes-2-Pro-Mistral-7B | 12,579 | 490 | ['transformers', 'safetensors', 'mistral', 'text-generation', 'Mistral', 'instruct', 'finetune', 'chatml', 'DPO', 'RLHF', 'gpt4', 'synthetic data', 'distillation', 'function calling', 'json mode', 'conversational', 'en', 'dataset:teknium/OpenHermes-2.5', 'base_model:mistralai/Mistral-7B-v0.1', 'base_model:finetune:mistralai/Mistral-7B-v0.1', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2024-03-11T12:55:27Z" | ['MistralForCausalLM'] | mistral | 32 | 32,032 | null | 4,096 | 14,336 | 32 | silu | null | 32,768 | null | 0.00001 | null | null | 2,024 | 3 | 11 |
chavinlo/gpt4-x-alpaca | 2,151 | 485 | ['transformers', 'pytorch', 'llama', 'text-generation', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2023-03-31T13:03:46Z" | ['LlamaForCausalLM'] | llama | 40 | 32,001 | null | 5,120 | 13,824 | 40 | silu | null | null | null | 0.000001 | null | null | 2,023 | 3 | 31 |
EleutherAI/gpt-neo-2.7B | 87,569 | 477 | ['transformers', 'pytorch', 'jax', 'rust', 'safetensors', 'gpt_neo', 'text-generation', 'text generation', 'causal-lm', 'en', 'dataset:EleutherAI/pile', 'arxiv:2101.00027', 'license:mit', 'autotrain_compatible', 'endpoints_compatible', 'region:us'] | "2022-03-02T23:29:04Z" | ['GPTNeoForCausalLM'] | gpt_neo | null | 50,257 | null | 2,560 | null | null | null | null | 2,048 | gelu_new | null | null | null | 2,022 | 3 | 2 |
mosaicml/mpt-7b-instruct | 9,929 | 469 | ['transformers', 'pytorch', 'mpt', 'text-generation', 'Composer', 'MosaicML', 'llm-foundry', 'custom_code', 'dataset:mosaicml/dolly_hhrlhf', 'arxiv:2205.14135', 'arxiv:2108.12409', 'arxiv:2010.04245', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'region:us'] | "2023-05-05T00:52:12Z" | ['MPTForCausalLM'] | mpt | null | 50,432 | null | null | null | null | null | null | null | null | null | null | null | 2,023 | 5 | 5 |
deepseek-ai/DeepSeek-R1-Distill-Qwen-14B | 677,313 | 469 | ['transformers', 'safetensors', 'qwen2', 'text-generation', 'conversational', 'arxiv:2501.12948', 'license:mit', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2025-01-20T09:18:27Z" | ['Qwen2ForCausalLM'] | qwen2 | 40 | 152,064 | null | 5,120 | 13,824 | 48 | silu | null | 131,072 | null | 0.00001 | null | null | 2,025 | 1 | 20 |
TheBloke/Llama-2-7B-Chat-GGUF | 77,189 | 458 | ['transformers', 'gguf', 'llama', 'facebook', 'meta', 'pytorch', 'llama-2', 'text-generation', 'en', 'arxiv:2307.09288', 'base_model:meta-llama/Llama-2-7b-chat-hf', 'base_model:quantized:meta-llama/Llama-2-7b-chat-hf', 'license:llama2', 'region:us'] | "2023-09-04T16:38:41Z" | null | llama | null | null | null | null | null | null | null | null | null | null | null | null | null | 2,023 | 9 | 4 |
openlm-research/open_llama_13b | 3,572 | 455 | ['transformers', 'pytorch', 'llama', 'text-generation', 'dataset:togethercomputer/RedPajama-Data-1T', 'arxiv:2302.13971', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2023-06-15T10:51:45Z" | ['LlamaForCausalLM'] | llama | 40 | 32,000 | 0 | 5,120 | 13,824 | 40 | silu | null | 2,048 | null | 0.000001 | null | null | 2,023 | 6 | 15 |
cognitivecomputations/dolphin-2.9-llama3-8b | 2,712 | 444 | ['transformers', 'safetensors', 'llama', 'text-generation', 'generated_from_trainer', 'axolotl', 'conversational', 'dataset:cognitivecomputations/Dolphin-2.9', 'dataset:teknium/OpenHermes-2.5', 'dataset:m-a-p/CodeFeedback-Filtered-Instruction', 'dataset:cognitivecomputations/dolphin-coder', 'dataset:cognitivecomputations/samantha-data', 'dataset:HuggingFaceH4/ultrachat_200k', 'dataset:microsoft/orca-math-word-problems-200k', 'dataset:abacusai/SystemChat-1.1', 'dataset:Locutusque/function-calling-chatml', 'dataset:internlm/Agent-FLAN', 'base_model:meta-llama/Meta-Llama-3-8B', 'base_model:finetune:meta-llama/Meta-Llama-3-8B', 'license:other', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2024-04-20T23:14:52Z" | ['LlamaForCausalLM'] | llama | 32 | 128,258 | null | 4,096 | 14,336 | 32 | silu | null | 8,192 | null | 0.00001 | null | null | 2,024 | 4 | 20 |
Qwen/Qwen2.5-Coder-7B-Instruct | 174,917 | 441 | ['transformers', 'safetensors', 'qwen2', 'text-generation', 'code', 'codeqwen', 'chat', 'qwen', 'qwen-coder', 'conversational', 'en', 'arxiv:2409.12186', 'arxiv:2309.00071', 'arxiv:2407.10671', 'base_model:Qwen/Qwen2.5-Coder-7B', 'base_model:finetune:Qwen/Qwen2.5-Coder-7B', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2024-09-17T13:38:49Z" | ['Qwen2ForCausalLM'] | qwen2 | 28 | 152,064 | null | 3,584 | 18,944 | 28 | silu | null | 32,768 | null | 0.000001 | null | null | 2,024 | 9 | 17 |
AWS/MistralLite | 2,499 | 430 | ['transformers', 'pytorch', 'mistral', 'text-generation', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'region:us'] | "2023-10-16T00:57:56Z" | ['MistralForCausalLM'] | mistral | 32 | 32,003 | null | 4,096 | 14,336 | 32 | silu | null | 32,768 | null | 0.00001 | null | null | 2,023 | 10 | 16 |
NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO | 8,323 | 427 | ['transformers', 'safetensors', 'mixtral', 'text-generation', 'Mixtral', 'instruct', 'finetune', 'chatml', 'DPO', 'RLHF', 'gpt4', 'synthetic data', 'distillation', 'conversational', 'en', 'dataset:teknium/OpenHermes-2.5', 'base_model:mistralai/Mixtral-8x7B-v0.1', 'base_model:finetune:mistralai/Mixtral-8x7B-v0.1', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2024-01-11T06:14:47Z" | ['MixtralForCausalLM'] | mixtral | 32 | 32,002 | null | 4,096 | 14,336 | 32 | silu | null | 32,768 | null | 0.00001 | null | null | 2,024 | 1 | 11 |
BioMistral/BioMistral-7B | 11,476 | 426 | ['transformers', 'pytorch', 'tensorboard', 'mistral', 'text-generation', 'medical', 'biology', 'conversational', 'fr', 'en', 'de', 'nl', 'es', 'pt', 'pl', 'ro', 'it', 'arxiv:2402.10373', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2024-02-14T11:33:32Z" | ['MistralForCausalLM'] | mistral | 32 | 32,000 | null | 4,096 | 14,336 | 32 | silu | null | 32,768 | null | 0.00001 | null | null | 2,024 | 2 | 14 |
TheBloke/Mixtral-8x7B-v0.1-GGUF | 30,791 | 425 | ['transformers', 'gguf', 'mixtral', 'fr', 'it', 'de', 'es', 'en', 'base_model:mistralai/Mixtral-8x7B-v0.1', 'base_model:quantized:mistralai/Mixtral-8x7B-v0.1', 'license:apache-2.0', 'region:us'] | "2023-12-11T13:23:32Z" | null | mixtral | null | null | null | null | null | null | null | null | null | null | null | null | null | 2,023 | 12 | 11 |
TheBloke/Mistral-7B-Instruct-v0.2-GGUF | 91,100 | 423 | ['transformers', 'gguf', 'mistral', 'finetuned', 'text-generation', 'arxiv:2310.06825', 'base_model:mistralai/Mistral-7B-Instruct-v0.2', 'base_model:quantized:mistralai/Mistral-7B-Instruct-v0.2', 'license:apache-2.0', 'region:us', 'conversational'] | "2023-12-11T22:18:46Z" | null | mistral | null | null | null | null | null | null | null | null | null | null | null | null | null | 2,023 | 12 | 11 |
NousResearch/Hermes-2-Pro-Llama-3-8B | 30,559 | 418 | ['transformers', 'safetensors', 'llama', 'text-generation', 'Llama-3', 'instruct', 'finetune', 'chatml', 'DPO', 'RLHF', 'gpt4', 'synthetic data', 'distillation', 'function calling', 'json mode', 'axolotl', 'conversational', 'en', 'dataset:teknium/OpenHermes-2.5', 'base_model:NousResearch/Meta-Llama-3-8B', 'base_model:finetune:NousResearch/Meta-Llama-3-8B', 'license:llama3', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2024-04-30T00:16:18Z" | ['LlamaForCausalLM'] | llama | 32 | 128,288 | null | 4,096 | 14,336 | 32 | silu | null | 8,192 | null | 0.00001 | null | null | 2,024 | 4 | 30 |
Qwen/Qwen2-VL-2B-Instruct | 1,182,625 | 406 | ['transformers', 'safetensors', 'qwen2_vl', 'image-text-to-text', 'multimodal', 'conversational', 'en', 'arxiv:2409.12191', 'arxiv:2308.12966', 'base_model:Qwen/Qwen2-VL-2B', 'base_model:finetune:Qwen/Qwen2-VL-2B', 'license:apache-2.0', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2024-08-28T09:02:15Z" | ['Qwen2VLForConditionalGeneration'] | qwen2_vl | 12 | 151,936 | null | 1,536 | 8,960 | 28 | silu | null | 32,768 | null | 0.000001 | null | null | 2,024 | 8 | 28 |
aaditya/Llama3-OpenBioLLM-70B | 24,929 | 402 | ['transformers', 'pytorch', 'llama', 'text-generation', 'llama-3', 'Mixtral', 'instruct', 'finetune', 'chatml', 'DPO', 'RLHF', 'gpt4', 'distillation', 'heathcare', 'medical', 'clinical', 'med', 'lifescience', 'Pharmaceutical', 'Pharma', 'conversational', 'en', 'arxiv:2305.18290', 'arxiv:2303.13375', 'arxiv:2212.13138', 'arxiv:2305.09617', 'arxiv:2402.07023', 'base_model:meta-llama/Meta-Llama-3-70B-Instruct', 'base_model:finetune:meta-llama/Meta-Llama-3-70B-Instruct', 'license:llama3', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2024-04-24T02:21:12Z" | ['LlamaForCausalLM'] | llama | 64 | 128,256 | null | 8,192 | 28,672 | 80 | silu | null | 8,192 | null | 0.00001 | null | null | 2,024 | 4 | 24 |
ICTNLP/Llama-3.1-8B-Omni | 7,016 | 398 | ['llama-omni', 'safetensors', 'omni_speech2s_llama', 'large language models', 'speech-language models', 'speech interaction', 'speech-to-speech', 'en', 'arxiv:2409.06666', 'base_model:meta-llama/Llama-3.1-8B-Instruct', 'base_model:finetune:meta-llama/Llama-3.1-8B-Instruct', 'license:other', 'region:us'] | "2024-09-10T05:23:09Z" | ['OmniSpeech2SLlamaForCausalLM'] | omni_speech2s_llama | 32 | 128,256 | null | 4,096 | 14,336 | 32 | silu | null | 131,072 | null | 0.00001 | null | null | 2,024 | 9 | 10 |
georgesung/llama2_7b_chat_uncensored | 4,007 | 382 | ['transformers', 'pytorch', 'tensorboard', 'safetensors', 'llama', 'text-generation', 'dataset:georgesung/wizard_vicuna_70k_unfiltered', 'license:other', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2023-07-20T10:45:03Z" | ['LlamaForCausalLM'] | llama | 32 | 32,000 | 0 | 4,096 | 11,008 | 32 | silu | null | 2,048 | null | 0.00001 | null | null | 2,023 | 7 | 20 |
microsoft/Phi-3-medium-128k-instruct | 16,786 | 381 | ['transformers', 'safetensors', 'phi3', 'text-generation', 'nlp', 'code', 'conversational', 'custom_code', 'multilingual', 'license:mit', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2024-05-07T15:27:32Z" | ['Phi3ForCausalLM'] | phi3 | 40 | 32,064 | null | 5,120 | 17,920 | 40 | silu | null | 131,072 | null | 0.00001 | null | null | 2,024 | 5 | 7 |
Qwen/Qwen-7B | 28,885 | 379 | ['transformers', 'safetensors', 'qwen', 'text-generation', 'custom_code', 'zh', 'en', 'arxiv:2309.16609', 'license:other', 'autotrain_compatible', 'region:us'] | "2023-08-03T02:51:18Z" | ['QWenLMHeadModel'] | qwen | 32 | 151,936 | null | 4,096 | 22,016 | 32 | null | null | 32,768 | null | null | null | null | 2,023 | 8 | 3 |
Qwen/Qwen2.5-VL-72B-Instruct | 335,945 | 374 | ['transformers', 'safetensors', 'qwen2_5_vl', 'image-text-to-text', 'multimodal', 'conversational', 'en', 'arxiv:2309.00071', 'arxiv:2409.12191', 'arxiv:2308.12966', 'base_model:Qwen/Qwen2.5-VL-72B-Instruct', 'base_model:finetune:Qwen/Qwen2.5-VL-72B-Instruct', 'license:other', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2025-01-27T04:12:04Z" | ['Qwen2_5_VLForConditionalGeneration'] | qwen2_5_vl | 64 | 152,064 | null | 8,192 | 29,568 | 80 | silu | null | 128,000 | null | 0.000001 | null | null | 2,025 | 1 | 27 |
Qwen/Qwen2-Audio-7B-Instruct | 150,420 | 370 | ['transformers', 'safetensors', 'qwen2_audio', 'text2text-generation', 'chat', 'audio', 'audio-text-to-text', 'en', 'arxiv:2407.10759', 'arxiv:2311.07919', 'license:apache-2.0', 'autotrain_compatible', 'endpoints_compatible', 'region:us'] | "2024-07-31T09:22:21Z" | ['Qwen2AudioForConditionalGeneration'] | qwen2_audio | null | 156,032 | null | null | null | null | null | null | null | null | null | null | null | 2,024 | 7 | 31 |
Qwen/Qwen2.5-1.5B-Instruct | 1,083,905 | 367 | ['transformers', 'safetensors', 'qwen2', 'text-generation', 'chat', 'conversational', 'en', 'arxiv:2407.10671', 'base_model:Qwen/Qwen2.5-1.5B', 'base_model:finetune:Qwen/Qwen2.5-1.5B', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2024-09-17T14:10:29Z" | ['Qwen2ForCausalLM'] | qwen2 | 12 | 151,936 | null | 1,536 | 8,960 | 28 | silu | null | 32,768 | null | 0.000001 | null | null | 2,024 | 9 | 17 |
OpenAssistant/oasst-sft-4-pythia-12b-epoch-3.5 | 498,245 | 366 | ['transformers', 'pytorch', 'gpt_neox', 'text-generation', 'sft', 'en', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2023-04-03T20:06:28Z" | ['GPTNeoXForCausalLM'] | gpt_neox | 40 | 50,288 | null | 5,120 | 20,480 | 36 | gelu | 0.00001 | 2,048 | null | null | null | null | 2,023 | 4 | 3 |
sambanovasystems/BLOOMChat-176B-v1 | 2,244 | 365 | ['transformers', 'pytorch', 'bloom', 'text-generation', 'license:other', 'autotrain_compatible', 'text-generation-inference', 'region:us'] | "2023-05-10T21:17:39Z" | ['BloomForCausalLM'] | bloom | null | 250,880 | 3 | 14,336 | null | null | null | null | null | null | null | null | null | 2,023 | 5 | 10 |
TheBloke/Llama-2-13B-chat-GPTQ | 12,448 | 362 | ['transformers', 'safetensors', 'llama', 'text-generation', 'facebook', 'meta', 'pytorch', 'llama-2', 'en', 'arxiv:2307.09288', 'base_model:meta-llama/Llama-2-13b-chat-hf', 'base_model:quantized:meta-llama/Llama-2-13b-chat-hf', 'license:llama2', 'autotrain_compatible', 'text-generation-inference', '4-bit', 'gptq', 'region:us'] | "2023-07-18T18:28:36Z" | ['LlamaForCausalLM'] | llama | 40 | 32,000 | 0 | 5,120 | 13,824 | 40 | silu | null | 4,096 | null | 0.00001 | null | null | 2,023 | 7 | 18 |
stabilityai/stablelm-tuned-alpha-7b | 4,278 | 360 | ['transformers', 'pytorch', 'gpt_neox', 'text-generation', 'causal-lm', 'en', 'dataset:dmayhem93/ChatCombined', 'dataset:tatsu-lab/alpaca', 'dataset:nomic-ai/gpt4all_prompt_generations', 'dataset:Dahoas/full-hh-rlhf', 'dataset:jeffwan/sharegpt_vicuna', 'dataset:HuggingFaceH4/databricks_dolly_15k', 'license:cc-by-nc-sa-4.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2023-04-19T04:29:58Z" | ['GPTNeoXForCausalLM'] | gpt_neox | 48 | 50,432 | null | 6,144 | 24,576 | 16 | gelu | 0.00001 | 4,096 | null | null | null | null | 2,023 | 4 | 19 |
Qwen/Qwen-14B-Chat | 2,615 | 360 | ['transformers', 'safetensors', 'qwen', 'text-generation', 'custom_code', 'zh', 'en', 'arxiv:2309.16609', 'arxiv:2305.08322', 'arxiv:2009.03300', 'arxiv:2305.05280', 'arxiv:2210.03629', 'autotrain_compatible', 'region:us'] | "2023-09-24T03:27:58Z" | ['QWenLMHeadModel'] | qwen | 40 | 152,064 | null | 5,120 | 27,392 | 40 | null | null | 8,192 | null | null | null | null | 2,023 | 9 | 24 |
microsoft/Phi-4-mini-instruct | 197,412 | 360 | ['transformers', 'safetensors', 'phi3', 'text-generation', 'nlp', 'code', 'conversational', 'custom_code', 'multilingual', 'ar', 'zh', 'cs', 'da', 'nl', 'en', 'fi', 'fr', 'de', 'he', 'hu', 'it', 'ja', 'ko', 'no', 'pl', 'pt', 'ru', 'es', 'sv', 'th', 'tr', 'uk', 'arxiv:2503.01743', 'license:mit', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2025-02-19T01:00:58Z" | ['Phi3ForCausalLM'] | phi3 | 24 | 200,064 | 199,999 | 3,072 | 8,192 | 32 | silu | null | 131,072 | null | 0.00001 | null | null | 2,025 | 2 | 19 |
microsoft/DialoGPT-medium | 320,044 | 359 | ['transformers', 'pytorch', 'tf', 'jax', 'rust', 'gpt2', 'text-generation', 'conversational', 'arxiv:1911.00536', 'license:mit', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2022-03-02T23:29:05Z" | ['GPT2LMHeadModel'] | gpt2 | null | 50,257 | null | null | null | null | null | null | null | gelu_new | null | null | null | 2,022 | 3 | 2 |
Qwen/Qwen-VL-Chat | 39,037 | 359 | ['transformers', 'pytorch', 'qwen', 'text-generation', 'custom_code', 'zh', 'en', 'arxiv:2308.12966', 'autotrain_compatible', 'region:us'] | "2023-08-20T04:45:22Z" | ['QWenLMHeadModel'] | qwen | 32 | 151,936 | null | 4,096 | 22,016 | 32 | null | null | 8,192 | null | null | null | null | 2,023 | 8 | 20 |
Qwen/Qwen-72B | 5,072 | 355 | ['transformers', 'safetensors', 'qwen', 'text-generation', 'custom_code', 'zh', 'en', 'arxiv:2309.16609', 'license:other', 'autotrain_compatible', 'region:us'] | "2023-11-26T16:16:31Z" | ['QWenLMHeadModel'] | qwen | 64 | 152,064 | null | 8,192 | 49,152 | 80 | null | null | 32,768 | null | null | null | null | 2,023 | 11 | 26 |
bigscience/bloom-560m | 184,845 | 351 | ['transformers', 'pytorch', 'jax', 'onnx', 'safetensors', 'bloom', 'text-generation', 'ak', 'ar', 'as', 'bm', 'bn', 'ca', 'code', 'en', 'es', 'eu', 'fon', 'fr', 'gu', 'hi', 'id', 'ig', 'ki', 'kn', 'lg', 'ln', 'ml', 'mr', 'ne', 'nso', 'ny', 'or', 'pa', 'pt', 'rn', 'rw', 'sn', 'st', 'sw', 'ta', 'te', 'tn', 'ts', 'tum', 'tw', 'ur', 'vi', 'wo', 'xh', 'yo', 'zh', 'zhs', 'zht', 'zu', 'arxiv:1909.08053', 'arxiv:2110.02861', 'arxiv:2108.12409', 'license:bigscience-bloom-rail-1.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2022-05-19T11:51:24Z" | ['BloomForCausalLM'] | bloom | 16 | 250,880 | 3 | null | null | null | null | null | null | null | null | null | null | 2,022 | 5 | 19 |
codellama/CodeLlama-7b-hf | 67,904 | 350 | ['transformers', 'pytorch', 'safetensors', 'llama', 'text-generation', 'llama-2', 'code', 'arxiv:2308.12950', 'license:llama2', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us'] | "2023-08-24T16:31:11Z" | ['LlamaForCausalLM'] | llama | 32 | 32,016 | null | 4,096 | 11,008 | 32 | silu | null | 16,384 | null | 0.00001 | null | null | 2,023 | 8 | 24 |
mosaicml/mpt-30b | 5,849 | 342 | ['transformers', 'pytorch', 'mpt', 'text-generation', 'Composer', 'MosaicML', 'llm-foundry', 'StreamingDatasets', 'custom_code', 'dataset:allenai/c4', 'dataset:mc4', 'dataset:togethercomputer/RedPajama-Data-1T', 'dataset:bigcode/the-stack-dedup', 'dataset:allenai/s2orc', 'arxiv:2108.12409', 'arxiv:2302.13971', 'arxiv:2205.14135', 'arxiv:2010.04245', 'arxiv:1909.08053', 'arxiv:2302.06675', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'region:us'] | "2023-06-20T16:29:39Z" | ['MPTForCausalLM'] | mpt | null | 50,432 | null | null | null | null | null | null | null | null | null | null | null | 2,023 | 6 | 20 |
Subsets and Splits