_id
stringlengths
24
24
id
stringlengths
7
122
author
stringlengths
2
41
base_models
dict
downloads
int64
0
206M
downloads_all_time
int64
0
2.81B
gated
stringclasses
3 values
created_at
timestamp[us, tz=UTC]date
2022-03-02 23:29:04
2025-11-03 14:41:54
last_modified
timestamp[us, tz=UTC]date
2020-12-11 21:34:15
2026-03-29 10:14:05
library_name
stringclasses
36 values
likes
int64
0
13.1k
trending_score
float64
0
47
model_index
stringlengths
30
911k
pipeline_tag
stringclasses
46 values
safetensors
stringlengths
30
122
siblings
listlengths
0
10k
sha
stringlengths
40
40
tags
listlengths
2
1.82k
gguf
stringclasses
789 values
config
stringlengths
2
53.7k
transformers_info
dict
card_data
stringlengths
234
905k
card
stringlengths
0
638k
spaces
null
licenses
listlengths
1
3
datasets
listlengths
1
289
languages
listlengths
1
1.81k
safetensors_params
float64
0
1,019B
gguf_params
float64
0
122B
metrics
listlengths
1
15
architectures
listlengths
1
6
tasks
listlengths
1
6
modalities
listlengths
1
4
input_modalities
listlengths
1
3
output_modalities
listlengths
1
4
6698d8a0653e4babe21e1e7d
meta-llama/Llama-3.1-8B-Instruct
meta-llama
{ "models": [ { "_id": "66944f1fe0c5c2e493a804f5", "id": "meta-llama/Llama-3.1-8B" } ], "relation": "finetune" }
8,384,864
133,735,004
manual
2024-07-18T08:56:00Z
2024-09-25T17:00:57Z
transformers
5,631
47
null
text-generation
{"parameters": {"BF16": 8030261248}, "total": 8030261248}
[ ".gitattributes", "LICENSE", "README.md", "USE_POLICY.md", "config.json", "generation_config.json", "model-00001-of-00004.safetensors", "model-00002-of-00004.safetensors", "model-00003-of-00004.safetensors", "model-00004-of-00004.safetensors", "model.safetensors.index.json", "original/consolid...
0e9e39f249a16976918f6564b8830bc894c89659
[ "transformers", "safetensors", "llama", "text-generation", "facebook", "meta", "pytorch", "llama-3", "conversational", "en", "de", "fr", "it", "pt", "hi", "es", "th", "arxiv:2204.05149", "base_model:meta-llama/Llama-3.1-8B", "base_model:finetune:meta-llama/Llama-3.1-8B", "lic...
null
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{{- bos_token }}\n{%- if custom_tools is defined %}\n {%- set tools = custom_tools %}\n{%- endif %}\n{%- if not tools_in_user_message is defined %}\n {%- set tools_in_user_messag...
{ "auto_model": "AutoModelForCausalLM", "custom_class": null, "pipeline_tag": "text-generation", "processor": "AutoTokenizer" }
{"base_model": "meta-llama/Meta-Llama-3.1-8B", "datasets": null, "eval_results": null, "language": ["en", "de", "fr", "it", "pt", "hi", "es", "th"], "library_name": null, "license": "llama3.1", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "text-generation", "tags": ["...
null
null
[ "llama3.1" ]
null
[ "en", "de", "fr", "it", "pt", "hi", "es", "th" ]
8,030,261,248
null
null
[ "AutoModelForCausalLM", "llama", "LlamaForCausalLM" ]
[ "text-generation" ]
[ "text" ]
[ "text" ]
[ "text" ]
654a84cadff2f49007ce6c37
openai/whisper-large-v3
openai
null
4,832,162
115,357,760
False
2023-11-07T18:41:14Z
2024-08-12T10:20:10Z
transformers
5,522
30
null
automatic-speech-recognition
null
[ ".gitattributes", "README.md", "added_tokens.json", "config.json", "flax_model.msgpack", "generation_config.json", "merges.txt", "model.fp32-00001-of-00002.safetensors", "model.fp32-00002-of-00002.safetensors", "model.safetensors", "model.safetensors.index.fp32.json", "normalizer.json", "pre...
06f233fe06e710322aca913c1bc4249a0d71fce1
[ "transformers", "pytorch", "jax", "safetensors", "whisper", "automatic-speech-recognition", "audio", "hf-asr-leaderboard", "en", "zh", "de", "es", "ru", "ko", "fr", "ja", "pt", "tr", "pl", "ca", "nl", "ar", "sv", "it", "id", "hi", "fi", "vi", "he", "uk", "...
null
{"architectures": ["WhisperForConditionalGeneration"], "model_type": "whisper", "tokenizer_config": {"bos_token": "<|endoftext|>", "eos_token": "<|endoftext|>", "pad_token": "<|endoftext|>", "unk_token": "<|endoftext|>"}}
{ "auto_model": "AutoModelForSpeechSeq2Seq", "custom_class": null, "pipeline_tag": "automatic-speech-recognition", "processor": "AutoProcessor" }
{"base_model": null, "datasets": null, "eval_results": null, "language": ["en", "zh", "de", "es", "ru", "ko", "fr", "ja", "pt", "tr", "pl", "ca", "nl", "ar", "sv", "it", "id", "hi", "fi", "vi", "he", "uk", "el", "ms", "cs", "ro", "da", "hu", "ta", "no", "th", "ur", "hr", "bg", "lt", "la", "mi", "ml", "cy", "sk", "te", ...
# Whisper Whisper is a state-of-the-art model for automatic speech recognition (ASR) and speech translation, proposed in the paper [Robust Speech Recognition via Large-Scale Weak Supervision](https://huggingface.co/papers/2212.04356) by Alec Radford et al. from OpenAI. Trained on >5M hours of labeled data, Whisper d...
null
[ "apache-2.0" ]
null
[ "en", "zh", "de", "es", "ru", "ko", "fr", "ja", "pt", "tr", "pl", "ca", "nl", "ar", "sv", "it", "id", "hi", "fi", "vi", "he", "uk", "el", "ms", "cs", "ro", "da", "hu", "ta", "no", "th", "ur", "hr", "bg", "lt", "la", "mi", "ml", "cy", "sk"...
null
null
null
[ "WhisperForConditionalGeneration", "whisper", "AutoModelForSpeechSeq2Seq" ]
[ "automatic-speech-recognition" ]
[ "multimodal" ]
[ "audio" ]
[ "text" ]
68913522f16f3c8aaffccf1f
openai/gpt-oss-120b
openai
null
4,363,728
29,644,344
False
2025-08-04T22:33:06Z
2025-08-26T17:25:03Z
transformers
4,621
23
null
text-generation
{"parameters": {"BF16": 2167371072, "U8": 118244966400}, "total": 120412337472}
[ ".gitattributes", "LICENSE", "README.md", "USAGE_POLICY", "chat_template.jinja", "config.json", "generation_config.json", "metal/model.bin", "model-00000-of-00014.safetensors", "model-00001-of-00014.safetensors", "model-00002-of-00014.safetensors", "model-00003-of-00014.safetensors", "model-...
b5c939de8f754692c1647ca79fbf85e8c1e70f8a
[ "transformers", "safetensors", "gpt_oss", "text-generation", "vllm", "conversational", "arxiv:2508.10925", "license:apache-2.0", "eval-results", "endpoints_compatible", "8-bit", "mxfp4", "deploy:azure", "region:us" ]
null
{"architectures": ["GptOssForCausalLM"], "model_type": "gpt_oss", "quantization_config": {"quant_method": "mxfp4"}, "tokenizer_config": {"bos_token": "<|startoftext|>", "eos_token": "<|return|>", "pad_token": "<|endoftext|>"}, "chat_template_jinja": "{#-\n In addition to the normal inputs of `messages` and `tools`, th...
{ "auto_model": "AutoModelForCausalLM", "custom_class": null, "pipeline_tag": "text-generation", "processor": "AutoTokenizer" }
{"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "apache-2.0", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "text-generation", "tags": ["vllm"]}
<p align="center"> <img alt="gpt-oss-120b" src="https://raw.githubusercontent.com/openai/gpt-oss/main/docs/gpt-oss-120b.svg"> </p> <p align="center"> <a href="https://gpt-oss.com"><strong>Try gpt-oss</strong></a> · <a href="https://cookbook.openai.com/topic/gpt-oss"><strong>Guides</strong></a> · <a href="https...
null
[ "apache-2.0" ]
null
null
120,412,337,472
null
null
[ "AutoModelForCausalLM", "GptOssForCausalLM", "gpt_oss" ]
[ "text-generation" ]
[ "text" ]
[ "text" ]
[ "text" ]
621ffdc036468d709f17434d
openai-community/gpt2
openai-community
null
11,636,493
836,599,686
False
2022-03-02T23:29:04Z
2024-02-19T10:57:45Z
transformers
3,153
22
null
text-generation
{"parameters": {"F32": 137022720}, "total": 137022720}
[ ".gitattributes", "64-8bits.tflite", "64-fp16.tflite", "64.tflite", "README.md", "config.json", "flax_model.msgpack", "generation_config.json", "merges.txt", "model.safetensors", "onnx/config.json", "onnx/decoder_model.onnx", "onnx/decoder_model_merged.onnx", "onnx/decoder_with_past_model....
607a30d783dfa663caf39e06633721c8d4cfcd7e
[ "transformers", "pytorch", "tf", "jax", "tflite", "rust", "onnx", "safetensors", "gpt2", "text-generation", "exbert", "en", "doi:10.57967/hf/0039", "license:mit", "text-generation-inference", "endpoints_compatible", "deploy:azure", "region:us" ]
null
{"architectures": ["GPT2LMHeadModel"], "model_type": "gpt2", "tokenizer_config": {}}
{ "auto_model": "AutoModelForCausalLM", "custom_class": null, "pipeline_tag": "text-generation", "processor": "AutoTokenizer" }
{"base_model": null, "datasets": null, "eval_results": null, "language": "en", "library_name": null, "license": "mit", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": null, "tags": ["exbert"]}
# GPT-2 Test the whole generation capabilities here: https://transformer.huggingface.co/doc/gpt2-large Pretrained model on English language using a causal language modeling (CLM) objective. It was introduced in [this paper](https://d4mucfpksywv.cloudfront.net/better-language-models/language_models_are_unsupervised_mu...
null
[ "mit" ]
null
[ "en" ]
137,022,720
null
null
[ "GPT2LMHeadModel", "AutoModelForCausalLM", "gpt2" ]
[ "text-generation" ]
[ "text" ]
[ "text" ]
[ "text" ]
66e81cbcd683a3f4e5291bbf
Qwen/Qwen2.5-7B-Instruct
Qwen
{ "models": [ { "_id": "66e6d06401e2adb00b0be996", "id": "Qwen/Qwen2.5-7B" } ], "relation": "finetune" }
17,917,551
109,933,464
False
2024-09-16T11:55:40Z
2025-01-12T02:10:10Z
transformers
1,165
22
null
text-generation
{"parameters": {"BF16": 7615616512}, "total": 7615616512}
[ ".gitattributes", "LICENSE", "README.md", "config.json", "generation_config.json", "merges.txt", "model-00001-of-00004.safetensors", "model-00002-of-00004.safetensors", "model-00003-of-00004.safetensors", "model-00004-of-00004.safetensors", "model.safetensors.index.json", "tokenizer.json", "...
a09a35458c702b33eeacc393d103063234e8bc28
[ "transformers", "safetensors", "qwen2", "text-generation", "chat", "conversational", "en", "arxiv:2309.00071", "arxiv:2407.10671", "base_model:Qwen/Qwen2.5-7B", "base_model:finetune:Qwen/Qwen2.5-7B", "license:apache-2.0", "text-generation-inference", "endpoints_compatible", "deploy:azure...
null
{"architectures": ["Qwen2ForCausalLM"], "model_type": "qwen2", "tokenizer_config": {"bos_token": null, "chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0]['role'] == 'system' %}\n {{- messages[0]['content'] }}\n {%- else %}\n {{- 'You are Qwen, created by Aliba...
{ "auto_model": "AutoModelForCausalLM", "custom_class": null, "pipeline_tag": "text-generation", "processor": "AutoTokenizer" }
{"base_model": "Qwen/Qwen2.5-7B", "datasets": null, "eval_results": null, "language": ["en"], "library_name": "transformers", "license": "apache-2.0", "license_name": null, "license_link": "https://huggingface.co/Qwen/Qwen2.5-7B-Instruct/blob/main/LICENSE", "metrics": null, "model_name": null, "pipeline_tag": "text-gen...
# Qwen2.5-7B-Instruct <a href="https://chat.qwenlm.ai/" target="_blank" style="margin: 2px;"> <img alt="Chat" src="https://img.shields.io/badge/%F0%9F%92%9C%EF%B8%8F%20Qwen%20Chat%20-536af5" style="display: inline-block; vertical-align: middle;"/> </a> ## Introduction Qwen2.5 is the latest series of Qwen large la...
null
[ "apache-2.0", "https://huggingface.co/Qwen/Qwen2.5-7B-Instruct/blob/main/LICENSE" ]
null
[ "en" ]
7,615,616,512
null
null
[ "AutoModelForCausalLM", "Qwen2ForCausalLM", "qwen2" ]
[ "text-generation" ]
[ "text" ]
[ "text" ]
[ "text" ]
67b79c8700245b72c5706777
google/gemma-3-4b-it
google
{ "models": [ { "_id": "67b79c6c01ad68cfed14677a", "id": "google/gemma-3-4b-pt" } ], "relation": "finetune" }
1,702,746
15,276,165
manual
2025-02-20T21:20:07Z
2025-03-21T20:20:53Z
transformers
1,265
22
null
image-text-to-text
null
[ ".gitattributes", "README.md", "added_tokens.json", "chat_template.json", "config.json", "generation_config.json", "model-00001-of-00002.safetensors", "model-00002-of-00002.safetensors", "model.safetensors.index.json", "preprocessor_config.json", "processor_config.json", "special_tokens_map.js...
093f9f388b31de276ce2de164bdc2081324b9767
[ "transformers", "safetensors", "gemma3", "image-text-to-text", "conversational", "arxiv:1905.07830", "arxiv:1905.10044", "arxiv:1911.11641", "arxiv:1904.09728", "arxiv:1705.03551", "arxiv:1911.01547", "arxiv:1907.10641", "arxiv:1903.00161", "arxiv:2009.03300", "arxiv:2304.06364", "arxi...
null
{"architectures": ["Gemma3ForConditionalGeneration"], "model_type": "gemma3", "processor_config": {"chat_template": "{{ bos_token }}\n{%- if messages[0]['role'] == 'system' -%}\n {%- if messages[0]['content'] is string -%}\n {%- set first_user_prefix = messages[0]['content'] + '\n\n' -%}\n {%- else -%}\n ...
{ "auto_model": "AutoModelForImageTextToText", "custom_class": null, "pipeline_tag": "image-text-to-text", "processor": "AutoProcessor" }
{"base_model": "google/gemma-3-4b-pt", "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "gemma", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "image-text-to-text", "tags": null, "extra_gated_heading": "Access Gemma o...
null
null
[ "gemma" ]
null
null
null
null
null
[ "AutoModelForImageTextToText", "Gemma3ForConditionalGeneration", "gemma3" ]
[ "image-text-to-text" ]
[ "multimodal" ]
[ "text", "image" ]
[ "text" ]
621ffdc136468d709f180294
sentence-transformers/all-MiniLM-L6-v2
sentence-transformers
null
206,073,068
2,406,656,367
False
2022-03-02T23:29:05Z
2025-03-06T13:37:44Z
sentence-transformers
4,620
21
null
sentence-similarity
{"parameters": {"I64": 512, "F32": 22713216}, "total": 22713728}
[ ".gitattributes", "1_Pooling/config.json", "README.md", "config.json", "config_sentence_transformers.json", "data_config.json", "model.safetensors", "modules.json", "onnx/model.onnx", "onnx/model_O1.onnx", "onnx/model_O2.onnx", "onnx/model_O3.onnx", "onnx/model_O4.onnx", "onnx/model_qint8_...
c9745ed1d9f207416be6d2e6f8de32d1f16199bf
[ "sentence-transformers", "pytorch", "tf", "rust", "onnx", "safetensors", "openvino", "bert", "feature-extraction", "sentence-similarity", "transformers", "en", "dataset:s2orc", "dataset:flax-sentence-embeddings/stackexchange_xml", "dataset:ms_marco", "dataset:gooaq", "dataset:yahoo_a...
null
{"architectures": ["BertModel"], "model_type": "bert", "tokenizer_config": {"unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]"}}
{ "auto_model": "AutoModel", "custom_class": null, "pipeline_tag": "feature-extraction", "processor": "AutoTokenizer" }
{"base_model": null, "datasets": ["s2orc", "flax-sentence-embeddings/stackexchange_xml", "ms_marco", "gooaq", "yahoo_answers_topics", "code_search_net", "search_qa", "eli5", "snli", "multi_nli", "wikihow", "natural_questions", "trivia_qa", "embedding-data/sentence-compression", "embedding-data/flickr30k-captions", "emb...
# all-MiniLM-L6-v2 This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 384 dimensional dense vector space and can be used for tasks like clustering or semantic search. ## Usage (Sentence-Transformers) Using this model becomes easy when you have [sentence-transformers](ht...
null
[ "apache-2.0" ]
[ "s2orc", "flax-sentence-embeddings/stackexchange_xml", "ms_marco", "gooaq", "yahoo_answers_topics", "code_search_net", "search_qa", "eli5", "snli", "multi_nli", "wikihow", "natural_questions", "trivia_qa", "embedding-data/sentence-compression", "embedding-data/flickr30k-captions", "emb...
[ "en" ]
22,713,728
null
null
[ "BertModel", "AutoModel", "bert" ]
[ "sentence-similarity", "feature-extraction" ]
[ "text", "multimodal" ]
[ "text" ]
[ "logits", "embeddings" ]
66944f1fe0c5c2e493a804f5
meta-llama/Llama-3.1-8B
meta-llama
null
1,387,690
20,616,230
manual
2024-07-14T22:20:15Z
2024-10-16T22:00:37Z
transformers
2,128
16
null
text-generation
{"parameters": {"BF16": 8030261248}, "total": 8030261248}
[ ".gitattributes", "LICENSE", "README.md", "USE_POLICY.md", "config.json", "generation_config.json", "model-00001-of-00004.safetensors", "model-00002-of-00004.safetensors", "model-00003-of-00004.safetensors", "model-00004-of-00004.safetensors", "model.safetensors.index.json", "original/consolid...
d04e592bb4f6aa9cfee91e2e20afa771667e1d4b
[ "transformers", "safetensors", "llama", "text-generation", "facebook", "meta", "pytorch", "llama-3", "en", "de", "fr", "it", "pt", "hi", "es", "th", "arxiv:2204.05149", "license:llama3.1", "text-generation-inference", "endpoints_compatible", "region:us" ]
null
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "eos_token": "<|end_of_text|>"}}
{ "auto_model": "AutoModelForCausalLM", "custom_class": null, "pipeline_tag": "text-generation", "processor": "AutoTokenizer" }
{"base_model": null, "datasets": null, "eval_results": null, "language": ["en", "de", "fr", "it", "pt", "hi", "es", "th"], "library_name": "transformers", "license": "llama3.1", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "text-generation", "tags": ["facebook", "meta...
null
null
[ "llama3.1" ]
null
[ "en", "de", "fr", "it", "pt", "hi", "es", "th" ]
8,030,261,248
null
null
[ "AutoModelForCausalLM", "llama", "LlamaForCausalLM" ]
[ "text-generation" ]
[ "text" ]
[ "text" ]
[ "text" ]
67fddfa9a7fe1f21ec1d3026
microsoft/bitnet-b1.58-2B-4T
microsoft
null
16,753
190,249
False
2025-04-15T04:25:13Z
2025-12-17T18:13:05Z
transformers
1,404
15
null
text-generation
{"parameters": {"BF16": 328775890, "U8": 521011200}, "total": 849787090}
[ ".gitattributes", "LICENSE", "README.md", "config.json", "data_summary_card.md", "generation_config.json", "model.safetensors", "special_tokens_map.json", "tokenizer.json", "tokenizer_config.json" ]
04c3b9ad9361b824064a1f25ea60a8be9599b127
[ "transformers", "safetensors", "bitnet", "text-generation", "chat", "large-language-model", "conversational", "custom_code", "en", "arxiv:2504.12285", "license:mit", "endpoints_compatible", "8-bit", "deploy:azure", "region:us" ]
null
{"architectures": ["BitNetForCausalLM"], "auto_map": {"AutoConfig": "configuration_bitnet.BitNetConfig", "AutoModelForCausalLM": "modeling_bitnet.BitNetForCausalLM"}, "model_type": "bitnet", "quantization_config": {"quant_method": "bitnet"}, "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% se...
{ "auto_model": "AutoModelForCausalLM", "custom_class": null, "pipeline_tag": "text-generation", "processor": "AutoTokenizer" }
{"base_model": null, "datasets": null, "eval_results": null, "language": ["en"], "library_name": "transformers", "license": "mit", "license_name": null, "license_link": "https://huggingface.co/microsoft/bitnet-b1.58-2B-4T/blob/main/LICENSE", "metrics": null, "model_name": null, "pipeline_tag": "text-generation", "tags"...
# BitNet b1.58 2B4T - Scaling Native 1-bit LLM This repository contains the weights for **BitNet b1.58 2B4T**, the first open-source, native 1-bit Large Language Model (LLM) at the 2-billion parameter scale, developed by Microsoft Research. Trained on a corpus of 4 trillion tokens, this model demonstrates that native...
null
[ "mit", "https://huggingface.co/microsoft/bitnet-b1.58-2B-4T/blob/main/LICENSE" ]
null
[ "en" ]
849,787,090
null
null
[ "AutoModelForCausalLM", "BitNetForCausalLM", "bitnet" ]
[ "text-generation" ]
[ "text" ]
[ "text" ]
[ "text" ]
680da79d48c19d19cf82ab5e
Qwen/Qwen3-8B
Qwen
{ "models": [ { "_id": "680f0c3e6eacf22a04ee1219", "id": "Qwen/Qwen3-8B-Base" } ], "relation": "finetune" }
9,481,229
43,490,001
False
2025-04-27T03:42:21Z
2025-07-26T03:49:13Z
transformers
1,013
15
null
text-generation
{"parameters": {"BF16": 8190735360}, "total": 8190735360}
[ ".gitattributes", "LICENSE", "README.md", "config.json", "generation_config.json", "merges.txt", "model-00001-of-00005.safetensors", "model-00002-of-00005.safetensors", "model-00003-of-00005.safetensors", "model-00004-of-00005.safetensors", "model-00005-of-00005.safetensors", "model.safetensor...
b968826d9c46dd6066d109eabc6255188de91218
[ "transformers", "safetensors", "qwen3", "text-generation", "conversational", "arxiv:2309.00071", "arxiv:2505.09388", "base_model:Qwen/Qwen3-8B-Base", "base_model:finetune:Qwen/Qwen3-8B-Base", "license:apache-2.0", "text-generation-inference", "endpoints_compatible", "deploy:azure", "region...
null
{"architectures": ["Qwen3ForCausalLM"], "model_type": "qwen3", "tokenizer_config": {"bos_token": null, "chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0].role == 'system' %}\n {{- messages[0].content + '\\n\\n' }}\n {%- endif %}\n {{- \"# Tools\\n\\nYou may call o...
{ "auto_model": "AutoModelForCausalLM", "custom_class": null, "pipeline_tag": "text-generation", "processor": "AutoTokenizer" }
{"base_model": ["Qwen/Qwen3-8B-Base"], "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "apache-2.0", "license_name": null, "license_link": "https://huggingface.co/Qwen/Qwen3-8B/blob/main/LICENSE", "metrics": null, "model_name": null, "pipeline_tag": "text-generation"...
null
null
[ "apache-2.0", "https://huggingface.co/Qwen/Qwen3-8B/blob/main/LICENSE" ]
null
null
8,190,735,360
null
null
[ "AutoModelForCausalLM", "Qwen3ForCausalLM", "qwen3" ]
[ "text-generation" ]
[ "text" ]
[ "text" ]
[ "text" ]
683f05dc98de733bf4c59f3d
Qwen/Qwen3-Embedding-0.6B
Qwen
{ "models": [ { "_id": "680f0da2f99a20ddfc9a9f26", "id": "Qwen/Qwen3-0.6B-Base" } ], "relation": "finetune" }
5,507,791
36,810,259
False
2025-06-03T14:25:32Z
2025-06-20T09:31:05Z
sentence-transformers
950
15
null
feature-extraction
{"parameters": {"BF16": 595776512}, "total": 595776512}
[ ".gitattributes", "1_Pooling/config.json", "README.md", "config.json", "config_sentence_transformers.json", "generation_config.json", "merges.txt", "model.safetensors", "modules.json", "tokenizer.json", "tokenizer_config.json", "vocab.json" ]
c54f2e6e80b2d7b7de06f51cec4959f6b3e03418
[ "sentence-transformers", "safetensors", "qwen3", "text-generation", "transformers", "sentence-similarity", "feature-extraction", "text-embeddings-inference", "arxiv:2506.05176", "base_model:Qwen/Qwen3-0.6B-Base", "base_model:finetune:Qwen/Qwen3-0.6B-Base", "license:apache-2.0", "endpoints_co...
null
{"architectures": ["Qwen3ForCausalLM"], "model_type": "qwen3", "tokenizer_config": {"bos_token": null, "chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0].role == 'system' %}\n {{- messages[0].content + '\\n\\n' }}\n {%- endif %}\n {{- \"# Tools\\n\\nYou may call o...
{ "auto_model": "AutoModelForCausalLM", "custom_class": null, "pipeline_tag": "text-generation", "processor": "AutoTokenizer" }
{"base_model": ["Qwen/Qwen3-0.6B-Base"], "datasets": null, "eval_results": null, "language": null, "library_name": null, "license": "apache-2.0", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": null, "tags": ["transformers", "sentence-transformers", "sentence-similarity"...
# Qwen3-Embedding-0.6B <p align="center"> <img src="https://qianwen-res.oss-accelerate-overseas.aliyuncs.com/logo_qwen3.png" width="400"/> <p> ## Highlights The Qwen3 Embedding model series is the latest proprietary model of the Qwen family, specifically designed for text embedding and ranking tasks. Building up...
null
[ "apache-2.0" ]
null
null
595,776,512
null
null
[ "AutoModelForCausalLM", "Qwen3ForCausalLM", "qwen3" ]
[ "sentence-similarity", "feature-extraction", "text-generation" ]
[ "text", "multimodal" ]
[ "text" ]
[ "logits", "text", "embeddings" ]
68913539bd3d0a833438591d
openai/gpt-oss-20b
openai
null
6,734,875
54,662,871
False
2025-08-04T22:33:29Z
2025-08-26T17:25:47Z
transformers
4,485
15
null
text-generation
{"parameters": {"BF16": 1804459584, "U8": 19707494400}, "total": 21511953984}
[ ".gitattributes", "LICENSE", "README.md", "USAGE_POLICY", "chat_template.jinja", "config.json", "generation_config.json", "metal/model.bin", "model-00000-of-00002.safetensors", "model-00001-of-00002.safetensors", "model-00002-of-00002.safetensors", "model.safetensors.index.json", "original/c...
6cee5e81ee83917806bbde320786a8fb61efebee
[ "transformers", "safetensors", "gpt_oss", "text-generation", "vllm", "conversational", "arxiv:2508.10925", "license:apache-2.0", "eval-results", "endpoints_compatible", "8-bit", "mxfp4", "deploy:azure", "region:us" ]
null
{"architectures": ["GptOssForCausalLM"], "model_type": "gpt_oss", "quantization_config": {"quant_method": "mxfp4"}, "tokenizer_config": {"bos_token": "<|startoftext|>", "eos_token": "<|return|>", "pad_token": "<|endoftext|>"}, "chat_template_jinja": "{#-\n In addition to the normal inputs of `messages` and `tools`, th...
{ "auto_model": "AutoModelForCausalLM", "custom_class": null, "pipeline_tag": "text-generation", "processor": "AutoTokenizer" }
{"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "apache-2.0", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "text-generation", "tags": ["vllm"]}
<p align="center"> <img alt="gpt-oss-20b" src="https://raw.githubusercontent.com/openai/gpt-oss/main/docs/gpt-oss-20b.svg"> </p> <p align="center"> <a href="https://gpt-oss.com"><strong>Try gpt-oss</strong></a> · <a href="https://cookbook.openai.com/topic/gpt-oss"><strong>Guides</strong></a> · <a href="https:/...
null
[ "apache-2.0" ]
null
null
21,511,953,984
null
null
[ "AutoModelForCausalLM", "GptOssForCausalLM", "gpt_oss" ]
[ "text-generation" ]
[ "text" ]
[ "text" ]
[ "text" ]
680da718233834890aa01f51
Qwen/Qwen3-0.6B
Qwen
{ "models": [ { "_id": "680f0da2f99a20ddfc9a9f26", "id": "Qwen/Qwen3-0.6B-Base" } ], "relation": "finetune" }
13,807,272
74,174,117
False
2025-04-27T03:40:08Z
2025-07-26T03:46:27Z
transformers
1,160
13
null
text-generation
{"parameters": {"BF16": 751632384}, "total": 751632384}
[ ".gitattributes", "LICENSE", "README.md", "config.json", "generation_config.json", "merges.txt", "model.safetensors", "tokenizer.json", "tokenizer_config.json", "vocab.json" ]
c1899de289a04d12100db370d81485cdf75e47ca
[ "transformers", "safetensors", "qwen3", "text-generation", "conversational", "arxiv:2505.09388", "base_model:Qwen/Qwen3-0.6B-Base", "base_model:finetune:Qwen/Qwen3-0.6B-Base", "license:apache-2.0", "text-generation-inference", "endpoints_compatible", "deploy:azure", "region:us" ]
null
{"architectures": ["Qwen3ForCausalLM"], "model_type": "qwen3", "tokenizer_config": {"bos_token": null, "chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0].role == 'system' %}\n {{- messages[0].content + '\\n\\n' }}\n {%- endif %}\n {{- \"# Tools\\n\\nYou may call o...
{ "auto_model": "AutoModelForCausalLM", "custom_class": null, "pipeline_tag": "text-generation", "processor": "AutoTokenizer" }
{"base_model": ["Qwen/Qwen3-0.6B-Base"], "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "apache-2.0", "license_name": null, "license_link": "https://huggingface.co/Qwen/Qwen3-0.6B/blob/main/LICENSE", "metrics": null, "model_name": null, "pipeline_tag": "text-generat...
null
null
[ "apache-2.0", "https://huggingface.co/Qwen/Qwen3-0.6B/blob/main/LICENSE" ]
null
null
751,632,384
null
null
[ "AutoModelForCausalLM", "Qwen3ForCausalLM", "qwen3" ]
[ "text-generation" ]
[ "text" ]
[ "text" ]
[ "text" ]
678dc6fff905d106be796d8a
deepseek-ai/DeepSeek-R1
deepseek-ai
null
2,203,466
17,068,124
False
2025-01-20T03:46:07Z
2025-03-27T04:01:59Z
transformers
13,107
12
null
text-generation
{"parameters": {"BF16": 3918786560, "F8_E4M3": 680571043840, "F32": 41555600}, "total": 684531386000}
[ ".gitattributes", "LICENSE", "README.md", "config.json", "configuration_deepseek.py", "figures/benchmark.jpg", "generation_config.json", "model-00001-of-000163.safetensors", "model-00002-of-000163.safetensors", "model-00003-of-000163.safetensors", "model-00004-of-000163.safetensors", "model-00...
56d4cbbb4d29f4355bab4b9a39ccb717a14ad5ad
[ "transformers", "safetensors", "deepseek_v3", "text-generation", "conversational", "custom_code", "arxiv:2501.12948", "license:mit", "eval-results", "text-generation-inference", "endpoints_compatible", "fp8", "region:us" ]
null
{"architectures": ["DeepseekV3ForCausalLM"], "auto_map": {"AutoConfig": "configuration_deepseek.DeepseekV3Config", "AutoModel": "modeling_deepseek.DeepseekV3Model", "AutoModelForCausalLM": "modeling_deepseek.DeepseekV3ForCausalLM"}, "model_type": "deepseek_v3", "quantization_config": {"quant_method": "fp8"}, "tokenizer...
{ "auto_model": "AutoModelForCausalLM", "custom_class": null, "pipeline_tag": "text-generation", "processor": "AutoTokenizer" }
{"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "mit", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": null, "tags": null}
# DeepSeek-R1 <!-- markdownlint-disable first-line-h1 --> <!-- markdownlint-disable html --> <!-- markdownlint-disable no-duplicate-header --> <div align="center"> <img src="https://github.com/deepseek-ai/DeepSeek-V2/blob/main/figures/logo.svg?raw=true" width="60%" alt="DeepSeek-V3" /> </div> <hr> <div align="center...
null
[ "mit" ]
null
null
684,531,386,000
null
null
[ "AutoModelForCausalLM", "deepseek_v3", "DeepseekV3ForCausalLM" ]
[ "text-generation" ]
[ "text" ]
[ "text" ]
[ "text" ]
6891e3bb084ce75acffb033d
Qwen/Qwen3-4B-Instruct-2507
Qwen
null
6,235,673
28,945,408
False
2025-08-05T10:58:03Z
2025-09-17T06:56:53Z
transformers
785
12
null
text-generation
{"parameters": {"BF16": 4022468096}, "total": 4022468096}
[ ".gitattributes", "LICENSE", "README.md", "config.json", "generation_config.json", "merges.txt", "model-00001-of-00003.safetensors", "model-00002-of-00003.safetensors", "model-00003-of-00003.safetensors", "model.safetensors.index.json", "tokenizer.json", "tokenizer_config.json", "vocab.json"...
cdbee75f17c01a7cc42f958dc650907174af0554
[ "transformers", "safetensors", "qwen3", "text-generation", "conversational", "arxiv:2505.09388", "license:apache-2.0", "eval-results", "text-generation-inference", "endpoints_compatible", "deploy:azure", "region:us" ]
null
{"architectures": ["Qwen3ForCausalLM"], "model_type": "qwen3", "tokenizer_config": {"bos_token": null, "chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0].role == 'system' %}\n {{- messages[0].content + '\\n\\n' }}\n {%- endif %}\n {{- \"# Tools\\n\\nYou may call o...
{ "auto_model": "AutoModelForCausalLM", "custom_class": null, "pipeline_tag": "text-generation", "processor": "AutoTokenizer" }
{"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "apache-2.0", "license_name": null, "license_link": "https://huggingface.co/Qwen/Qwen3-4B-Instruct-2507/blob/main/LICENSE", "metrics": null, "model_name": null, "pipeline_tag": "text-generation", "t...
# Qwen3-4B-Instruct-2507 <a href="https://chat.qwen.ai" target="_blank" style="margin: 2px;"> <img alt="Chat" src="https://img.shields.io/badge/%F0%9F%92%9C%EF%B8%8F%20Qwen%20Chat%20-536af5" style="display: inline-block; vertical-align: middle;"/> </a> ## Highlights We introduce the updated version of the **Qwen3...
null
[ "apache-2.0", "https://huggingface.co/Qwen/Qwen3-4B-Instruct-2507/blob/main/LICENSE" ]
null
null
4,022,468,096
null
null
[ "AutoModelForCausalLM", "Qwen3ForCausalLM", "qwen3" ]
[ "text-generation" ]
[ "text" ]
[ "text" ]
[ "text" ]
68ea05fb43df37d95ad2491d
Qwen/Qwen3-VL-8B-Instruct
Qwen
null
4,649,140
20,365,833
False
2025-10-11T07:23:39Z
2025-10-15T16:16:59Z
transformers
837
12
null
image-text-to-text
{"parameters": {"BF16": 8767123696}, "total": 8767123696}
[ ".gitattributes", "README.md", "chat_template.json", "config.json", "generation_config.json", "merges.txt", "model-00001-of-00004.safetensors", "model-00002-of-00004.safetensors", "model-00003-of-00004.safetensors", "model-00004-of-00004.safetensors", "model.safetensors.index.json", "preproces...
0c351dd01ed87e9c1b53cbc748cba10e6187ff3b
[ "transformers", "safetensors", "qwen3_vl", "image-text-to-text", "conversational", "arxiv:2505.09388", "arxiv:2502.13923", "arxiv:2409.12191", "arxiv:2308.12966", "license:apache-2.0", "endpoints_compatible", "deploy:azure", "region:us" ]
null
{"architectures": ["Qwen3VLForConditionalGeneration"], "model_type": "qwen3_vl", "processor_config": {"chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0].role == 'system' %}\n {%- if messages[0].content is string %}\n {{- messages[0].content }}\n {%- el...
{ "auto_model": "AutoModelForImageTextToText", "custom_class": null, "pipeline_tag": "image-text-to-text", "processor": "AutoProcessor" }
{"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "apache-2.0", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "image-text-to-text", "tags": null}
null
null
[ "apache-2.0" ]
null
null
8,767,123,696
null
null
[ "AutoModelForImageTextToText", "qwen3_vl", "Qwen3VLForConditionalGeneration" ]
[ "image-text-to-text" ]
[ "multimodal" ]
[ "text", "image" ]
[ "text" ]
680da769038897aa905586a7
Qwen/Qwen3-4B
Qwen
{ "models": [ { "_id": "680f0c5b4e1c1da6491f3b9d", "id": "Qwen/Qwen3-4B-Base" } ], "relation": "finetune" }
6,491,995
30,446,215
False
2025-04-27T03:41:29Z
2025-07-26T03:46:39Z
transformers
582
11
null
text-generation
null
[ ".gitattributes", "LICENSE", "README.md", "config.json", "generation_config.json", "merges.txt", "model-00001-of-00003.safetensors", "model-00002-of-00003.safetensors", "model-00003-of-00003.safetensors", "model.safetensors.index.json", "tokenizer.json", "tokenizer_config.json", "vocab.json"...
1cfa9a7208912126459214e8b04321603b3df60c
[ "transformers", "safetensors", "qwen3", "text-generation", "conversational", "arxiv:2309.00071", "arxiv:2505.09388", "base_model:Qwen/Qwen3-4B-Base", "base_model:finetune:Qwen/Qwen3-4B-Base", "license:apache-2.0", "text-generation-inference", "endpoints_compatible", "deploy:azure", "region...
null
{"architectures": ["Qwen3ForCausalLM"], "model_type": "qwen3", "tokenizer_config": {"bos_token": null, "chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0].role == 'system' %}\n {{- messages[0].content + '\\n\\n' }}\n {%- endif %}\n {{- \"# Tools\\n\\nYou may call o...
{ "auto_model": "AutoModelForCausalLM", "custom_class": null, "pipeline_tag": "text-generation", "processor": "AutoTokenizer" }
{"base_model": ["Qwen/Qwen3-4B-Base"], "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "apache-2.0", "license_name": null, "license_link": "https://huggingface.co/Qwen/Qwen3-4B/blob/main/LICENSE", "metrics": null, "model_name": null, "pipeline_tag": "text-generation"...
null
null
[ "apache-2.0", "https://huggingface.co/Qwen/Qwen3-4B/blob/main/LICENSE" ]
null
null
null
null
null
[ "AutoModelForCausalLM", "Qwen3ForCausalLM", "qwen3" ]
[ "text-generation" ]
[ "text" ]
[ "text" ]
[ "text" ]
683f090e77b41ac9fc98c51f
Qwen/Qwen3-Embedding-8B
Qwen
{ "models": [ { "_id": "680f0c3e6eacf22a04ee1219", "id": "Qwen/Qwen3-8B-Base" } ], "relation": "finetune" }
1,648,122
9,165,432
False
2025-06-03T14:39:10Z
2025-07-07T09:02:21Z
sentence-transformers
628
11
null
feature-extraction
{"parameters": {"BF16": 7567295488}, "total": 7567295488}
[ ".gitattributes", "1_Pooling/config.json", "LICENSE", "README.md", "config.json", "config_sentence_transformers.json", "generation_config.json", "merges.txt", "model-00001-of-00004.safetensors", "model-00002-of-00004.safetensors", "model-00003-of-00004.safetensors", "model-00004-of-00004.safet...
1d8ad4ca9b3dd8059ad90a75d4983776a23d44af
[ "sentence-transformers", "safetensors", "qwen3", "text-generation", "transformers", "sentence-similarity", "feature-extraction", "text-embeddings-inference", "arxiv:2506.05176", "base_model:Qwen/Qwen3-8B-Base", "base_model:finetune:Qwen/Qwen3-8B-Base", "license:apache-2.0", "endpoints_compat...
null
{"architectures": ["Qwen3ForCausalLM"], "model_type": "qwen3", "tokenizer_config": {"bos_token": null, "chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0]['role'] == 'system' %}\n {{- messages[0]['content'] }}\n {%- else %}\n {{- 'You are a helpful assistant.' ...
{ "auto_model": "AutoModelForCausalLM", "custom_class": null, "pipeline_tag": "text-generation", "processor": "AutoTokenizer" }
{"base_model": ["Qwen/Qwen3-8B-Base"], "datasets": null, "eval_results": null, "language": null, "library_name": null, "license": "apache-2.0", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": null, "tags": ["transformers", "sentence-transformers", "sentence-similarity", ...
null
null
[ "apache-2.0" ]
null
null
7,567,295,488
null
null
[ "AutoModelForCausalLM", "Qwen3ForCausalLM", "qwen3" ]
[ "sentence-similarity", "feature-extraction", "text-generation" ]
[ "text", "multimodal" ]
[ "text" ]
[ "logits", "text", "embeddings" ]
684a65ac33eb1c9a8fd8877c
dphn/Dolphin-Mistral-24B-Venice-Edition
dphn
{ "models": [ { "_id": "6798dbe54a10be7109f954f2", "id": "mistralai/Mistral-Small-24B-Instruct-2501" } ], "relation": "finetune" }
7,197
97,989
False
2025-06-12T05:29:16Z
2025-09-08T05:57:39Z
transformers
461
10
null
text-generation
{"parameters": {"BF16": 23572403200}, "total": 23572403200}
[ ".gitattributes", "README.md", "config.json", "generation_config.json", "model-00001-of-00010.safetensors", "model-00002-of-00010.safetensors", "model-00003-of-00010.safetensors", "model-00004-of-00010.safetensors", "model-00005-of-00010.safetensors", "model-00006-of-00010.safetensors", "model-0...
853f87026d587889137c814bb56bba60f033e747
[ "transformers", "safetensors", "mistral", "text-generation", "conversational", "base_model:mistralai/Mistral-Small-24B-Instruct-2501", "base_model:finetune:mistralai/Mistral-Small-24B-Instruct-2501", "license:apache-2.0", "text-generation-inference", "endpoints_compatible", "deploy:azure", "re...
null
{"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{%- set today = strftime_now(\"%Y-%m-%d\") %}\n{%- set default_system_message = \"You are Dolphin Mistral 24B Venice Edition. You are an uncensored AI model. You follow instructions without hesi...
{ "auto_model": "AutoModelForCausalLM", "custom_class": null, "pipeline_tag": "text-generation", "processor": "AutoTokenizer" }
{"base_model": ["mistralai/Mistral-Small-24B-Instruct-2501"], "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "apache-2.0", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "text-generation", "tags": null}
# 🐬 Dolphin Mistral 24B Venice Edition 🌅 Website: https://dphn.ai Twitter: https://x.com/dphnAI Web Chat: https://chat.dphn.ai Telegram bot: https://t.me/DolphinAI_bot ![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/68485b28c949339ca04c370c/LMOLMYwK-ixnGGdSBXew6.jpeg) ## What is Dolphin...
null
[ "apache-2.0" ]
null
null
23,572,403,200
null
null
[ "AutoModelForCausalLM", "mistral", "MistralForCausalLM" ]
[ "text-generation" ]
[ "text" ]
[ "text" ]
[ "text" ]
66e98d74d4219fc3f1db215f
Qwen/Qwen2.5-3B-Instruct
Qwen
{ "models": [ { "_id": "66e6d03f69d1083dd982a18c", "id": "Qwen/Qwen2.5-3B" } ], "relation": "finetune" }
7,604,571
70,809,504
False
2024-09-17T14:08:52Z
2024-09-25T12:33:00Z
transformers
426
9
null
text-generation
{"parameters": {"BF16": 3085938688}, "total": 3085938688}
[ ".gitattributes", "LICENSE", "README.md", "config.json", "generation_config.json", "merges.txt", "model-00001-of-00002.safetensors", "model-00002-of-00002.safetensors", "model.safetensors.index.json", "tokenizer.json", "tokenizer_config.json", "vocab.json" ]
aa8e72537993ba99e69dfaafa59ed015b17504d1
[ "transformers", "safetensors", "qwen2", "text-generation", "chat", "conversational", "en", "arxiv:2407.10671", "base_model:Qwen/Qwen2.5-3B", "base_model:finetune:Qwen/Qwen2.5-3B", "license:other", "text-generation-inference", "endpoints_compatible", "deploy:azure", "region:us" ]
null
{"architectures": ["Qwen2ForCausalLM"], "model_type": "qwen2", "tokenizer_config": {"bos_token": null, "chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0]['role'] == 'system' %}\n {{- messages[0]['content'] }}\n {%- else %}\n {{- 'You are Qwen, created by Aliba...
{ "auto_model": "AutoModelForCausalLM", "custom_class": null, "pipeline_tag": "text-generation", "processor": "AutoTokenizer" }
{"base_model": "Qwen/Qwen2.5-3B", "datasets": null, "eval_results": null, "language": ["en"], "library_name": "transformers", "license": "other", "license_name": "qwen-research", "license_link": "https://huggingface.co/Qwen/Qwen2.5-3B-Instruct/blob/main/LICENSE", "metrics": null, "model_name": null, "pipeline_tag": "te...
# Qwen2.5-3B-Instruct ## Introduction Qwen2.5 is the latest series of Qwen large language models. For Qwen2.5, we release a number of base language models and instruction-tuned language models ranging from 0.5 to 72 billion parameters. Qwen2.5 brings the following improvements upon Qwen2: - Significantly **more know...
null
[ "other", "qwen-research", "https://huggingface.co/Qwen/Qwen2.5-3B-Instruct/blob/main/LICENSE" ]
null
[ "en" ]
3,085,938,688
null
null
[ "AutoModelForCausalLM", "Qwen2ForCausalLM", "qwen2" ]
[ "text-generation" ]
[ "text" ]
[ "text" ]
[ "text" ]
67c35b9bb236f0d365bf29d3
google/gemma-3-27b-it
google
{ "models": [ { "_id": "67c35b8b87a7f49a82593992", "id": "google/gemma-3-27b-pt" } ], "relation": "finetune" }
1,078,578
12,381,516
manual
2025-03-01T19:10:19Z
2025-03-21T20:29:02Z
transformers
1,935
9
null
image-text-to-text
{"parameters": {"BF16": 27432406640}, "total": 27432406640}
[ ".gitattributes", "README.md", "added_tokens.json", "chat_template.json", "config.json", "generation_config.json", "model-00001-of-00012.safetensors", "model-00002-of-00012.safetensors", "model-00003-of-00012.safetensors", "model-00004-of-00012.safetensors", "model-00005-of-00012.safetensors", ...
005ad3404e59d6023443cb575daa05336842228a
[ "transformers", "safetensors", "gemma3", "image-text-to-text", "conversational", "arxiv:1905.07830", "arxiv:1905.10044", "arxiv:1911.11641", "arxiv:1904.09728", "arxiv:1705.03551", "arxiv:1911.01547", "arxiv:1907.10641", "arxiv:1903.00161", "arxiv:2009.03300", "arxiv:2304.06364", "arxi...
null
{"architectures": ["Gemma3ForConditionalGeneration"], "model_type": "gemma3", "processor_config": {"chat_template": "{{ bos_token }}\n{%- if messages[0]['role'] == 'system' -%}\n {%- if messages[0]['content'] is string -%}\n {%- set first_user_prefix = messages[0]['content'] + '\n\n' -%}\n {%- else -%}\n ...
{ "auto_model": "AutoModelForImageTextToText", "custom_class": null, "pipeline_tag": "image-text-to-text", "processor": "AutoProcessor" }
{"base_model": "google/gemma-3-27b-pt", "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "gemma", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "image-text-to-text", "tags": null, "extra_gated_heading": "Access Gemma ...
null
null
[ "gemma" ]
null
null
27,432,406,640
null
null
[ "AutoModelForImageTextToText", "Gemma3ForConditionalGeneration", "gemma3" ]
[ "image-text-to-text" ]
[ "multimodal" ]
[ "text", "image" ]
[ "text" ]
690088e0369d9d3b7594d73c
OpenMOSS-Team/RoboOmni
OpenMOSS-Team
null
148
378
False
2025-10-28T09:12:00Z
2025-10-30T06:53:06Z
transformers
15
9
null
robotics
null
[ ".gitattributes", "README.md", "added_tokens.json", "chat_template.jinja", "config.json", "generation_config.json", "merges.txt", "model.safetensors", "model.safetensors.index.json", "preprocessor_config.json", "special_tokens_map.json", "tokenizer.json", "tokenizer_config.json", "vocab.js...
3a84313809c3b16fad1295a2e3814754bf179cfe
[ "transformers", "safetensors", "qwen2_5_omni_thinker", "text-generation", "robotics", "en", "dataset:fnlp/OmniAction", "arxiv:2510.23763", "license:cc-by-nc-4.0", "endpoints_compatible", "region:us" ]
null
{"architectures": ["Qwen2_5OmniThinkerForConditionalGeneration"], "model_type": "qwen2_5_omni_thinker", "tokenizer_config": {"bos_token": null, "eos_token": "<|im_end|>", "pad_token": "<|endoftext|>", "unk_token": null}, "chat_template_jinja": "{% set audio_count = namespace(value=0) %}{% set image_count = namespace(va...
{ "auto_model": "AutoModelForSeq2SeqLM", "custom_class": null, "pipeline_tag": "text-generation", "processor": null }
{"base_model": null, "datasets": ["fnlp/OmniAction"], "eval_results": null, "language": ["en"], "library_name": "transformers", "license": "cc-by-nc-4.0", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "robotics", "tags": null}
null
null
[ "cc-by-nc-4.0" ]
[ "fnlp/OmniAction" ]
[ "en" ]
null
null
null
[ "AutoModelForSeq2SeqLM", "qwen2_5_omni_thinker", "Qwen2_5OmniThinkerForConditionalGeneration" ]
[ "robotics", "text-generation" ]
[ "text" ]
[ "text" ]
[ "text" ]
621ffdc036468d709f174338
google-bert/bert-base-uncased
google-bert
null
71,392,969
2,805,772,843
False
2022-03-02T23:29:04Z
2024-02-19T11:06:12Z
transformers
2,595
8
null
fill-mask
{"parameters": {"F32": 110106428}, "total": 110106428}
[ ".gitattributes", "LICENSE", "README.md", "config.json", "coreml/fill-mask/float32_model.mlpackage/Data/com.apple.CoreML/model.mlmodel", "coreml/fill-mask/float32_model.mlpackage/Data/com.apple.CoreML/weights/weight.bin", "coreml/fill-mask/float32_model.mlpackage/Manifest.json", "flax_model.msgpack", ...
86b5e0934494bd15c9632b12f734a8a67f723594
[ "transformers", "pytorch", "tf", "jax", "rust", "coreml", "onnx", "safetensors", "bert", "fill-mask", "exbert", "en", "dataset:bookcorpus", "dataset:wikipedia", "arxiv:1810.04805", "license:apache-2.0", "endpoints_compatible", "deploy:azure", "region:us" ]
null
{"architectures": ["BertForMaskedLM"], "model_type": "bert", "tokenizer_config": {}}
{ "auto_model": "AutoModelForMaskedLM", "custom_class": null, "pipeline_tag": "fill-mask", "processor": "AutoTokenizer" }
{"base_model": null, "datasets": ["bookcorpus", "wikipedia"], "eval_results": null, "language": "en", "library_name": null, "license": "apache-2.0", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": null, "tags": ["exbert"]}
# BERT base model (uncased) Pretrained model on English language using a masked language modeling (MLM) objective. It was introduced in [this paper](https://arxiv.org/abs/1810.04805) and first released in [this repository](https://github.com/google-research/bert). This model is uncased: it does not make a difference b...
null
[ "apache-2.0" ]
[ "bookcorpus", "wikipedia" ]
[ "en" ]
110,106,428
null
null
[ "AutoModelForMaskedLM", "bert", "BertForMaskedLM" ]
[ "fill-mask" ]
[ "text" ]
[ "text" ]
[ "logits" ]
66fba7309482f97131bf08d6
openai/whisper-large-v3-turbo
openai
{ "models": [ { "_id": "654a84cadff2f49007ce6c37", "id": "openai/whisper-large-v3" } ], "relation": "finetune" }
5,038,910
72,108,416
False
2024-10-01T07:39:28Z
2024-10-04T14:51:11Z
transformers
2,871
8
null
automatic-speech-recognition
null
[ ".gitattributes", "README.md", "added_tokens.json", "config.json", "generation_config.json", "merges.txt", "model.safetensors", "normalizer.json", "preprocessor_config.json", "special_tokens_map.json", "tokenizer.json", "tokenizer_config.json", "vocab.json" ]
41f01f3fe87f28c78e2fbf8b568835947dd65ed9
[ "transformers", "safetensors", "whisper", "automatic-speech-recognition", "audio", "en", "zh", "de", "es", "ru", "ko", "fr", "ja", "pt", "tr", "pl", "ca", "nl", "ar", "sv", "it", "id", "hi", "fi", "vi", "he", "uk", "el", "ms", "cs", "ro", "da", "hu", ...
null
{"architectures": ["WhisperForConditionalGeneration"], "model_type": "whisper", "tokenizer_config": {"bos_token": "<|endoftext|>", "eos_token": "<|endoftext|>", "pad_token": "<|endoftext|>", "unk_token": "<|endoftext|>"}}
{ "auto_model": "AutoModelForSpeechSeq2Seq", "custom_class": null, "pipeline_tag": "automatic-speech-recognition", "processor": "AutoProcessor" }
{"base_model": ["openai/whisper-large-v3"], "datasets": null, "eval_results": null, "language": ["en", "zh", "de", "es", "ru", "ko", "fr", "ja", "pt", "tr", "pl", "ca", "nl", "ar", "sv", "it", "id", "hi", "fi", "vi", "he", "uk", "el", "ms", "cs", "ro", "da", "hu", "ta", "no", "th", "ur", "hr", "bg", "lt", "la", "mi", "...
# Whisper Whisper is a state-of-the-art model for automatic speech recognition (ASR) and speech translation, proposed in the paper [Robust Speech Recognition via Large-Scale Weak Supervision](https://huggingface.co/papers/2212.04356) by Alec Radford et al. from OpenAI. Trained on >5M hours of labeled data, Whisper d...
null
[ "mit" ]
null
[ "en", "zh", "de", "es", "ru", "ko", "fr", "ja", "pt", "tr", "pl", "ca", "nl", "ar", "sv", "it", "id", "hi", "fi", "vi", "he", "uk", "el", "ms", "cs", "ro", "da", "hu", "ta", "no", "th", "ur", "hr", "bg", "lt", "la", "mi", "ml", "cy", "sk"...
null
null
null
[ "WhisperForConditionalGeneration", "whisper", "AutoModelForSpeechSeq2Seq" ]
[ "automatic-speech-recognition" ]
[ "multimodal" ]
[ "audio" ]
[ "text" ]
6795ffcd88cd7c0294702a72
Qwen/Qwen2.5-VL-7B-Instruct
Qwen
null
4,611,358
51,675,220
False
2025-01-26T09:26:37Z
2025-04-06T16:23:01Z
transformers
1,476
8
null
image-text-to-text
{"parameters": {"BF16": 8292166656}, "total": 8292166656}
[ ".gitattributes", "README.md", "chat_template.json", "config.json", "generation_config.json", "merges.txt", "model-00001-of-00005.safetensors", "model-00002-of-00005.safetensors", "model-00003-of-00005.safetensors", "model-00004-of-00005.safetensors", "model-00005-of-00005.safetensors", "model...
cc594898137f460bfe9f0759e9844b3ce807cfb5
[ "transformers", "safetensors", "qwen2_5_vl", "image-text-to-text", "multimodal", "conversational", "en", "arxiv:2309.00071", "arxiv:2409.12191", "arxiv:2308.12966", "license:apache-2.0", "eval-results", "text-generation-inference", "endpoints_compatible", "deploy:azure", "region:us" ]
null
{"architectures": ["Qwen2_5_VLForConditionalGeneration"], "model_type": "qwen2_5_vl", "processor_config": {"chat_template": "{% set image_count = namespace(value=0) %}{% set video_count = namespace(value=0) %}{% for message in messages %}{% if loop.first and message['role'] != 'system' %}<|im_start|>system\nYou are a h...
{ "auto_model": "AutoModelForImageTextToText", "custom_class": null, "pipeline_tag": "image-text-to-text", "processor": "AutoProcessor" }
{"base_model": null, "datasets": null, "eval_results": null, "language": ["en"], "library_name": "transformers", "license": "apache-2.0", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "image-text-to-text", "tags": ["multimodal"]}
# Qwen2.5-VL-7B-Instruct <a href="https://chat.qwenlm.ai/" target="_blank" style="margin: 2px;"> <img alt="Chat" src="https://img.shields.io/badge/%F0%9F%92%9C%EF%B8%8F%20Qwen%20Chat%20-536af5" style="display: inline-block; vertical-align: middle;"/> </a> ## Introduction In the past five months since Qwen2-VL’s r...
null
[ "apache-2.0" ]
null
[ "en" ]
8,292,166,656
null
null
[ "qwen2_5_vl", "AutoModelForImageTextToText", "Qwen2_5_VLForConditionalGeneration" ]
[ "image-text-to-text" ]
[ "multimodal" ]
[ "text", "image" ]
[ "text" ]
65f44de295b7d70871c368eb
BAAI/bge-reranker-v2-m3
BAAI
null
5,692,258
47,134,293
False
2024-03-15T13:32:18Z
2024-06-24T14:08:45Z
sentence-transformers
934
7
null
text-classification
{"parameters": {"F32": 567755777}, "total": 567755777}
[ ".gitattributes", "README.md", "assets/BEIR-bge-en-v1.5.png", "assets/BEIR-e5-mistral.png", "assets/CMTEB-retrieval-bge-zh-v1.5.png", "assets/llama-index.png", "assets/miracl-bge-m3.png", "config.json", "model.safetensors", "sentencepiece.bpe.model", "special_tokens_map.json", "tokenizer.json"...
953dc6f6f85a1b2dbfca4c34a2796e7dde08d41e
[ "sentence-transformers", "safetensors", "xlm-roberta", "text-classification", "transformers", "text-embeddings-inference", "multilingual", "arxiv:2312.15503", "arxiv:2402.03216", "license:apache-2.0", "endpoints_compatible", "deploy:azure", "region:us" ]
null
{"architectures": ["XLMRobertaForSequenceClassification"], "model_type": "xlm-roberta", "tokenizer_config": {"bos_token": "<s>", "cls_token": "<s>", "eos_token": "</s>", "mask_token": "<mask>", "pad_token": "<pad>", "sep_token": "</s>", "unk_token": "<unk>"}}
{ "auto_model": "AutoModelForSequenceClassification", "custom_class": null, "pipeline_tag": "text-classification", "processor": "AutoTokenizer" }
{"base_model": null, "datasets": null, "eval_results": null, "language": ["multilingual"], "library_name": null, "license": "apache-2.0", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "text-classification", "tags": ["transformers", "sentence-transformers", "text-embedd...
# Reranker **More details please refer to our Github: [FlagEmbedding](https://github.com/FlagOpen/FlagEmbedding/tree/master).** - [Model List](#model-list) - [Usage](#usage) - [Fine-tuning](#fine-tune) - [Evaluation](#evaluation) - [Citation](#citation) Different from embedding model, reranker uses question and docu...
null
[ "apache-2.0" ]
null
[ "multilingual" ]
567,755,777
null
null
[ "AutoModelForSequenceClassification", "XLMRobertaForSequenceClassification", "xlm-roberta" ]
[ "text-classification" ]
[ "text" ]
[ "text" ]
[ "logits" ]
666ce1af091b7c8e3f9aae7d
microsoft/Florence-2-large
microsoft
null
1,208,523
19,527,799
False
2024-06-15T00:34:55Z
2025-08-04T17:09:36Z
transformers
1,786
7
null
image-text-to-text
{"parameters": {"F16": 776721497}, "total": 776721497}
[ ".gitattributes", "CODE_OF_CONDUCT.md", "LICENSE", "README.md", "SECURITY.md", "SUPPORT.md", "config.json", "configuration_florence2.py", "generation_config.json", "model.safetensors", "modeling_florence2.py", "preprocessor_config.json", "processing_florence2.py", "pytorch_model.bin", "s...
21a599d414c4d928c9032694c424fb94458e3594
[ "transformers", "pytorch", "safetensors", "florence2", "image-text-to-text", "vision", "custom_code", "arxiv:2311.06242", "license:mit", "endpoints_compatible", "region:us" ]
null
{"architectures": ["Florence2ForConditionalGeneration"], "auto_map": {"AutoConfig": "configuration_florence2.Florence2Config", "AutoModelForCausalLM": "modeling_florence2.Florence2ForConditionalGeneration"}, "model_type": "florence2", "tokenizer_config": {}}
{ "auto_model": "AutoModelForImageTextToText", "custom_class": null, "pipeline_tag": "image-text-to-text", "processor": "AutoProcessor" }
{"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": null, "license": "mit", "license_name": null, "license_link": "https://huggingface.co/microsoft/Florence-2-large/resolve/main/LICENSE", "metrics": null, "model_name": null, "pipeline_tag": "image-text-to-text", "tags": ["visi...
null
null
[ "mit", "https://huggingface.co/microsoft/Florence-2-large/resolve/main/LICENSE" ]
null
null
776,721,497
null
null
[ "florence2", "AutoModelForImageTextToText", "Florence2ForConditionalGeneration" ]
[ "image-text-to-text" ]
[ "multimodal" ]
[ "text", "image" ]
[ "text" ]
66e98dd5899bdb384bd953b4
Qwen/Qwen2.5-1.5B-Instruct
Qwen
{ "models": [ { "_id": "66e6d032663c452ad2400439", "id": "Qwen/Qwen2.5-1.5B" } ], "relation": "finetune" }
9,542,607
152,014,423
False
2024-09-17T14:10:29Z
2024-09-25T12:32:50Z
transformers
650
7
null
text-generation
{"parameters": {"BF16": 1543714304}, "total": 1543714304}
[ ".gitattributes", "LICENSE", "README.md", "config.json", "generation_config.json", "merges.txt", "model.safetensors", "tokenizer.json", "tokenizer_config.json", "vocab.json" ]
989aa7980e4cf806f80c7fef2b1adb7bc71aa306
[ "transformers", "safetensors", "qwen2", "text-generation", "chat", "conversational", "en", "arxiv:2407.10671", "base_model:Qwen/Qwen2.5-1.5B", "base_model:finetune:Qwen/Qwen2.5-1.5B", "license:apache-2.0", "text-generation-inference", "endpoints_compatible", "deploy:azure", "region:us" ]
null
{"architectures": ["Qwen2ForCausalLM"], "model_type": "qwen2", "tokenizer_config": {"bos_token": null, "chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0]['role'] == 'system' %}\n {{- messages[0]['content'] }}\n {%- else %}\n {{- 'You are Qwen, created by Aliba...
{ "auto_model": "AutoModelForCausalLM", "custom_class": null, "pipeline_tag": "text-generation", "processor": "AutoTokenizer" }
{"base_model": "Qwen/Qwen2.5-1.5B", "datasets": null, "eval_results": null, "language": ["en"], "library_name": "transformers", "license": "apache-2.0", "license_name": null, "license_link": "https://huggingface.co/Qwen/Qwen2.5-1.5B-Instruct/blob/main/LICENSE", "metrics": null, "model_name": null, "pipeline_tag": "text...
# Qwen2.5-1.5B-Instruct ## Introduction Qwen2.5 is the latest series of Qwen large language models. For Qwen2.5, we release a number of base language models and instruction-tuned language models ranging from 0.5 to 72 billion parameters. Qwen2.5 brings the following improvements upon Qwen2: - Significantly **more kn...
null
[ "apache-2.0", "https://huggingface.co/Qwen/Qwen2.5-1.5B-Instruct/blob/main/LICENSE" ]
null
[ "en" ]
1,543,714,304
null
null
[ "AutoModelForCausalLM", "Qwen2ForCausalLM", "qwen2" ]
[ "text-generation" ]
[ "text" ]
[ "text" ]
[ "text" ]
678e11922b39b4ed1381531b
deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B
deepseek-ai
null
633,494
17,413,391
False
2025-01-20T09:04:18Z
2025-02-24T03:32:35Z
transformers
1,464
7
null
text-generation
{"parameters": {"BF16": 1777088000}, "total": 1777088000}
[ ".gitattributes", "LICENSE", "README.md", "config.json", "figures/benchmark.jpg", "generation_config.json", "model.safetensors", "tokenizer.json", "tokenizer_config.json" ]
ad9f0ae0864d7fbcd1cd905e3c6c5b069cc8b562
[ "transformers", "safetensors", "qwen2", "text-generation", "conversational", "arxiv:2501.12948", "license:mit", "text-generation-inference", "endpoints_compatible", "region:us" ]
null
{"architectures": ["Qwen2ForCausalLM"], "model_type": "qwen2", "tokenizer_config": {"bos_token": {"__type": "AddedToken", "content": "<\uff5cbegin\u2581of\u2581sentence\uff5c>", "lstrip": false, "normalized": true, "rstrip": false, "single_word": false}, "eos_token": {"__type": "AddedToken", "content": "<\uff5cend\u258...
{ "auto_model": "AutoModelForCausalLM", "custom_class": null, "pipeline_tag": "text-generation", "processor": "AutoTokenizer" }
{"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "mit", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": null, "tags": null}
# DeepSeek-R1 <!-- markdownlint-disable first-line-h1 --> <!-- markdownlint-disable html --> <!-- markdownlint-disable no-duplicate-header --> <div align="center"> <img src="https://github.com/deepseek-ai/DeepSeek-V2/blob/main/figures/logo.svg?raw=true" width="60%" alt="DeepSeek-V3" /> </div> <hr> <div align="center...
null
[ "mit" ]
null
null
1,777,088,000
null
null
[ "AutoModelForCausalLM", "Qwen2ForCausalLM", "qwen2" ]
[ "text-generation" ]
[ "text" ]
[ "text" ]
[ "text" ]
67b52d4a824d77f2bba8b0af
microsoft/Phi-4-mini-instruct
microsoft
null
605,689
3,848,451
False
2025-02-19T01:00:58Z
2025-12-10T20:24:40Z
transformers
706
7
null
text-generation
null
[ ".gitattributes", "CODE_OF_CONDUCT.md", "LICENSE", "NOTICE.md", "README.md", "SECURITY.md", "added_tokens.json", "config.json", "configuration_phi3.py", "data_summary_card.md", "generation_config.json", "merges.txt", "model-00001-of-00002.safetensors", "model-00002-of-00002.safetensors", ...
cfbefacb99257ffa30c83adab238a50856ac3083
[ "transformers", "safetensors", "phi3", "text-generation", "nlp", "code", "conversational", "custom_code", "multilingual", "ar", "zh", "cs", "da", "nl", "en", "fi", "fr", "de", "he", "hu", "it", "ja", "ko", "no", "pl", "pt", "ru", "es", "sv", "th", "tr", ...
null
{"architectures": ["Phi3ForCausalLM"], "auto_map": {"AutoConfig": "configuration_phi3.Phi3Config", "AutoModelForCausalLM": "modeling_phi3.Phi3ForCausalLM", "AutoTokenizer": "Xenova/gpt-4o"}, "model_type": "phi3", "tokenizer_config": {"bos_token": "<|endoftext|>", "chat_template": "{% for message in messages %}{% if mes...
{ "auto_model": "AutoModelForCausalLM", "custom_class": null, "pipeline_tag": "text-generation", "processor": "AutoTokenizer" }
{"base_model": null, "datasets": null, "eval_results": null, "language": ["multilingual", "ar", "zh", "cs", "da", "nl", "en", "fi", "fr", "de", "he", "hu", "it", "ja", "ko", "no", "pl", "pt", "ru", "es", "sv", "th", "tr", "uk"], "library_name": "transformers", "license": "mit", "license_name": null, "license_link": "ht...
null
null
[ "mit", "https://huggingface.co/microsoft/Phi-4-mini-instruct/resolve/main/LICENSE" ]
null
[ "multilingual", "ar", "zh", "cs", "da", "nl", "en", "fi", "fr", "de", "he", "hu", "it", "ja", "ko", "no", "pl", "pt", "ru", "es", "sv", "th", "tr", "uk" ]
null
null
null
[ "AutoModelForCausalLM", "phi3", "Phi3ForCausalLM" ]
[ "text-generation" ]
[ "text" ]
[ "text" ]
[ "text" ]
680da7517bcb5198014ad394
Qwen/Qwen3-1.7B
Qwen
{ "models": [ { "_id": "680f0cb340199096ff7bae3e", "id": "Qwen/Qwen3-1.7B-Base" } ], "relation": "finetune" }
6,900,591
30,148,110
False
2025-04-27T03:41:05Z
2025-07-26T03:46:32Z
transformers
435
7
null
text-generation
{"parameters": {"BF16": 2031739904}, "total": 2031739904}
[ ".gitattributes", "LICENSE", "README.md", "config.json", "generation_config.json", "merges.txt", "model-00001-of-00002.safetensors", "model-00002-of-00002.safetensors", "model.safetensors.index.json", "tokenizer.json", "tokenizer_config.json", "vocab.json" ]
70d244cc86ccca08cf5af4e1e306ecf908b1ad5e
[ "transformers", "safetensors", "qwen3", "text-generation", "conversational", "arxiv:2505.09388", "base_model:Qwen/Qwen3-1.7B-Base", "base_model:finetune:Qwen/Qwen3-1.7B-Base", "license:apache-2.0", "text-generation-inference", "endpoints_compatible", "deploy:azure", "region:us" ]
null
{"architectures": ["Qwen3ForCausalLM"], "model_type": "qwen3", "tokenizer_config": {"bos_token": null, "chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0].role == 'system' %}\n {{- messages[0].content + '\\n\\n' }}\n {%- endif %}\n {{- \"# Tools\\n\\nYou may call o...
{ "auto_model": "AutoModelForCausalLM", "custom_class": null, "pipeline_tag": "text-generation", "processor": "AutoTokenizer" }
{"base_model": ["Qwen/Qwen3-1.7B-Base"], "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "apache-2.0", "license_name": null, "license_link": "https://huggingface.co/Qwen/Qwen3-1.7B/blob/main/LICENSE", "metrics": null, "model_name": null, "pipeline_tag": "text-generat...
null
null
[ "apache-2.0", "https://huggingface.co/Qwen/Qwen3-1.7B/blob/main/LICENSE" ]
null
null
2,031,739,904
null
null
[ "AutoModelForCausalLM", "Qwen3ForCausalLM", "qwen3" ]
[ "text-generation" ]
[ "text" ]
[ "text" ]
[ "text" ]
682b8c278503c5bbf012e9a9
ibm-granite/granite-docling-258M
ibm-granite
null
68,982
1,071,300
False
2025-05-19T19:53:11Z
2025-09-23T08:52:16Z
transformers
1,143
7
null
image-text-to-text
null
[ ".gitattributes", "README.md", "added_tokens.json", "assets/granite_docling_split_page.png", "assets/new_arxiv.png", "chat_template.jinja", "config.json", "generation_config.json", "granite_docling.png", "merges.txt", "model.safetensors", "preprocessor_config.json", "processor_config.json", ...
982fe3b40f2fa73c365bdb1bcacf6c81b7184bfe
[ "transformers", "safetensors", "idefics3", "image-text-to-text", "text-generation", "documents", "code", "formula", "chart", "ocr", "layout", "table", "document-parse", "docling", "granite", "extraction", "math", "conversational", "en", "dataset:ds4sd/SynthCodeNet", "dataset:...
null
{"architectures": ["Idefics3ForConditionalGeneration"], "model_type": "idefics3", "tokenizer_config": {"bos_token": "<|start_of_role|>", "eos_token": "<|end_of_text|>", "pad_token": "<|end_of_text|>", "unk_token": "<|unk|>"}, "chat_template_jinja": "{%- for message in messages -%}\n{{- '<|start_of_role|>' + message['ro...
{ "auto_model": "AutoModelForImageTextToText", "custom_class": null, "pipeline_tag": "image-text-to-text", "processor": "AutoProcessor" }
{"base_model": null, "datasets": ["ds4sd/SynthCodeNet", "ds4sd/SynthFormulaNet", "ds4sd/SynthChartNet", "HuggingFaceM4/DoclingMatix"], "eval_results": null, "language": ["en"], "library_name": "transformers", "license": "apache-2.0", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipe...
# granite-docling-258m <div style="display: flex; align-items: center;"> <img src="https://huggingface.co/ibm-granite/granite-docling-258M/resolve/main/granite_docling.png" alt="Granite Docling Logo" style="width: 200px; height: auto; margin-right: 20px;"> <div> <p>Granite Docling is a multimodal Image-...
null
[ "apache-2.0" ]
[ "ds4sd/SynthCodeNet", "ds4sd/SynthFormulaNet", "ds4sd/SynthChartNet", "HuggingFaceM4/DoclingMatix" ]
[ "en" ]
null
null
null
[ "AutoModelForImageTextToText", "Idefics3ForConditionalGeneration", "idefics3" ]
[ "image-text-to-text", "text-generation" ]
[ "text", "multimodal" ]
[ "text", "image" ]
[ "text" ]
686ceee17e3b40a013a9afdc
HuggingFaceTB/SmolLM3-3B
HuggingFaceTB
{ "models": [ { "_id": "6853f5779e7c268e15850d9c", "id": "HuggingFaceTB/SmolLM3-3B-Base" } ], "relation": "finetune" }
1,092,049
2,453,821
False
2025-07-08T10:11:45Z
2025-09-10T12:28:11Z
transformers
920
7
null
text-generation
{"parameters": {"BF16": 3075098624}, "total": 3075098624}
[ ".gitattributes", "README.md", "chat_template.jinja", "config.json", "generation_config.json", "model-00001-of-00002.safetensors", "model-00002-of-00002.safetensors", "model.safetensors.index.json", "notebook.ipynb", "special_tokens_map.json", "tokenizer.json", "tokenizer_config.json" ]
a07cc9a04f16550a088caea529712d1d335b0ac1
[ "transformers", "safetensors", "smollm3", "text-generation", "conversational", "en", "fr", "es", "it", "pt", "zh", "ar", "ru", "base_model:HuggingFaceTB/SmolLM3-3B-Base", "base_model:finetune:HuggingFaceTB/SmolLM3-3B-Base", "license:apache-2.0", "endpoints_compatible", "deploy:azur...
null
{"architectures": ["SmolLM3ForCausalLM"], "model_type": "smollm3", "tokenizer_config": {"bos_token": null, "eos_token": "<|im_end|>", "pad_token": "<|im_end|>"}, "chat_template_jinja": "{# \u2500\u2500\u2500\u2500\u2500 defaults \u2500\u2500\u2500\u2500\u2500 #}\n{%- if enable_thinking is not defined -%}\n{%- set enabl...
{ "auto_model": "AutoModelForCausalLM", "custom_class": null, "pipeline_tag": "text-generation", "processor": "AutoTokenizer" }
{"base_model": ["HuggingFaceTB/SmolLM3-3B-Base"], "datasets": null, "eval_results": null, "language": ["en", "fr", "es", "it", "pt", "zh", "ar", "ru"], "library_name": "transformers", "license": "apache-2.0", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": null, "tags": ...
# SmolLM3 ![image/png](https://cdn-uploads.huggingface.co/production/uploads/61c141342aac764ce1654e43/zy0dqTCCt5IHmuzwoqtJ9.png) ## Table of Contents 1. [Model Summary](#model-summary) 2. [How to use](#how-to-use) 3. [Evaluation](#evaluation) 4. [Training](#training) 5. [Limitations](#limitations) 6. [License](#l...
null
[ "apache-2.0" ]
null
[ "en", "fr", "es", "it", "pt", "zh", "ar", "ru" ]
3,075,098,624
null
null
[ "AutoModelForCausalLM", "SmolLM3ForCausalLM", "smollm3" ]
[ "text-generation" ]
[ "text" ]
[ "text" ]
[ "text" ]
621ffdc136468d709f17ea63
openai/clip-vit-large-patch14
openai
null
24,178,746
1,101,271,075
False
2022-03-02T23:29:05Z
2023-09-15T15:49:35Z
transformers
1,978
6
null
zero-shot-image-classification
{"parameters": {"I64": 334, "F32": 427616512}, "total": 427616846}
[ ".gitattributes", "README.md", "config.json", "flax_model.msgpack", "merges.txt", "model.safetensors", "preprocessor_config.json", "pytorch_model.bin", "special_tokens_map.json", "tf_model.h5", "tokenizer.json", "tokenizer_config.json", "vocab.json" ]
32bd64288804d66eefd0ccbe215aa642df71cc41
[ "transformers", "pytorch", "tf", "jax", "safetensors", "clip", "zero-shot-image-classification", "vision", "arxiv:2103.00020", "arxiv:1908.04913", "endpoints_compatible", "region:us" ]
null
{"architectures": ["CLIPModel"], "model_type": "clip", "tokenizer_config": {"unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "bos_token": {"content": "<|startoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "...
{ "auto_model": "AutoModelForZeroShotImageClassification", "custom_class": null, "pipeline_tag": "zero-shot-image-classification", "processor": "AutoProcessor" }
{"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": null, "license": null, "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": null, "tags": ["vision"], "widget": [{"src": "https://huggingface.co/datasets/mishig/sample_images/resolv...
# Model Card: CLIP Disclaimer: The model card is taken and modified from the official CLIP repository, it can be found [here](https://github.com/openai/CLIP/blob/main/model-card.md). ## Model Details The CLIP model was developed by researchers at OpenAI to learn about what contributes to robustness in computer visio...
null
null
null
null
427,616,846
null
null
[ "AutoModelForZeroShotImageClassification", "CLIPModel", "clip" ]
[ "zero-shot-image-classification" ]
[ "multimodal" ]
[ "text", "image" ]
[ "logits" ]
621ffdc136468d709f1802ec
sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2
sentence-transformers
null
26,743,180
314,376,005
False
2022-03-02T23:29:05Z
2026-01-28T10:02:26Z
sentence-transformers
1,172
6
null
sentence-similarity
{"parameters": {"I64": 512, "F32": 117653760}, "total": 117654272}
[ ".gitattributes", "1_Pooling/config.json", "README.md", "config.json", "config_sentence_transformers.json", "model.safetensors", "modules.json", "onnx/model.onnx", "onnx/model_O1.onnx", "onnx/model_O2.onnx", "onnx/model_O3.onnx", "onnx/model_O4.onnx", "onnx/model_qint8_arm64.onnx", "onnx/m...
e8f8c211226b894fcb81acc59f3b34ba3efd5f42
[ "sentence-transformers", "pytorch", "tf", "onnx", "safetensors", "openvino", "bert", "feature-extraction", "sentence-similarity", "transformers", "multilingual", "ar", "bg", "ca", "cs", "da", "de", "el", "en", "es", "et", "fa", "fi", "fr", "gl", "gu", "he", "hi"...
null
{"architectures": ["BertModel"], "model_type": "bert", "tokenizer_config": {"unk_token": "<unk>", "sep_token": "</s>", "pad_token": "<pad>", "cls_token": "<s>", "mask_token": {"content": "<mask>", "single_word": false, "lstrip": true, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "bos_token": "<s>", "eo...
{ "auto_model": "AutoModel", "custom_class": null, "pipeline_tag": "feature-extraction", "processor": "AutoTokenizer" }
{"base_model": null, "datasets": null, "eval_results": null, "language": ["multilingual", "ar", "bg", "ca", "cs", "da", "de", "el", "en", "es", "et", "fa", "fi", "fr", "gl", "gu", "he", "hi", "hr", "hu", "hy", "id", "it", "ja", "ka", "ko", "ku", "lt", "lv", "mk", "mn", "mr", "ms", "my", "nb", "nl", "pl", "pt", "ro", "r...
# sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2 This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 384 dimensional dense vector space and can be used for tasks like clustering or semantic search. ## Usage (Sentence-Transformers) Using this model become...
null
[ "apache-2.0" ]
null
[ "multilingual", "ar", "bg", "ca", "cs", "da", "de", "el", "en", "es", "et", "fa", "fi", "fr", "gl", "gu", "he", "hi", "hr", "hu", "hy", "id", "it", "ja", "ka", "ko", "ku", "lt", "lv", "mk", "mn", "mr", "ms", "my", "nb", "nl", "pl", "pt", "r...
117,654,272
null
null
[ "BertModel", "AutoModel", "bert" ]
[ "sentence-similarity", "feature-extraction" ]
[ "text", "multimodal" ]
[ "text" ]
[ "logits", "embeddings" ]
676c000762cee1f3abc3ed5f
deepseek-ai/DeepSeek-V3
deepseek-ai
null
754,886
14,403,499
False
2024-12-25T12:52:23Z
2025-03-27T04:01:45Z
transformers
4,019
6
null
text-generation
{"parameters": {"BF16": 3918786560, "F8_E4M3": 680571043840, "F32": 41555600}, "total": 684531386000}
[ ".gitattributes", "LICENSE-CODE", "LICENSE-MODEL", "README.md", "README_WEIGHTS.md", "config.json", "configuration_deepseek.py", "figures/benchmark.png", "figures/niah.png", "inference/configs/config_16B.json", "inference/configs/config_236B.json", "inference/configs/config_671B.json", "infe...
e815299b0bcbac849fa540c768ef21845365c9eb
[ "transformers", "safetensors", "deepseek_v3", "text-generation", "conversational", "custom_code", "arxiv:2412.19437", "eval-results", "text-generation-inference", "endpoints_compatible", "fp8", "region:us" ]
null
{"architectures": ["DeepseekV3ForCausalLM"], "auto_map": {"AutoConfig": "configuration_deepseek.DeepseekV3Config", "AutoModel": "modeling_deepseek.DeepseekV3Model", "AutoModelForCausalLM": "modeling_deepseek.DeepseekV3ForCausalLM"}, "model_type": "deepseek_v3", "quantization_config": {"quant_method": "fp8"}, "tokenizer...
{ "auto_model": "AutoModelForCausalLM", "custom_class": null, "pipeline_tag": "text-generation", "processor": "AutoTokenizer" }
{"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": null, "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": null, "tags": null}
<!-- markdownlint-disable first-line-h1 --> <!-- markdownlint-disable html --> <!-- markdownlint-disable no-duplicate-header --> <div align="center"> <img src="https://github.com/deepseek-ai/DeepSeek-V2/blob/main/figures/logo.svg?raw=true" width="60%" alt="DeepSeek-V3" /> </div> <hr> <div align="center" style="line-...
null
null
null
null
684,531,386,000
null
null
[ "AutoModelForCausalLM", "deepseek_v3", "DeepseekV3ForCausalLM" ]
[ "text-generation" ]
[ "text" ]
[ "text" ]
[ "text" ]
67c35be6eae05d8f94fae4c2
google/gemma-3-12b-it
google
{ "models": [ { "_id": "67c35bd50e998b1f7c4d3139", "id": "google/gemma-3-12b-pt" } ], "relation": "finetune" }
2,534,085
12,192,318
manual
2025-03-01T19:11:34Z
2025-03-21T20:28:56Z
transformers
696
6
null
image-text-to-text
null
[ ".gitattributes", "README.md", "added_tokens.json", "chat_template.json", "config.json", "generation_config.json", "model-00001-of-00005.safetensors", "model-00002-of-00005.safetensors", "model-00003-of-00005.safetensors", "model-00004-of-00005.safetensors", "model-00005-of-00005.safetensors", ...
96b6f1eccf38110c56df3a15bffe176da04bfd80
[ "transformers", "safetensors", "gemma3", "image-text-to-text", "conversational", "arxiv:1905.07830", "arxiv:1905.10044", "arxiv:1911.11641", "arxiv:1904.09728", "arxiv:1705.03551", "arxiv:1911.01547", "arxiv:1907.10641", "arxiv:1903.00161", "arxiv:2009.03300", "arxiv:2304.06364", "arxi...
null
{"architectures": ["Gemma3ForConditionalGeneration"], "model_type": "gemma3", "processor_config": {"chat_template": "{{ bos_token }}\n{%- if messages[0]['role'] == 'system' -%}\n {%- if messages[0]['content'] is string -%}\n {%- set first_user_prefix = messages[0]['content'] + '\n\n' -%}\n {%- else -%}\n ...
{ "auto_model": "AutoModelForImageTextToText", "custom_class": null, "pipeline_tag": "image-text-to-text", "processor": "AutoProcessor" }
{"base_model": "google/gemma-3-12b-pt", "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "gemma", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "image-text-to-text", "tags": null, "extra_gated_heading": "Access Gemma ...
null
null
[ "gemma" ]
null
null
null
null
null
[ "AutoModelForImageTextToText", "Gemma3ForConditionalGeneration", "gemma3" ]
[ "image-text-to-text" ]
[ "multimodal" ]
[ "text", "image" ]
[ "text" ]
67f524d174216a9575d48e68
google/gemma-3-12b-it-qat-q4_0-unquantized
google
{ "models": [ { "_id": "67c35be6eae05d8f94fae4c2", "id": "google/gemma-3-12b-it" } ], "relation": "finetune" }
27,707
77,234
manual
2025-04-08T13:29:53Z
2025-04-15T21:07:07Z
transformers
80
6
null
image-text-to-text
null
[ ".gitattributes", "README.md", "added_tokens.json", "chat_template.json", "config.json", "generation_config.json", "model-00001-of-00005.safetensors", "model-00002-of-00005.safetensors", "model-00003-of-00005.safetensors", "model-00004-of-00005.safetensors", "model-00005-of-00005.safetensors", ...
68f7ee4fbd59087436ada77ed2d62f373fdd4482
[ "transformers", "safetensors", "gemma3", "image-text-to-text", "gemma", "google", "conversational", "arxiv:1905.07830", "arxiv:1905.10044", "arxiv:1911.11641", "arxiv:1904.09728", "arxiv:1705.03551", "arxiv:1911.01547", "arxiv:1907.10641", "arxiv:1903.00161", "arxiv:2009.03300", "arx...
null
{"architectures": ["Gemma3ForConditionalGeneration"], "model_type": "gemma3", "processor_config": {"chat_template": "{{ bos_token }}\n{%- if messages[0]['role'] == 'system' -%}\n {%- if messages[0]['content'] is string -%}\n {%- set first_user_prefix = messages[0]['content'] + '\n\n' -%}\n {%- else -%}\n ...
{ "auto_model": "AutoModelForImageTextToText", "custom_class": null, "pipeline_tag": "image-text-to-text", "processor": "AutoProcessor" }
{"base_model": "google/gemma-3-12b-it", "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "gemma", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "image-text-to-text", "tags": ["gemma3", "gemma", "google"], "extra_gated...
null
null
[ "gemma" ]
null
null
null
null
null
[ "AutoModelForImageTextToText", "Gemma3ForConditionalGeneration", "gemma3" ]
[ "image-text-to-text" ]
[ "multimodal" ]
[ "text", "image" ]
[ "text" ]
684b11e729c32c69e56d1db9
google/gemma-3n-E2B-it
google
{ "models": [ { "_id": "683f3c2354280d882006f816", "id": "google/gemma-3n-E4B-it" } ], "relation": "finetune" }
279,054
1,756,622
manual
2025-06-12T17:44:07Z
2025-07-14T13:55:52Z
transformers
289
6
null
image-text-to-text
null
[ ".gitattributes", "README.md", "chat_template.jinja", "config.json", "generation_config.json", "model-00001-of-00003.safetensors", "model-00002-of-00003.safetensors", "model-00003-of-00003.safetensors", "model.safetensors.index.json", "notebook.ipynb", "preprocessor_config.json", "processor_co...
5e092ebca197cdcd8d8b195040accf22693501bc
[ "transformers", "safetensors", "gemma3n", "image-text-to-text", "automatic-speech-recognition", "automatic-speech-translation", "audio-text-to-text", "video-text-to-text", "conversational", "arxiv:1905.07830", "arxiv:1905.10044", "arxiv:1911.11641", "arxiv:1904.09728", "arxiv:1705.03551", ...
null
{"architectures": ["Gemma3nForConditionalGeneration"], "model_type": "gemma3n", "tokenizer_config": {"bos_token": "<bos>", "eos_token": "<eos>", "pad_token": "<pad>", "unk_token": "<unk>", "use_default_system_prompt": false}, "chat_template_jinja": "{{ bos_token }}\n{%- if messages[0]['role'] == 'system' -%}\n {%- i...
{ "auto_model": "AutoModelForImageTextToText", "custom_class": null, "pipeline_tag": "image-text-to-text", "processor": "AutoProcessor" }
{"base_model": "google/gemma-3n-E4B-it", "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "gemma", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "image-text-to-text", "tags": ["automatic-speech-recognition", "automati...
null
null
[ "gemma" ]
null
null
null
null
null
[ "AutoModelForImageTextToText", "gemma3n", "Gemma3nForConditionalGeneration" ]
[ "image-text-to-text", "video-text-to-text", "automatic-speech-recognition" ]
[ "vision", "multimodal" ]
[ "text", "audio", "image" ]
[ "text" ]
621ffdc036468d709f174347
distilbert/distilbert-base-uncased
distilbert
null
7,004,758
627,000,229
False
2022-03-02T23:29:04Z
2024-05-06T13:44:53Z
transformers
848
5
null
fill-mask
{"parameters": {"F32": 66985530}, "total": 66985530}
[ ".gitattributes", "LICENSE", "README.md", "config.json", "flax_model.msgpack", "model.safetensors", "pytorch_model.bin", "rust_model.ot", "tf_model.h5", "tokenizer.json", "tokenizer_config.json", "vocab.txt" ]
12040accade4e8a0f71eabdb258fecc2e7e948be
[ "transformers", "pytorch", "tf", "jax", "rust", "safetensors", "distilbert", "fill-mask", "exbert", "en", "dataset:bookcorpus", "dataset:wikipedia", "arxiv:1910.01108", "license:apache-2.0", "endpoints_compatible", "deploy:azure", "region:us" ]
null
{"architectures": ["DistilBertForMaskedLM"], "model_type": "distilbert", "tokenizer_config": {}}
{ "auto_model": "AutoModelForMaskedLM", "custom_class": null, "pipeline_tag": "fill-mask", "processor": "AutoTokenizer" }
{"base_model": null, "datasets": ["bookcorpus", "wikipedia"], "eval_results": null, "language": "en", "library_name": null, "license": "apache-2.0", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": null, "tags": ["exbert"]}
# DistilBERT base model (uncased) This model is a distilled version of the [BERT base model](https://huggingface.co/bert-base-uncased). It was introduced in [this paper](https://arxiv.org/abs/1910.01108). The code for the distillation process can be found [here](https://github.com/huggingface/transformers/tree/main/ex...
null
[ "apache-2.0" ]
[ "bookcorpus", "wikipedia" ]
[ "en" ]
66,985,530
null
null
[ "distilbert", "AutoModelForMaskedLM", "DistilBertForMaskedLM" ]
[ "fill-mask" ]
[ "text" ]
[ "text" ]
[ "logits" ]
64fff537d522560505ad6567
BAAI/bge-small-en-v1.5
BAAI
null
11,538,782
118,876,425
False
2023-09-12T05:20:55Z
2024-02-22T03:36:23Z
sentence-transformers
429
5
[{"name": "bge-small-en-v1.5", "results": [{"task": {"type": "Classification"}, "dataset": {"type": "mteb/amazon_counterfactual", "name": "MTEB AmazonCounterfactualClassification (en)", "config": "en", "split": "test", "revision": "e8379541af4e31359cca9fbcf4b00f2671dba205"}, "metrics": [{"type": "accuracy", "value": 73...
feature-extraction
{"parameters": {"I64": 512, "F32": 33360000}, "total": 33360512}
[ ".gitattributes", "1_Pooling/config.json", "README.md", "config.json", "config_sentence_transformers.json", "model.safetensors", "modules.json", "onnx/model.onnx", "pytorch_model.bin", "sentence_bert_config.json", "special_tokens_map.json", "tokenizer.json", "tokenizer_config.json", "vocab...
5c38ec7c405ec4b44b94cc5a9bb96e735b38267a
[ "sentence-transformers", "pytorch", "onnx", "safetensors", "bert", "feature-extraction", "sentence-similarity", "transformers", "mteb", "en", "arxiv:2401.03462", "arxiv:2312.15503", "arxiv:2311.13534", "arxiv:2310.07554", "arxiv:2309.07597", "license:mit", "model-index", "text-embe...
null
{"architectures": ["BertModel"], "model_type": "bert", "tokenizer_config": {"cls_token": "[CLS]", "mask_token": "[MASK]", "pad_token": "[PAD]", "sep_token": "[SEP]", "unk_token": "[UNK]"}}
{ "auto_model": "AutoModel", "custom_class": null, "pipeline_tag": "feature-extraction", "processor": "AutoTokenizer" }
{"language": ["en"], "license": "mit", "tags": ["sentence-transformers", "feature-extraction", "sentence-similarity", "transformers", "mteb"], "model-index": [{"name": "bge-small-en-v1.5", "results": [{"task": {"type": "Classification"}, "dataset": {"name": "MTEB AmazonCounterfactualClassification (en)", "type": "mteb/...
<h1 align="center">FlagEmbedding</h1> <h4 align="center"> <p> <a href=#model-list>Model List</a> | <a href=#frequently-asked-questions>FAQ</a> | <a href=#usage>Usage</a> | <a href="#evaluation">Evaluation</a> | <a href="#train">Train</a> | <a href="#contact">Conta...
null
[ "mit" ]
null
[ "en" ]
33,360,512
null
null
[ "BertModel", "AutoModel", "bert" ]
[ "sentence-similarity", "feature-extraction" ]
[ "text", "multimodal" ]
[ "text" ]
[ "logits", "embeddings" ]
65770c3426ef61bbf101d4da
mistralai/Mistral-7B-Instruct-v0.2
mistralai
null
2,719,964
53,687,337
False
2023-12-11T13:18:44Z
2025-07-24T16:57:21Z
transformers
3,098
5
null
text-generation
{"parameters": {"BF16": 7241732096}, "total": 7241732096}
[ ".gitattributes", "README.md", "config.json", "generation_config.json", "model-00001-of-00003.safetensors", "model-00002-of-00003.safetensors", "model-00003-of-00003.safetensors", "model.safetensors.index.json", "pytorch_model-00001-of-00003.bin", "pytorch_model-00002-of-00003.bin", "pytorch_mod...
63a8b081895390a26e140280378bc85ec8bce07a
[ "transformers", "pytorch", "safetensors", "mistral", "text-generation", "finetuned", "mistral-common", "conversational", "arxiv:2310.06825", "license:apache-2.0", "text-generation-inference", "deploy:azure", "region:us" ]
null
{"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{%- if messages[0]['role'] == 'system' %}\n {%- set system_message = messages[0]['content'] %}\n {%- set loop_messages = messages[1:] %}\n{%- else %}\n {%- set loop_messages = messages ...
{ "auto_model": "AutoModelForCausalLM", "custom_class": null, "pipeline_tag": "text-generation", "processor": "AutoTokenizer" }
{"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "apache-2.0", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": null, "tags": ["finetuned", "mistral-common"], "new_version": "mistralai/Mistral-7B-Ins...
null
null
[ "apache-2.0" ]
null
null
7,241,732,096
null
null
[ "AutoModelForCausalLM", "mistral", "MistralForCausalLM" ]
[ "text-generation" ]
[ "text" ]
[ "text" ]
[ "text" ]
658fb85235c41262d661dc48
TinyLlama/TinyLlama-1.1B-Chat-v1.0
TinyLlama
null
2,817,808
35,656,120
False
2023-12-30T06:27:30Z
2024-03-17T05:07:08Z
transformers
1,556
5
null
text-generation
{"parameters": {"BF16": 1100048384}, "total": 1100048384}
[ ".gitattributes", "README.md", "config.json", "eval_results.json", "generation_config.json", "model.safetensors", "special_tokens_map.json", "tokenizer.json", "tokenizer.model", "tokenizer_config.json" ]
fe8a4ea1ffedaf415f4da2f062534de366a451e6
[ "transformers", "safetensors", "llama", "text-generation", "conversational", "en", "dataset:cerebras/SlimPajama-627B", "dataset:bigcode/starcoderdata", "dataset:HuggingFaceH4/ultrachat_200k", "dataset:HuggingFaceH4/ultrafeedback_binarized", "license:apache-2.0", "text-generation-inference", ...
null
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{% for message in messages %}\n{% if message['role'] == 'user' %}\n{{ '<|user|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'system' %}\n{{ '<|system|>\n' + message['content']...
{ "auto_model": "AutoModelForCausalLM", "custom_class": null, "pipeline_tag": "text-generation", "processor": "AutoTokenizer" }
{"base_model": null, "datasets": ["cerebras/SlimPajama-627B", "bigcode/starcoderdata", "HuggingFaceH4/ultrachat_200k", "HuggingFaceH4/ultrafeedback_binarized"], "eval_results": null, "language": ["en"], "library_name": null, "license": "apache-2.0", "license_name": null, "license_link": null, "metrics": null, "model_na...
<div align="center"> # TinyLlama-1.1B </div> https://github.com/jzhang38/TinyLlama The TinyLlama project aims to **pretrain** a **1.1B Llama model on 3 trillion tokens**. With some proper optimization, we can achieve this within a span of "just" 90 days using 16 A100-40G GPUs 🚀🚀. The training has started on 2023-0...
null
[ "apache-2.0" ]
[ "cerebras/SlimPajama-627B", "bigcode/starcoderdata", "HuggingFaceH4/ultrachat_200k", "HuggingFaceH4/ultrafeedback_binarized" ]
[ "en" ]
1,100,048,384
null
null
[ "AutoModelForCausalLM", "llama", "LlamaForCausalLM" ]
[ "text-generation" ]
[ "text" ]
[ "text" ]
[ "text" ]
65e60cda4b2e0f45e49dc03d
vikhyatk/moondream2
vikhyatk
null
4,627,664
24,499,658
False
2024-03-04T18:03:06Z
2025-09-23T01:22:59Z
transformers
1,399
5
null
image-text-to-text
{"parameters": {"BF16": 1927237104}, "total": 1927237104}
[ ".gitattributes", "README.md", "added_tokens.json", "config.json", "config.py", "configuration_moondream.py", "fourier_features.py", "generation_config.json", "handler.py", "hf_moondream.py", "image_crops.py", "layers.py", "lora.py", "merges.txt", "model.safetensors", "moondream.py", ...
6b714b26eea5cbd9f31e4edb2541c170afa935ba
[ "transformers", "safetensors", "moondream1", "text-generation", "image-text-to-text", "custom_code", "doi:10.57967/hf/6762", "license:apache-2.0", "endpoints_compatible", "region:us" ]
null
{"architectures": ["HfMoondream"], "auto_map": {"AutoConfig": "hf_moondream.HfConfig", "AutoModelForCausalLM": "hf_moondream.HfMoondream"}, "model_type": "moondream1", "tokenizer_config": {"bos_token": "<|endoftext|>", "eos_token": "<|endoftext|>", "unk_token": "<|endoftext|>"}}
{ "auto_model": "AutoModelForCausalLM", "custom_class": "hf_moondream.HfMoondream", "pipeline_tag": "text-generation", "processor": null }
{"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": null, "license": "apache-2.0", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "image-text-to-text", "tags": null, "new_version": "moondream/moondream3-preview"}
⚠️ This repository contains the latest version of Moondream 2, our previous generation model. The latest version of Moondream is [Moondream 3 (Preview)](https://huggingface.co/moondream/moondream3-preview). --- Moondream is a small vision language model designed to run efficiently everywhere. [Website](https://moon...
null
[ "apache-2.0" ]
null
null
1,927,237,104
null
null
[ "HfMoondream", "hf_moondream.HfMoondream", "AutoModelForCausalLM", "moondream1" ]
[ "image-text-to-text", "text-generation" ]
[ "text", "multimodal" ]
[ "text", "image" ]
[ "text" ]
6641487429be04778b9180f0
google/paligemma-3b-pt-224
google
null
81,889
887,103
manual
2024-05-12T22:53:40Z
2024-09-21T10:14:25Z
transformers
426
5
null
image-text-to-text
null
[ ".gitattributes", "README.md", "added_tokens.json", "config.json", "generation_config.json", "model-00001-of-00003.safetensors", "model-00002-of-00003.safetensors", "model-00003-of-00003.safetensors", "model.safetensors.index.json", "preprocessor_config.json", "special_tokens_map.json", "token...
35e4f46485b4d07967e7e9935bc3786aad50687c
[ "transformers", "safetensors", "paligemma", "image-text-to-text", "arxiv:2310.09199", "arxiv:2303.15343", "arxiv:2403.08295", "arxiv:1706.03762", "arxiv:2010.11929", "arxiv:2209.06794", "arxiv:2209.04372", "arxiv:2103.01913", "arxiv:2205.12522", "arxiv:2110.11624", "arxiv:2108.03353", ...
null
{"architectures": ["PaliGemmaForConditionalGeneration"], "model_type": "paligemma", "tokenizer_config": {"bos_token": "<bos>", "eos_token": "<eos>", "pad_token": "<pad>", "unk_token": "<unk>", "use_default_system_prompt": false}}
{ "auto_model": "AutoModelForImageTextToText", "custom_class": null, "pipeline_tag": "image-text-to-text", "processor": "AutoProcessor" }
{"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "gemma", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "image-text-to-text", "tags": null, "extra_gated_heading": "Access PaliGemma on Hugging Face...
null
null
[ "gemma" ]
null
null
null
null
null
[ "PaliGemmaForConditionalGeneration", "AutoModelForImageTextToText", "paligemma" ]
[ "image-text-to-text" ]
[ "multimodal" ]
[ "text", "image" ]
[ "text" ]
66592ee81ba271cba98959cb
Tiiny/TurboSparse-Mixtral
Tiiny
null
20
452
False
2024-05-31T01:59:04Z
2024-06-14T02:13:18Z
transformers
42
5
null
feature-extraction
{"parameters": {"BF16": 51796250624}, "total": 51796250624}
[ ".gitattributes", "README.md", "added_tokens.json", "config.json", "configuration_turbosparsemixtral.py", "generation_config.json", "model-00001-of-00021.safetensors", "model-00002-of-00021.safetensors", "model-00003-of-00021.safetensors", "model-00004-of-00021.safetensors", "model-00005-of-0002...
706a33e3e53182035d2e8e2c7f4ed6d0fc3e272e
[ "transformers", "safetensors", "turbosparsemixtral", "feature-extraction", "custom_code", "en", "arxiv:2406.05955", "license:apache-2.0", "region:us" ]
null
{"architectures": ["TurboSparseMixtralForCausalLM"], "auto_map": {"AutoConfig": "configuration_turbosparsemixtral.TurboSparseMixtralConfig", "AutoModel": "modeling_turbosparsemixtral.TurboSparseMixtralForCausalLM", "AutoModelForCausalLM": "modeling_turbosparsemixtral.TurboSparseMixtralForCausalLM"}, "model_type": "turb...
{ "auto_model": "AutoModel", "custom_class": "modeling_turbosparsemixtral.TurboSparseMixtralForCausalLM", "pipeline_tag": "feature-extraction", "processor": null }
{"base_model": null, "datasets": null, "eval_results": null, "language": ["en"], "library_name": null, "license": "apache-2.0", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": null, "tags": null}
null
null
[ "apache-2.0" ]
null
[ "en" ]
51,796,250,624
null
null
[ "AutoModel", "modeling_turbosparsemixtral.TurboSparseMixtralForCausalLM", "TurboSparseMixtralForCausalLM", "turbosparsemixtral" ]
[ "feature-extraction" ]
[ "multimodal" ]
[ "text" ]
[ "embeddings" ]
66672b6f2e5f1c08ecf1c61d
openvla/openvla-7b
openvla
null
1,196,859
11,449,520
False
2024-06-10T16:35:59Z
2026-02-17T03:43:23Z
transformers
197
5
null
robotics
{"parameters": {"BF16": 7541237184}, "total": 7541237184}
[ ".gitattributes", "README.md", "added_tokens.json", "config.json", "configuration_prismatic.py", "generation_config.json", "model-00001-of-00003.safetensors", "model-00002-of-00003.safetensors", "model-00003-of-00003.safetensors", "model.safetensors.index.json", "modeling_prismatic.py", "prepr...
47a0ec7fc4ec123775a391911046cf33cf9ed83f
[ "transformers", "safetensors", "openvla", "feature-extraction", "robotics", "vla", "image-text-to-text", "multimodal", "pretraining", "custom_code", "en", "arxiv:2406.09246", "license:mit", "region:us" ]
null
{"architectures": ["OpenVLAForActionPrediction"], "auto_map": {"AutoConfig": "configuration_prismatic.OpenVLAConfig", "AutoModelForVision2Seq": "modeling_prismatic.OpenVLAForActionPrediction"}, "model_type": "openvla", "tokenizer_config": {"bos_token": "<s>", "eos_token": "</s>", "pad_token": "<PAD>", "unk_token": "<un...
{ "auto_model": "AutoModelForVision2Seq", "custom_class": "modeling_prismatic.OpenVLAForActionPrediction", "pipeline_tag": "feature-extraction", "processor": null }
{"base_model": null, "datasets": null, "eval_results": null, "language": ["en"], "library_name": "transformers", "license": "mit", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "robotics", "tags": ["robotics", "vla", "image-text-to-text", "multimodal", "pretraining"]}
# OpenVLA 7B OpenVLA 7B (`openvla-7b`) is an open vision-language-action model trained on 970K robot manipulation episodes from the [Open X-Embodiment](https://robotics-transformer-x.github.io/) dataset. The model takes language instructions and camera images as input and generates robot actions. It supports controll...
null
[ "mit" ]
null
[ "en" ]
7,541,237,184
null
null
[ "AutoModelForVision2Seq", "modeling_prismatic.OpenVLAForActionPrediction", "openvla", "OpenVLAForActionPrediction" ]
[ "image-text-to-text", "robotics", "feature-extraction" ]
[ "multimodal" ]
[ "text", "image" ]
[ "text", "embeddings" ]
666be1e54e6657b7d1a78f80
deepseek-ai/DeepSeek-Coder-V2-Lite-Instruct
deepseek-ai
null
319,323
5,373,292
False
2024-06-14T06:23:33Z
2024-07-03T05:16:11Z
transformers
570
5
null
text-generation
{"parameters": {"BF16": 15706484224}, "total": 15706484224}
[ ".gitattributes", "README.md", "config.json", "configuration_deepseek.py", "generation_config.json", "model-00001-of-000004.safetensors", "model-00002-of-000004.safetensors", "model-00003-of-000004.safetensors", "model-00004-of-000004.safetensors", "model.safetensors.index.json", "modeling_deeps...
e434a23f91ba5b4923cf6c9d9a238eb4a08e3a11
[ "transformers", "safetensors", "deepseek_v2", "text-generation", "conversational", "custom_code", "arxiv:2401.06066", "license:other", "text-generation-inference", "endpoints_compatible", "region:us" ]
null
{"architectures": ["DeepseekV2ForCausalLM"], "auto_map": {"AutoConfig": "configuration_deepseek.DeepseekV2Config", "AutoModel": "modeling_deepseek.DeepseekV2Model", "AutoModelForCausalLM": "modeling_deepseek.DeepseekV2ForCausalLM"}, "model_type": "deepseek_v2", "tokenizer_config": {"bos_token": {"__type": "AddedToken",...
{ "auto_model": "AutoModelForCausalLM", "custom_class": null, "pipeline_tag": "text-generation", "processor": "AutoTokenizer" }
{"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": null, "license": "other", "license_name": "deepseek-license", "license_link": "LICENSE", "metrics": null, "model_name": null, "pipeline_tag": null, "tags": null}
<!-- markdownlint-disable first-line-h1 --> <!-- markdownlint-disable html --> <!-- markdownlint-disable no-duplicate-header --> <div align="center"> <img src="https://github.com/deepseek-ai/DeepSeek-V2/blob/main/figures/logo.svg?raw=true" width="60%" alt="DeepSeek-V2" /> </div> <hr> <div align="center" style="line-...
null
[ "other", "deepseek-license", "LICENSE" ]
null
null
15,706,484,224
null
null
[ "AutoModelForCausalLM", "DeepseekV2ForCausalLM", "deepseek_v2" ]
[ "text-generation" ]
[ "text" ]
[ "text" ]
[ "text" ]
66962a41534f204a2b9982fd
google/gemma-2-2b
google
null
536,112
24,639,131
manual
2024-07-16T08:07:29Z
2024-08-07T18:25:49Z
transformers
636
5
null
text-generation
null
[ ".gitattributes", "README.md", "config.json", "generation_config.json", "model-00001-of-00003.safetensors", "model-00002-of-00003.safetensors", "model-00003-of-00003.safetensors", "model.safetensors.index.json", "special_tokens_map.json", "tokenizer.json", "tokenizer.model", "tokenizer_config....
c5ebcd40d208330abc697524c919956e692655cf
[ "transformers", "safetensors", "gemma2", "text-generation", "arxiv:2009.03300", "arxiv:1905.07830", "arxiv:1911.11641", "arxiv:1904.09728", "arxiv:1905.10044", "arxiv:1907.10641", "arxiv:1811.00937", "arxiv:1809.02789", "arxiv:1911.01547", "arxiv:1705.03551", "arxiv:2107.03374", "arxiv...
null
{"architectures": ["Gemma2ForCausalLM"], "model_type": "gemma2", "tokenizer_config": {"bos_token": "<bos>", "eos_token": "<eos>", "pad_token": "<pad>", "unk_token": "<unk>", "use_default_system_prompt": false}}
{ "auto_model": "AutoModelForCausalLM", "custom_class": null, "pipeline_tag": "text-generation", "processor": "AutoTokenizer" }
{"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "gemma", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "text-generation", "tags": null, "extra_gated_heading": "Access Gemma on Hugging Face", "ext...
null
null
[ "gemma" ]
null
null
null
null
null
[ "AutoModelForCausalLM", "gemma2", "Gemma2ForCausalLM" ]
[ "text-generation" ]
[ "text" ]
[ "text" ]
[ "text" ]
669650bb11dbbf600cf4dcf0
google/gemma-2-2b-it
google
{ "models": [ { "_id": "66962a41534f204a2b9982fd", "id": "google/gemma-2-2b" } ], "relation": "finetune" }
390,438
8,672,814
manual
2024-07-16T10:51:39Z
2024-08-27T19:41:44Z
transformers
1,310
5
null
text-generation
{"parameters": {"BF16": 2614341888}, "total": 2614341888}
[ ".gitattributes", "README.md", "config.json", "generation_config.json", "model-00001-of-00002.safetensors", "model-00002-of-00002.safetensors", "model.safetensors.index.json", "special_tokens_map.json", "tokenizer.json", "tokenizer.model", "tokenizer_config.json" ]
299a8560bedf22ed1c72a8a11e7dce4a7f9f51f8
[ "transformers", "safetensors", "gemma2", "text-generation", "conversational", "arxiv:2009.03300", "arxiv:1905.07830", "arxiv:1911.11641", "arxiv:1904.09728", "arxiv:1905.10044", "arxiv:1907.10641", "arxiv:1811.00937", "arxiv:1809.02789", "arxiv:1911.01547", "arxiv:1705.03551", "arxiv:2...
null
{"architectures": ["Gemma2ForCausalLM"], "model_type": "gemma2", "tokenizer_config": {"bos_token": "<bos>", "chat_template": "{{ bos_token }}{% if messages[0]['role'] == 'system' %}{{ raise_exception('System role not supported') }}{% endif %}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0...
{ "auto_model": "AutoModelForCausalLM", "custom_class": null, "pipeline_tag": "text-generation", "processor": "AutoTokenizer" }
{"base_model": "google/gemma-2-2b", "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "gemma", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "text-generation", "tags": ["conversational"], "extra_gated_heading": "Access...
null
null
[ "gemma" ]
null
null
2,614,341,888
null
null
[ "AutoModelForCausalLM", "gemma2", "Gemma2ForCausalLM" ]
[ "text-generation" ]
[ "text" ]
[ "text" ]
[ "text" ]
6722d34c3189f1de8ad8ce22
HuggingFaceTB/SmolLM2-135M
HuggingFaceTB
null
1,056,461
10,564,948
False
2024-10-31T00:46:04Z
2025-02-06T10:34:41Z
transformers
177
5
null
text-generation
{"parameters": {"BF16": 134515008}, "total": 134515008}
[ ".gitattributes", "README.md", "config.json", "generation_config.json", "merges.txt", "model.safetensors", "special_tokens_map.json", "tokenizer.json", "tokenizer_config.json", "vocab.json" ]
93efa2f097d58c2a74874c7e644dbc9b0cee75a2
[ "transformers", "safetensors", "llama", "text-generation", "en", "arxiv:2502.02737", "license:apache-2.0", "text-generation-inference", "endpoints_compatible", "deploy:azure", "region:us" ]
null
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|endoftext|>", "eos_token": "<|endoftext|>", "unk_token": "<|endoftext|>"}}
{ "auto_model": "AutoModelForCausalLM", "custom_class": null, "pipeline_tag": "text-generation", "processor": "AutoTokenizer" }
{"base_model": null, "datasets": null, "eval_results": null, "language": ["en"], "library_name": "transformers", "license": "apache-2.0", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": null, "tags": null}
# SmolLM2 ![image/png](https://cdn-uploads.huggingface.co/production/uploads/61c141342aac764ce1654e43/XtSR4NkriicR6fGiWGowZ.png) ## Table of Contents 1. [Model Summary](##model-summary) 2. [Limitations](##limitations) 3. [Training](##training) 4. [License](##license) 5. [Citation](##citation) ## Model Summary Smo...
null
[ "apache-2.0" ]
null
[ "en" ]
134,515,008
null
null
[ "AutoModelForCausalLM", "llama", "LlamaForCausalLM" ]
[ "text-generation" ]
[ "text" ]
[ "text" ]
[ "text" ]
678e145ed35c6f68a26df940
deepseek-ai/DeepSeek-R1-Distill-Qwen-7B
deepseek-ai
null
620,266
12,793,856
False
2025-01-20T09:16:14Z
2025-02-24T03:32:20Z
transformers
799
5
null
text-generation
{"parameters": {"BF16": 7615616512}, "total": 7615616512}
[ ".gitattributes", "LICENSE", "README.md", "config.json", "figures/benchmark.jpg", "generation_config.json", "model-00001-of-000002.safetensors", "model-00002-of-000002.safetensors", "model.safetensors.index.json", "tokenizer.json", "tokenizer_config.json" ]
916b56a44061fd5cd7d6a8fb632557ed4f724f60
[ "transformers", "safetensors", "qwen2", "text-generation", "conversational", "arxiv:2501.12948", "license:mit", "text-generation-inference", "endpoints_compatible", "region:us" ]
null
{"architectures": ["Qwen2ForCausalLM"], "model_type": "qwen2", "tokenizer_config": {"bos_token": {"__type": "AddedToken", "content": "<\uff5cbegin\u2581of\u2581sentence\uff5c>", "lstrip": false, "normalized": true, "rstrip": false, "single_word": false}, "eos_token": {"__type": "AddedToken", "content": "<\uff5cend\u258...
{ "auto_model": "AutoModelForCausalLM", "custom_class": null, "pipeline_tag": "text-generation", "processor": "AutoTokenizer" }
{"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "mit", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": null, "tags": null}
# DeepSeek-R1 <!-- markdownlint-disable first-line-h1 --> <!-- markdownlint-disable html --> <!-- markdownlint-disable no-duplicate-header --> <div align="center"> <img src="https://github.com/deepseek-ai/DeepSeek-V2/blob/main/figures/logo.svg?raw=true" width="60%" alt="DeepSeek-V3" /> </div> <hr> <div align="center...
null
[ "mit" ]
null
null
7,615,616,512
null
null
[ "AutoModelForCausalLM", "Qwen2ForCausalLM", "qwen2" ]
[ "text-generation" ]
[ "text" ]
[ "text" ]
[ "text" ]
678e14e3ccefc7605ae5d807
deepseek-ai/DeepSeek-R1-Distill-Qwen-14B
deepseek-ai
null
567,723
6,242,713
False
2025-01-20T09:18:27Z
2025-02-24T03:31:45Z
transformers
615
5
null
text-generation
{"parameters": {"BF16": 14770033664}, "total": 14770033664}
[ ".gitattributes", "LICENSE", "README.md", "config.json", "figures/benchmark.jpg", "generation_config.json", "model-00001-of-000004.safetensors", "model-00002-of-000004.safetensors", "model-00003-of-000004.safetensors", "model-00004-of-000004.safetensors", "model.safetensors.index.json", "token...
1df8507178afcc1bef68cd8c393f61a886323761
[ "transformers", "safetensors", "qwen2", "text-generation", "conversational", "arxiv:2501.12948", "license:mit", "text-generation-inference", "endpoints_compatible", "region:us" ]
null
{"architectures": ["Qwen2ForCausalLM"], "model_type": "qwen2", "tokenizer_config": {"bos_token": {"__type": "AddedToken", "content": "<\uff5cbegin\u2581of\u2581sentence\uff5c>", "lstrip": false, "normalized": true, "rstrip": false, "single_word": false}, "eos_token": {"__type": "AddedToken", "content": "<\uff5cend\u258...
{ "auto_model": "AutoModelForCausalLM", "custom_class": null, "pipeline_tag": "text-generation", "processor": "AutoTokenizer" }
{"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "mit", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": null, "tags": null}
# DeepSeek-R1 <!-- markdownlint-disable first-line-h1 --> <!-- markdownlint-disable html --> <!-- markdownlint-disable no-duplicate-header --> <div align="center"> <img src="https://github.com/deepseek-ai/DeepSeek-V2/blob/main/figures/logo.svg?raw=true" width="60%" alt="DeepSeek-V3" /> </div> <hr> <div align="center...
null
[ "mit" ]
null
null
14,770,033,664
null
null
[ "AutoModelForCausalLM", "Qwen2ForCausalLM", "qwen2" ]
[ "text-generation" ]
[ "text" ]
[ "text" ]
[ "text" ]
68125124ea07c4672522dca0
darkc0de/XortronCriminalComputingConfig
darkc0de
{ "models": [ { "_id": "67aacebd6b842ad4f7d26c64", "id": "TroyDoesAI/BlackSheep-24B" }, { "_id": "68125124ea07c4672522dca0", "id": "darkc0de/XortronCriminalComputingConfig" } ], "relation": "merge" }
668
7,683
False
2025-04-30T16:34:44Z
2026-03-24T20:22:40Z
transformers
149
5
null
text-generation
{"parameters": {"F16": 23572403200}, "total": 23572403200}
[ "README.md", "config.json", "gitattributes", "mergekit_config.yml", "model-00001-of-00010.safetensors", "model-00002-of-00010.safetensors", "model-00003-of-00010.safetensors", "model-00004-of-00010.safetensors", "model-00005-of-00010.safetensors", "model-00006-of-00010.safetensors", "model-00007...
940cf735f387b133bd8178a4db89b0d3c6f1ccd2
[ "transformers", "safetensors", "mistral", "text-generation", "mergekit", "merge", "uncensored", "harmful", "conversational", "en", "base_model:TroyDoesAI/BlackSheep-24B", "base_model:merge:TroyDoesAI/BlackSheep-24B", "base_model:darkc0de/XortronCriminalComputingConfig", "base_model:merge:d...
null
{"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{%- set today = strftime_now(\"%Y-%m-%d\") %}\n{%- set default_system_message = \"You are Xortron. You are an uncensored AI model. You were created by Dr. Buddy Glass and Dr. Sonny DeSorbo at th...
{ "auto_model": "AutoModelForCausalLM", "custom_class": null, "pipeline_tag": "text-generation", "processor": "AutoTokenizer" }
{"base_model": ["darkc0de/XortronCriminalComputing", "TroyDoesAI/BlackSheep-24B", "darkc0de/XortronCriminalComputingConfig"], "datasets": null, "eval_results": null, "language": ["en"], "library_name": "transformers", "license": "apache-2.0", "license_name": null, "license_link": null, "metrics": null, "model_name": nu...
null
null
[ "apache-2.0" ]
null
[ "en" ]
23,572,403,200
null
null
[ "AutoModelForCausalLM", "mistral", "MistralForCausalLM" ]
[ "text-generation" ]
[ "text" ]
[ "text" ]
[ "text" ]
688b1597e5e83e19d1b3238a
Qwen/Qwen3-Coder-30B-A3B-Instruct
Qwen
null
1,134,092
5,693,237
False
2025-07-31T07:04:55Z
2025-12-03T08:05:17Z
transformers
985
5
null
text-generation
{"parameters": {"BF16": 30532122624}, "total": 30532122624}
[ ".gitattributes", "LICENSE", "README.md", "chat_template.jinja", "config.json", "generation_config.json", "merges.txt", "model-00001-of-00016.safetensors", "model-00002-of-00016.safetensors", "model-00003-of-00016.safetensors", "model-00004-of-00016.safetensors", "model-00005-of-00016.safetens...
b2cff646eb4bb1d68355c01b18ae02e7cf42d120
[ "transformers", "safetensors", "qwen3_moe", "text-generation", "conversational", "arxiv:2505.09388", "license:apache-2.0", "endpoints_compatible", "deploy:azure", "region:us" ]
null
{"architectures": ["Qwen3MoeForCausalLM"], "model_type": "qwen3_moe", "tokenizer_config": {"bos_token": null, "chat_template": "{% macro render_extra_keys(json_dict, handled_keys) %}\n {%- if json_dict is mapping %}\n {%- for json_key in json_dict if json_key not in handled_keys %}\n {%- if json_di...
{ "auto_model": "AutoModelForCausalLM", "custom_class": null, "pipeline_tag": "text-generation", "processor": "AutoTokenizer" }
{"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "apache-2.0", "license_name": null, "license_link": "https://huggingface.co/Qwen/Qwen3-Coder-30B-A3B-Instruct/blob/main/LICENSE", "metrics": null, "model_name": null, "pipeline_tag": "text-generatio...
# Qwen3-Coder-30B-A3B-Instruct <a href="https://chat.qwen.ai/" target="_blank" style="margin: 2px;"> <img alt="Chat" src="https://img.shields.io/badge/%F0%9F%92%9C%EF%B8%8F%20Qwen%20Chat%20-536af5" style="display: inline-block; vertical-align: middle;"/> </a> ## Highlights **Qwen3-Coder** is available in multiple...
null
[ "apache-2.0", "https://huggingface.co/Qwen/Qwen3-Coder-30B-A3B-Instruct/blob/main/LICENSE" ]
null
null
30,532,122,624
null
null
[ "qwen3_moe", "AutoModelForCausalLM", "Qwen3MoeForCausalLM" ]
[ "text-generation" ]
[ "text" ]
[ "text" ]
[ "text" ]
689252773b8900ddb9116aed
google/gemma-3-270m
google
null
87,092
830,635
manual
2025-08-05T18:50:31Z
2025-08-14T07:35:01Z
transformers
1,001
5
null
text-generation
{"parameters": {"BF16": 268098176}, "total": 268098176}
[ ".gitattributes", "README.md", "added_tokens.json", "config.json", "generation_config.json", "model.safetensors", "special_tokens_map.json", "tokenizer.json", "tokenizer.model", "tokenizer_config.json" ]
9b0cfec892e2bc2afd938c98eabe4e4a7b1e0ca1
[ "transformers", "safetensors", "gemma3_text", "text-generation", "gemma3", "gemma", "google", "arxiv:2503.19786", "arxiv:1905.07830", "arxiv:1905.10044", "arxiv:1911.11641", "arxiv:1705.03551", "arxiv:1911.01547", "arxiv:1907.10641", "arxiv:2311.07911", "arxiv:2311.12022", "arxiv:241...
null
{"architectures": ["Gemma3ForCausalLM"], "model_type": "gemma3_text", "tokenizer_config": {"bos_token": "<bos>", "eos_token": "<eos>", "pad_token": "<pad>", "unk_token": "<unk>", "use_default_system_prompt": false}}
{ "auto_model": "AutoModelForCausalLM", "custom_class": null, "pipeline_tag": "text-generation", "processor": "AutoTokenizer" }
{"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "gemma", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "text-generation", "tags": ["gemma3", "gemma", "google"], "extra_gated_heading": "Access Gem...
null
null
[ "gemma" ]
null
null
268,098,176
null
null
[ "gemma3_text", "AutoModelForCausalLM", "Gemma3ForCausalLM" ]
[ "text-generation" ]
[ "text" ]
[ "text" ]
[ "text" ]
68e51bdbe980d4b93323e6b8
LiquidAI/LFM2-8B-A1B
LiquidAI
null
47,879
142,143
False
2025-10-07T13:55:39Z
2026-03-04T12:20:04Z
transformers
342
5
null
text-generation
{"parameters": {"F32": 704, "BF16": 8339929856}, "total": 8339929856}
[ ".gitattributes", "LICENSE", "README.md", "chat_template.jinja", "config.json", "generation_config.json", "model-00001-of-00004.safetensors", "model-00002-of-00004.safetensors", "model-00003-of-00004.safetensors", "model-00004-of-00004.safetensors", "model.safetensors.index.json", "special_tok...
e2277e438c87cef87a57a8d437a7e74d58717939
[ "transformers", "safetensors", "lfm2_moe", "text-generation", "liquid", "lfm2", "edge", "moe", "conversational", "en", "ar", "zh", "fr", "de", "ja", "ko", "es", "arxiv:2511.23404", "license:other", "endpoints_compatible", "deploy:azure", "region:us" ]
null
{"architectures": ["Lfm2MoeForCausalLM"], "model_type": "lfm2_moe", "tokenizer_config": {"bos_token": "<|startoftext|>", "eos_token": "<|im_end|>", "pad_token": "<|pad|>", "use_default_system_prompt": false}, "chat_template_jinja": "{{- bos_token -}}{%- set system_prompt = \"\" -%}{%- set ns = namespace(system_prompt=\...
{ "auto_model": "AutoModelForCausalLM", "custom_class": null, "pipeline_tag": "text-generation", "processor": "AutoTokenizer" }
{"base_model": null, "datasets": null, "eval_results": null, "language": ["en", "ar", "zh", "fr", "de", "ja", "ko", "es"], "library_name": "transformers", "license": "other", "license_name": "lfm1.0", "license_link": "LICENSE", "metrics": null, "model_name": null, "pipeline_tag": "text-generation", "tags": ["liquid", "...
null
null
[ "other", "lfm1.0", "LICENSE" ]
null
[ "en", "ar", "zh", "fr", "de", "ja", "ko", "es" ]
8,339,929,856
null
null
[ "AutoModelForCausalLM", "lfm2_moe", "Lfm2MoeForCausalLM" ]
[ "text-generation" ]
[ "text" ]
[ "text" ]
[ "text" ]
68ea05ea8bfbf816c8e9ad2e
Qwen/Qwen3-VL-4B-Instruct
Qwen
null
2,023,717
5,797,256
False
2025-10-11T07:23:22Z
2025-10-15T16:15:55Z
transformers
361
5
null
image-text-to-text
{"parameters": {"BF16": 4437815808}, "total": 4437815808}
[ ".gitattributes", "README.md", "chat_template.json", "config.json", "generation_config.json", "merges.txt", "model-00001-of-00002.safetensors", "model-00002-of-00002.safetensors", "model.safetensors.index.json", "preprocessor_config.json", "tokenizer.json", "tokenizer_config.json", "video_pr...
ebb281ec70b05090aa6165b016eac8ec08e71b17
[ "transformers", "safetensors", "qwen3_vl", "image-text-to-text", "conversational", "arxiv:2505.09388", "arxiv:2502.13923", "arxiv:2409.12191", "arxiv:2308.12966", "license:apache-2.0", "endpoints_compatible", "deploy:azure", "region:us" ]
null
{"architectures": ["Qwen3VLForConditionalGeneration"], "model_type": "qwen3_vl", "processor_config": {"chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0].role == 'system' %}\n {%- if messages[0].content is string %}\n {{- messages[0].content }}\n {%- el...
{ "auto_model": "AutoModelForImageTextToText", "custom_class": null, "pipeline_tag": "image-text-to-text", "processor": "AutoProcessor" }
{"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "apache-2.0", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "image-text-to-text", "tags": null}
null
null
[ "apache-2.0" ]
null
null
4,437,815,808
null
null
[ "AutoModelForImageTextToText", "qwen3_vl", "Qwen3VLForConditionalGeneration" ]
[ "image-text-to-text" ]
[ "multimodal" ]
[ "text", "image" ]
[ "text" ]
68f6d55d42953c4ac0f8890f
datalab-to/chandra
datalab-to
null
164,256
1,120,608
False
2025-10-21T00:35:41Z
2026-03-26T15:55:54Z
transformers
510
5
null
image-text-to-text
{"parameters": {"BF16": 8767123696}, "total": 8767123696}
[ ".gitattributes", "LICENSE", "README.md", "added_tokens.json", "bench.png", "chat_template.jinja", "config.json", "generation_config.json", "handwritten_form.png", "merges.txt", "model-00001-of-00004.safetensors", "model-00002-of-00004.safetensors", "model-00003-of-00004.safetensors", "mod...
bd40c21576564d31ba99f204e28ccd5c4c1751fa
[ "transformers", "safetensors", "qwen3_vl", "image-text-to-text", "ocr", "vlm", "conversational", "license:openrail", "eval-results", "endpoints_compatible", "deploy:azure", "region:us" ]
null
{"architectures": ["Qwen3VLForConditionalGeneration"], "model_type": "qwen3_vl", "tokenizer_config": {"bos_token": null, "eos_token": "<|im_end|>", "pad_token": "<|endoftext|>", "unk_token": null}, "chat_template_jinja": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0].role == 'system' %}\n ...
{ "auto_model": "AutoModelForImageTextToText", "custom_class": null, "pipeline_tag": "image-text-to-text", "processor": "AutoProcessor" }
{"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "openrail", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": null, "tags": ["ocr", "vlm"], "new_version": "datalab-to/chandra-ocr-2"}
null
null
[ "openrail" ]
null
null
8,767,123,696
null
null
[ "AutoModelForImageTextToText", "qwen3_vl", "Qwen3VLForConditionalGeneration" ]
[ "image-text-to-text" ]
[ "multimodal" ]
[ "text", "image" ]
[ "text" ]
621ffdc136468d709f17adb7
facebook/bart-large-mnli
facebook
null
3,358,415
143,603,374
False
2022-03-02T23:29:05Z
2023-09-05T14:49:34Z
transformers
1,552
4
null
zero-shot-classification
{"parameters": {"F32": 407344133}, "total": 407344133}
[ ".gitattributes", "README.md", "config.json", "flax_model.msgpack", "merges.txt", "model.safetensors", "pytorch_model.bin", "rust_model.ot", "tokenizer.json", "tokenizer_config.json", "vocab.json" ]
d7645e127eaf1aefc7862fd59a17a5aa8558b8ce
[ "transformers", "pytorch", "jax", "rust", "safetensors", "bart", "text-classification", "zero-shot-classification", "dataset:multi_nli", "arxiv:1910.13461", "arxiv:1909.00161", "license:mit", "endpoints_compatible", "deploy:azure", "region:us" ]
null
{"architectures": ["BartForSequenceClassification"], "model_type": "bart", "tokenizer_config": {}}
{ "auto_model": "AutoModelForSequenceClassification", "custom_class": null, "pipeline_tag": "text-classification", "processor": "AutoTokenizer" }
{"base_model": null, "datasets": ["multi_nli"], "eval_results": null, "language": null, "library_name": null, "license": "mit", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "zero-shot-classification", "tags": null, "thumbnail": "https://huggingface.co/front/thumbnails...
# bart-large-mnli This is the checkpoint for [bart-large](https://huggingface.co/facebook/bart-large) after being trained on the [MultiNLI (MNLI)](https://huggingface.co/datasets/multi_nli) dataset. Additional information about this model: - The [bart-large](https://huggingface.co/facebook/bart-large) model page - [B...
null
[ "mit" ]
[ "multi_nli" ]
null
407,344,133
null
null
[ "BartForSequenceClassification", "bart", "AutoModelForSequenceClassification" ]
[ "zero-shot-classification", "text-classification" ]
[ "text" ]
[ "text" ]
[ "logits" ]
621ffdc136468d709f180297
sentence-transformers/all-mpnet-base-v2
sentence-transformers
null
29,268,568
1,158,149,103
False
2022-03-02T23:29:05Z
2025-08-19T10:14:25Z
sentence-transformers
1,262
4
null
sentence-similarity
{"parameters": {"I64": 514, "F32": 109486464}, "total": 109486978}
[ ".gitattributes", "1_Pooling/config.json", "README.md", "config.json", "config_sentence_transformers.json", "data_config.json", "model.safetensors", "modules.json", "onnx/model.onnx", "onnx/model_O1.onnx", "onnx/model_O2.onnx", "onnx/model_O3.onnx", "onnx/model_O4.onnx", "onnx/model_qint8_...
e8c3b32edf5434bc2275fc9bab85f82640a19130
[ "sentence-transformers", "pytorch", "onnx", "safetensors", "openvino", "mpnet", "fill-mask", "feature-extraction", "sentence-similarity", "transformers", "text-embeddings-inference", "en", "dataset:s2orc", "dataset:flax-sentence-embeddings/stackexchange_xml", "dataset:ms_marco", "datas...
null
{"architectures": ["MPNetForMaskedLM"], "model_type": "mpnet", "tokenizer_config": {"bos_token": "<s>", "eos_token": "</s>", "sep_token": "</s>", "cls_token": "<s>", "unk_token": "[UNK]", "pad_token": "<pad>", "mask_token": "<mask>"}}
{ "auto_model": "AutoModelForMaskedLM", "custom_class": null, "pipeline_tag": "fill-mask", "processor": "AutoTokenizer" }
{"base_model": null, "datasets": ["s2orc", "flax-sentence-embeddings/stackexchange_xml", "ms_marco", "gooaq", "yahoo_answers_topics", "code_search_net", "search_qa", "eli5", "snli", "multi_nli", "wikihow", "natural_questions", "trivia_qa", "embedding-data/sentence-compression", "embedding-data/flickr30k-captions", "emb...
# all-mpnet-base-v2 This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search. ## Usage (Sentence-Transformers) Using this model becomes easy when you have [sentence-transformers](h...
null
[ "apache-2.0" ]
[ "s2orc", "flax-sentence-embeddings/stackexchange_xml", "ms_marco", "gooaq", "yahoo_answers_topics", "code_search_net", "search_qa", "eli5", "snli", "multi_nli", "wikihow", "natural_questions", "trivia_qa", "embedding-data/sentence-compression", "embedding-data/flickr30k-captions", "emb...
[ "en" ]
109,486,978
null
null
[ "MPNetForMaskedLM", "AutoModelForMaskedLM", "mpnet" ]
[ "fill-mask", "sentence-similarity", "feature-extraction" ]
[ "text", "multimodal" ]
[ "text" ]
[ "logits", "embeddings" ]
63314bef9e3604f3f17e5b9e
openai/whisper-small
openai
null
1,779,882
118,833,474
False
2022-09-26T06:51:27Z
2024-02-29T10:57:38Z
transformers
544
4
[{"name": "whisper-small", "results": [{"task": {"name": "Automatic Speech Recognition", "type": "automatic-speech-recognition"}, "dataset": {"name": "LibriSpeech (clean)", "type": "librispeech_asr", "config": "clean", "split": "test", "args": {"language": "en"}}, "metrics": [{"name": "Test WER", "type": "wer", "value"...
automatic-speech-recognition
null
[ ".gitattributes", "README.md", "added_tokens.json", "config.json", "flax_model.msgpack", "generation_config.json", "merges.txt", "model.safetensors", "normalizer.json", "preprocessor_config.json", "pytorch_model.bin", "special_tokens_map.json", "tf_model.h5", "tokenizer.json", "tokenizer...
973afd24965f72e36ca33b3055d56a652f456b4d
[ "transformers", "pytorch", "tf", "jax", "safetensors", "whisper", "automatic-speech-recognition", "audio", "hf-asr-leaderboard", "en", "zh", "de", "es", "ru", "ko", "fr", "ja", "pt", "tr", "pl", "ca", "nl", "ar", "sv", "it", "id", "hi", "fi", "vi", "he", "...
null
{"architectures": ["WhisperForConditionalGeneration"], "model_type": "whisper", "tokenizer_config": {"bos_token": "<|endoftext|>", "eos_token": "<|endoftext|>", "pad_token": "<|endoftext|>", "unk_token": "<|endoftext|>"}}
{ "auto_model": "AutoModelForSpeechSeq2Seq", "custom_class": null, "pipeline_tag": "automatic-speech-recognition", "processor": "AutoProcessor" }
{"language": ["en", "zh", "de", "es", "ru", "ko", "fr", "ja", "pt", "tr", "pl", "ca", "nl", "ar", "sv", "it", "id", "hi", "fi", "vi", "he", "uk", "el", "ms", "cs", "ro", "da", "hu", "ta", "no", "th", "ur", "hr", "bg", "lt", "la", "mi", "ml", "cy", "sk", "te", "fa", "lv", "bn", "sr", "az", "sl", "kn", "et", "mk", "br", ...
# Whisper Whisper is a pre-trained model for automatic speech recognition (ASR) and speech translation. Trained on 680k hours of labelled data, Whisper models demonstrate a strong ability to generalise to many datasets and domains **without** the need for fine-tuning. Whisper was proposed in the paper [Robust Speec...
null
[ "apache-2.0" ]
null
[ "en", "zh", "de", "es", "ru", "ko", "fr", "ja", "pt", "tr", "pl", "ca", "nl", "ar", "sv", "it", "id", "hi", "fi", "vi", "he", "uk", "el", "ms", "cs", "ro", "da", "hu", "ta", "no", "th", "ur", "hr", "bg", "lt", "la", "mi", "ml", "cy", "sk"...
null
null
null
[ "WhisperForConditionalGeneration", "whisper", "AutoModelForSpeechSeq2Seq" ]
[ "automatic-speech-recognition" ]
[ "multimodal" ]
[ "audio" ]
[ "text" ]
65c5577b1080431ea9e083cd
google/gemma-7b
google
null
30,631
2,821,063
manual
2024-02-08T22:36:43Z
2024-06-27T14:09:40Z
transformers
3,290
4
null
text-generation
null
[ ".gitattributes", "README.md", "config.json", "examples/example_fsdp.py", "examples/example_sft_qlora.py", "examples/notebook_sft_peft.ipynb", "gemma-7b.gguf", "generation_config.json", "model-00001-of-00004.safetensors", "model-00002-of-00004.safetensors", "model-00003-of-00004.safetensors", ...
ff6768d9368919a1f025a54f9f5aa0ee591730bb
[ "transformers", "safetensors", "gguf", "gemma", "text-generation", "arxiv:2305.14314", "arxiv:2312.11805", "arxiv:2009.03300", "arxiv:1905.07830", "arxiv:1911.11641", "arxiv:1904.09728", "arxiv:1905.10044", "arxiv:1907.10641", "arxiv:1811.00937", "arxiv:1809.02789", "arxiv:1911.01547",...
{"total": 8538074112, "architecture": "gemma", "context_length": 8192, "bos_token": "<bos>", "eos_token": "<eos>"}
{"architectures": ["GemmaForCausalLM"], "model_type": "gemma", "tokenizer_config": {"bos_token": "<bos>", "eos_token": "<eos>", "pad_token": "<pad>", "unk_token": "<unk>", "use_default_system_prompt": false}}
{ "auto_model": "AutoModelForCausalLM", "custom_class": null, "pipeline_tag": "text-generation", "processor": "AutoTokenizer" }
{"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "gemma", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": null, "tags": null, "extra_gated_heading": "Access Gemma on Hugging Face", "extra_gated_prom...
null
null
[ "gemma" ]
null
null
null
8,538,074,112
null
[ "AutoModelForCausalLM", "gemma", "GemmaForCausalLM" ]
[ "text-generation" ]
[ "text" ]
[ "text" ]
[ "text" ]
65c7188324e59a0e4f89d863
nomic-ai/nomic-embed-text-v1.5
nomic-ai
null
10,410,034
60,578,496
False
2024-02-10T06:32:35Z
2025-07-21T17:44:14Z
sentence-transformers
785
4
[{"name": "epoch_0_model", "results": [{"task": {"type": "Classification"}, "dataset": {"type": "mteb/amazon_counterfactual", "name": "MTEB AmazonCounterfactualClassification (en)", "config": "en", "split": "test", "revision": "e8379541af4e31359cca9fbcf4b00f2671dba205"}, "metrics": [{"type": "accuracy", "value": 75.208...
sentence-similarity
{"parameters": {"F32": 136731648}, "total": 136731648}
[ ".gitattributes", "1_Pooling/config.json", "README.md", "config.json", "config_sentence_transformers.json", "model.safetensors", "modules.json", "onnx/model.onnx", "onnx/model_bnb4.onnx", "onnx/model_fp16.onnx", "onnx/model_int8.onnx", "onnx/model_q4.onnx", "onnx/model_q4f16.onnx", "onnx/m...
e5cf08aadaa33385f5990def41f7a23405aec398
[ "sentence-transformers", "onnx", "safetensors", "nomic_bert", "feature-extraction", "sentence-similarity", "mteb", "transformers", "transformers.js", "custom_code", "en", "arxiv:2402.01613", "arxiv:2205.13147", "license:apache-2.0", "model-index", "eval-results", "text-embeddings-inf...
null
{"architectures": ["NomicBertModel"], "auto_map": {"AutoConfig": "nomic-ai/nomic-bert-2048--configuration_hf_nomic_bert.NomicBertConfig", "AutoModel": "nomic-ai/nomic-bert-2048--modeling_hf_nomic_bert.NomicBertModel", "AutoModelForMaskedLM": "nomic-ai/nomic-bert-2048--modeling_hf_nomic_bert.NomicBertForPreTraining", "A...
{ "auto_model": "AutoModel", "custom_class": "nomic-ai/nomic-bert-2048--modeling_hf_nomic_bert.NomicBertModel", "pipeline_tag": "feature-extraction", "processor": null }
{"language": ["en"], "library_name": "sentence-transformers", "license": "apache-2.0", "pipeline_tag": "sentence-similarity", "tags": ["feature-extraction", "sentence-similarity", "mteb", "transformers", "transformers.js"], "model-index": [{"name": "epoch_0_model", "results": [{"task": {"type": "Classification"}, "data...
null
null
[ "apache-2.0" ]
null
[ "en" ]
136,731,648
null
null
[ "AutoModel", "NomicBertModel", "nomic_bert", "nomic-ai/nomic-bert-2048--modeling_hf_nomic_bert.NomicBertModel" ]
[ "sentence-similarity", "feature-extraction" ]
[ "text", "multimodal" ]
[ "text" ]
[ "logits", "embeddings" ]
65ee931c7443c0926729b6b5
sophosympatheia/Midnight-Miqu-70B-v1.5
sophosympatheia
{ "models": [ { "_id": "65e3b8a6607c2f5380421ffb", "id": "migtissera/Tess-70B-v1.6" }, { "_id": "65e012a1ebc538ecfac4cd1c", "id": "sophosympatheia/Midnight-Miqu-70B-v1.0" } ], "relation": "merge" }
11,205
142,689
False
2024-03-11T05:14:04Z
2024-12-10T16:50:16Z
transformers
252
4
[{"name": "Midnight-Miqu-70B-v1.5", "results": [{"task": {"type": "text-generation", "name": "Text Generation"}, "dataset": {"name": "IFEval (0-Shot)", "type": "HuggingFaceH4/ifeval", "args": {"num_few_shot": 0}}, "metrics": [{"type": "inst_level_strict_acc and prompt_level_strict_acc", "value": 61.18, "name": "strict ...
text-generation
{"parameters": {"F16": 68976648192}, "total": 68976648192}
[ ".gitattributes", "LICENSE", "README.md", "config.json", "mergekit_config.yml", "model-00001-of-00015.safetensors", "model-00002-of-00015.safetensors", "model-00003-of-00015.safetensors", "model-00004-of-00015.safetensors", "model-00005-of-00015.safetensors", "model-00006-of-00015.safetensors", ...
74888979a3734f0d7466d77dbf32120f575000e3
[ "transformers", "safetensors", "llama", "text-generation", "mergekit", "merge", "conversational", "arxiv:2311.03099", "base_model:migtissera/Tess-70B-v1.6", "base_model:merge:migtissera/Tess-70B-v1.6", "base_model:sophosympatheia/Midnight-Miqu-70B-v1.0", "base_model:merge:sophosympatheia/Midni...
null
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}...
{ "auto_model": "AutoModelForCausalLM", "custom_class": null, "pipeline_tag": "text-generation", "processor": "AutoTokenizer" }
{"base_model": ["sophosympatheia/Midnight-Miqu-70B-v1.0", "migtissera/Tess-70B-v1.6"], "library_name": "transformers", "license": "other", "tags": ["mergekit", "merge"], "model-index": [{"name": "Midnight-Miqu-70B-v1.5", "results": [{"task": {"type": "text-generation", "name": "Text Generation"}, "dataset": {"name": "I...
<div style="width: auto; margin-left: auto; margin-right: auto"> <img src="https://i.imgur.com/Tn9MBg6.png" alt="MidnightMiqu" style="width: 100%; min-width: 400px; display: block; margin: auto;"> </div> ### Overview Looking for the 103B version? You can get it from [FluffyKaeloky/Midnight-Miqu-103B-v1.5](https://hug...
null
[ "other" ]
null
null
68,976,648,192
null
null
[ "AutoModelForCausalLM", "llama", "LlamaForCausalLM" ]
[ "text-generation" ]
[ "text" ]
[ "text" ]
[ "text" ]
66969ad27a033bf62173f3e2
meta-llama/Llama-3.1-70B-Instruct
meta-llama
{ "models": [ { "_id": "66944fd095c7fa6e68c314ae", "id": "meta-llama/Llama-3.1-70B" } ], "relation": "finetune" }
1,036,359
20,229,486
manual
2024-07-16T16:07:46Z
2024-12-15T01:55:33Z
transformers
902
4
null
text-generation
{"parameters": {"BF16": 70553706496}, "total": 70553706496}
[ ".gitattributes", "LICENSE", "README.md", "USE_POLICY.md", "config.json", "generation_config.json", "model-00001-of-00030.safetensors", "model-00002-of-00030.safetensors", "model-00003-of-00030.safetensors", "model-00004-of-00030.safetensors", "model-00005-of-00030.safetensors", "model-00006-o...
1605565b47bb9346c5515c34102e054115b4f98b
[ "transformers", "safetensors", "llama", "text-generation", "facebook", "meta", "pytorch", "llama-3", "conversational", "en", "de", "fr", "it", "pt", "hi", "es", "th", "arxiv:2204.05149", "base_model:meta-llama/Llama-3.1-70B", "base_model:finetune:meta-llama/Llama-3.1-70B", "l...
null
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{{- bos_token }}\n{%- if custom_tools is defined %}\n {%- set tools = custom_tools %}\n{%- endif %}\n{%- if not tools_in_user_message is defined %}\n {%- set tools_in_user_messag...
{ "auto_model": "AutoModelForCausalLM", "custom_class": null, "pipeline_tag": "text-generation", "processor": "AutoTokenizer" }
{"base_model": "meta-llama/Meta-Llama-3.1-70B", "datasets": null, "eval_results": null, "language": ["en", "de", "fr", "it", "pt", "hi", "es", "th"], "library_name": "transformers", "license": "llama3.1", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "text-generation",...
null
null
[ "llama3.1" ]
null
[ "en", "de", "fr", "it", "pt", "hi", "es", "th" ]
70,553,706,496
null
null
[ "AutoModelForCausalLM", "llama", "LlamaForCausalLM" ]
[ "text-generation" ]
[ "text" ]
[ "text" ]
[ "text" ]
66b67e73621b62a4c02c1541
Orenguteng/Llama-3.1-8B-Lexi-Uncensored-V2
Orenguteng
null
3,873
369,961
False
2024-08-09T20:39:15Z
2024-09-25T02:49:53Z
transformers
281
4
[{"name": "Llama-3.1-8B-Lexi-Uncensored-V2", "results": [{"task": {"type": "text-generation", "name": "Text Generation"}, "dataset": {"name": "IFEval (0-Shot)", "type": "HuggingFaceH4/ifeval", "args": {"num_few_shot": 0}}, "metrics": [{"type": "inst_level_strict_acc and prompt_level_strict_acc", "value": 77.92, "name":...
text-generation
{"parameters": {"BF16": 8030261248}, "total": 8030261248}
[ ".gitattributes", "README.md", "config.json", "generation_config.json", "model-00001-of-00004.safetensors", "model-00002-of-00004.safetensors", "model-00003-of-00004.safetensors", "model-00004-of-00004.safetensors", "model.safetensors.index.json", "special_tokens_map.json", "tokenizer.json", "...
f4617caeabd21f1820ac89bd125c80eda70901a7
[ "transformers", "safetensors", "llama", "text-generation", "conversational", "license:llama3.1", "model-index", "text-generation-inference", "endpoints_compatible", "deploy:azure", "region:us" ]
null
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{{- bos_token }}\n{%- if custom_tools is defined %}\n {%- set tools = custom_tools %}\n{%- endif %}\n{%- if not tools_in_user_message is defined %}\n {%- set tools_in_user_messag...
{ "auto_model": "AutoModelForCausalLM", "custom_class": null, "pipeline_tag": "text-generation", "processor": "AutoTokenizer" }
{"library_name": "transformers", "license": "llama3.1", "model-index": [{"name": "Llama-3.1-8B-Lexi-Uncensored-V2", "results": [{"task": {"type": "text-generation", "name": "Text Generation"}, "dataset": {"name": "IFEval (0-Shot)", "type": "HuggingFaceH4/ifeval", "args": {"num_few_shot": 0}}, "metrics": [{"type": "inst...
![image/png](https://cdn-uploads.huggingface.co/production/uploads/644ad182f434a6a63b18eee6/7mnEJyioRzQaWz8xLM4KI.png) VERSION 2 Update Notes: --- - More compliant - Smarter - For best response, use this system prompt (feel free to expand upon it as you wish): Think step by step with a logical reasoning and intellect...
null
[ "llama3.1" ]
null
null
8,030,261,248
null
null
[ "AutoModelForCausalLM", "llama", "LlamaForCausalLM" ]
[ "text-generation" ]
[ "text" ]
[ "text" ]
[ "text" ]
67d7262eaf109e28c7ddcc49
mlabonne/gemma-3-27b-it-abliterated
mlabonne
{ "models": [ { "_id": "67c35b9bb236f0d365bf29d3", "id": "google/gemma-3-27b-it" } ], "relation": "finetune" }
4,371
275,755
False
2025-03-16T19:27:42Z
2025-03-21T16:10:45Z
transformers
305
4
null
image-text-to-text
null
[ ".gitattributes", "README.md", "added_tokens.json", "chat_template.json", "config.json", "generation_config.json", "model-00001-of-00012.safetensors", "model-00002-of-00012.safetensors", "model-00003-of-00012.safetensors", "model-00004-of-00012.safetensors", "model-00005-of-00012.safetensors", ...
eaa815dffdf0ea808a0421e4a48891bf52496a16
[ "transformers", "safetensors", "gemma3", "image-text-to-text", "conversational", "base_model:google/gemma-3-27b-it", "base_model:finetune:google/gemma-3-27b-it", "license:gemma", "text-generation-inference", "endpoints_compatible", "region:us" ]
null
{"architectures": ["Gemma3ForConditionalGeneration"], "model_type": "gemma3", "processor_config": {"chat_template": "{{ bos_token }}\n{%- if messages[0]['role'] == 'system' -%}\n {%- if messages[0]['content'] is string -%}\n {%- set first_user_prefix = messages[0]['content'] + '\n\n' -%}\n {%- else -%}\n ...
{ "auto_model": "AutoModelForImageTextToText", "custom_class": null, "pipeline_tag": "image-text-to-text", "processor": "AutoProcessor" }
{"base_model": "google/gemma-3-27b-it", "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "gemma", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "image-text-to-text", "tags": null}
# 💎 Gemma 3 27B IT Abliterated ![image/png](https://cdn-uploads.huggingface.co/production/uploads/61b8e2ba285851687028d395/WjFfc8hhj20r5XK07Yny9.png) <center><a href="https://huggingface.co/mlabonne/gemma-3-1b-it-abliterated">Gemma 3 1B Abliterated</a> • <a href="https://huggingface.co/mlabonne/gemma-3-4b-it-ablitera...
null
[ "gemma" ]
null
null
null
null
null
[ "AutoModelForImageTextToText", "Gemma3ForConditionalGeneration", "gemma3" ]
[ "image-text-to-text" ]
[ "multimodal" ]
[ "text", "image" ]
[ "text" ]
67ff7b22b5ba13b43f0c3811
ByteDance-Seed/UI-TARS-1.5-7B
ByteDance-Seed
null
150,091
1,602,459
False
2025-04-16T09:40:50Z
2025-04-18T01:35:38Z
transformers
533
4
null
image-text-to-text
{"parameters": {"F32": 8292166656}, "total": 8292166656}
[ ".gitattributes", "README.md", "added_tokens.json", "chat_template.json", "config.json", "merges.txt", "model-00001-of-00007.safetensors", "model-00002-of-00007.safetensors", "model-00003-of-00007.safetensors", "model-00004-of-00007.safetensors", "model-00005-of-00007.safetensors", "model-0000...
683d002dd99d8f95104d31e70391a39348857f4e
[ "transformers", "safetensors", "qwen2_5_vl", "image-text-to-text", "multimodal", "gui", "conversational", "en", "arxiv:2501.12326", "arxiv:2404.07972", "arxiv:2409.08264", "arxiv:2401.13919", "arxiv:2504.01382", "arxiv:2405.14573", "arxiv:2410.23218", "arxiv:2504.07981", "license:apa...
null
{"architectures": ["Qwen2_5_VLForConditionalGeneration"], "model_type": "qwen2_5_vl", "processor_config": {"chat_template": "{% set image_count = namespace(value=0) %}{% set video_count = namespace(value=0) %}{% for message in messages %}{% if loop.first and message['role'] != 'system' %}<|im_start|>system\nYou are a h...
{ "auto_model": "AutoModelForImageTextToText", "custom_class": null, "pipeline_tag": "image-text-to-text", "processor": "AutoProcessor" }
{"base_model": null, "datasets": null, "eval_results": null, "language": ["en"], "library_name": "transformers", "license": "apache-2.0", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "image-text-to-text", "tags": ["multimodal", "gui"]}
null
null
[ "apache-2.0" ]
null
[ "en" ]
8,292,166,656
null
null
[ "qwen2_5_vl", "AutoModelForImageTextToText", "Qwen2_5_VLForConditionalGeneration" ]
[ "image-text-to-text" ]
[ "multimodal" ]
[ "text", "image" ]
[ "text" ]
681163a79c2765c932331a55
XiaomiMiMo/MiMo-7B-Base
XiaomiMiMo
null
116,314
554,955
False
2025-04-29T23:41:27Z
2025-06-05T16:06:42Z
transformers
129
4
null
text-generation
{"parameters": {"BF16": 7833409536}, "total": 7833409536}
[ ".gitattributes", "README.md", "config.json", "configuration_mimo.py", "generation_config.json", "merges.txt", "model-00001-of-00004.safetensors", "model-00002-of-00004.safetensors", "model-00003-of-00004.safetensors", "model-00004-of-00004.safetensors", "model.safetensors.index.json", "modeli...
c72df4586cb8bdeebd65f36929cd3385a6566fbe
[ "transformers", "safetensors", "mimo", "text-generation", "conversational", "custom_code", "arxiv:2505.07608", "license:mit", "region:us" ]
null
{"architectures": ["MiMoForCausalLM"], "auto_map": {"AutoConfig": "configuration_mimo.MiMoConfig", "AutoModel": "modeling_mimo.MiMoModel", "AutoModelForCausalLM": "modeling_mimo.MiMoForCausalLM"}, "model_type": "mimo", "tokenizer_config": {"bos_token": null, "chat_template": "{%- if tools %}\n {{- '<|im_start|>syste...
{ "auto_model": "AutoModelForCausalLM", "custom_class": "modeling_mimo.MiMoForCausalLM", "pipeline_tag": "text-generation", "processor": null }
{"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "mit", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": null, "tags": null}
<div align="center"> <picture> <source srcset="https://github.com/XiaomiMiMo/MiMo/raw/main/figures/Xiaomi_MiMo_darkmode.png?raw=true" media="(prefers-color-scheme: dark)"> <img src="https://github.com/XiaomiMiMo/MiMo/raw/main/figures/Xiaomi_MiMo.png?raw=true" width="60%" alt="Xiaomi-MiMo" /> </picture> </di...
null
[ "mit" ]
null
null
7,833,409,536
null
null
[ "AutoModelForCausalLM", "modeling_mimo.MiMoForCausalLM", "mimo", "MiMoForCausalLM" ]
[ "text-generation" ]
[ "text" ]
[ "text" ]
[ "text" ]
681c309e234962a917f42f50
PocketDoc/Dans-PersonalityEngine-V1.3.0-24b
PocketDoc
{ "models": [ { "_id": "67d75f3fd56afd880495e9b6", "id": "mistralai/Mistral-Small-3.1-24B-Base-2503" } ], "relation": "finetune" }
181
91,747
False
2025-05-08T04:18:38Z
2025-05-23T07:07:01Z
transformers
139
4
null
text-generation
null
[ ".gitattributes", "README.md", "config.json", "generation_config.json", "model-00001-of-00010.safetensors", "model-00002-of-00010.safetensors", "model-00003-of-00010.safetensors", "model-00004-of-00010.safetensors", "model-00005-of-00010.safetensors", "model-00006-of-00010.safetensors", "model-0...
bb460a79d3d24f546106e8f738e9427f88d9dc40
[ "transformers", "safetensors", "mistral", "text-generation", "general-purpose", "roleplay", "storywriting", "chemistry", "biology", "code", "climate", "axolotl", "text-generation-inference", "finetune", "legal", "medical", "finance", "conversational", "en", "ar", "de", "fr"...
null
{"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "[gMASK]<sop>", "chat_template": "{{ bos_token }}{%- set loop_messages = messages %}{%- for message in loop_messages %}{%- if message['role'] == 'user' %}{%- set role_token = '<|user|>' %}{%- elif message['role'] == 'as...
{ "auto_model": "AutoModelForCausalLM", "custom_class": null, "pipeline_tag": "text-generation", "processor": "AutoTokenizer" }
{"base_model": ["mistralai/Mistral-Small-3.1-24B-Base-2503"], "datasets": ["PocketDoc/Dans-Prosemaxx-RP", "PocketDoc/Dans-Personamaxx-Logs-2", "PocketDoc/Dans-Personamaxx-VN", "PocketDoc/Dans-Kinomaxx-VanillaBackrooms", "PocketDoc/Dans-Prosemaxx-Gutenberg", "PocketDoc/Dans-Prosemaxx-Cowriter-3-XL", "PocketDoc/Dans-Pros...
null
null
[ "apache-2.0" ]
[ "PocketDoc/Dans-Prosemaxx-RP", "PocketDoc/Dans-Personamaxx-Logs-2", "PocketDoc/Dans-Personamaxx-VN", "PocketDoc/Dans-Kinomaxx-VanillaBackrooms", "PocketDoc/Dans-Prosemaxx-Gutenberg", "PocketDoc/Dans-Prosemaxx-Cowriter-3-XL", "PocketDoc/Dans-Prosemaxx-Adventure", "PocketDoc/Dans-Failuremaxx-Adventure-3...
[ "en", "ar", "de", "fr", "es", "hi", "pt", "ja", "ko" ]
null
null
null
[ "AutoModelForCausalLM", "mistral", "MistralForCausalLM" ]
[ "text-generation" ]
[ "text" ]
[ "text" ]
[ "text" ]
6848029d9a16b13edd1a3de4
nanonets/Nanonets-OCR-s
nanonets
{ "models": [ { "_id": "6795ff8f7534713f94918fc5", "id": "Qwen/Qwen2.5-VL-3B-Instruct" } ], "relation": "finetune" }
64,163
1,355,238
False
2025-06-10T10:02:05Z
2025-06-20T19:15:33Z
transformers
1,589
4
null
image-text-to-text
{"parameters": {"BF16": 3754622976}, "total": 3754622976}
[ ".gitattributes", "Modelfile", "README.md", "added_tokens.json", "chat_template.jinja", "config.json", "generation_config.json", "merges.txt", "model-00001-of-00002.safetensors", "model-00002-of-00002.safetensors", "model.safetensors.index.json", "preprocessor_config.json", "special_tokens_m...
3baad182cc87c65a1861f0c30357d3467e978172
[ "transformers", "safetensors", "qwen2_5_vl", "image-text-to-text", "OCR", "pdf2markdown", "conversational", "en", "base_model:Qwen/Qwen2.5-VL-3B-Instruct", "base_model:finetune:Qwen/Qwen2.5-VL-3B-Instruct", "text-generation-inference", "endpoints_compatible", "region:us" ]
null
{"architectures": ["Qwen2_5_VLForConditionalGeneration"], "model_type": "qwen2_5_vl", "tokenizer_config": {"bos_token": null, "eos_token": "<|im_end|>", "pad_token": "<|endoftext|>", "unk_token": null}, "chat_template_jinja": "{% set image_count = namespace(value=0) %}{% set video_count = namespace(value=0) %}{% for me...
{ "auto_model": "AutoModelForImageTextToText", "custom_class": null, "pipeline_tag": "image-text-to-text", "processor": "AutoProcessor" }
{"base_model": ["Qwen/Qwen2.5-VL-3B-Instruct"], "datasets": null, "eval_results": null, "language": ["en"], "library_name": "transformers", "license": null, "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "image-text-to-text", "tags": ["OCR", "pdf2markdown"]}
Nanonets-OCR-s by [Nanonets](https://nanonets.com) is a powerful, state-of-the-art image-to-markdown OCR model that goes far beyond traditional text extraction. It transforms documents into structured markdown with intelligent content recognition and semantic tagging, making it ideal for downstream processing by Large ...
null
null
null
[ "en" ]
3,754,622,976
null
null
[ "qwen2_5_vl", "AutoModelForImageTextToText", "Qwen2_5_VLForConditionalGeneration" ]
[ "image-text-to-text" ]
[ "multimodal" ]
[ "text", "image" ]
[ "text" ]
687c61c324649ecb26a748f0
zai-org/GLM-4.5-Air
zai-org
null
406,257
2,723,104
False
2025-07-20T03:25:55Z
2025-08-11T13:25:37Z
transformers
592
4
null
text-generation
{"parameters": {"BF16": 110468818944, "F32": 5888}, "total": 110468824832}
[ ".gitattributes", "README.md", "chat_template.jinja", "config.json", "generation_config.json", "model-00001-of-00047.safetensors", "model-00002-of-00047.safetensors", "model-00003-of-00047.safetensors", "model-00004-of-00047.safetensors", "model-00005-of-00047.safetensors", "model-00006-of-00047...
a24ceef6ce4f3536971efe9b778bdaa1bab18daa
[ "transformers", "safetensors", "glm4_moe", "text-generation", "conversational", "en", "zh", "arxiv:2508.06471", "license:mit", "endpoints_compatible", "deploy:azure", "region:us" ]
null
{"architectures": ["Glm4MoeForCausalLM"], "model_type": "glm4_moe", "tokenizer_config": {"eos_token": "<|endoftext|>", "pad_token": "<|endoftext|>"}, "chat_template_jinja": "[gMASK]<sop>\n{%- if tools -%}\n<|system|>\n# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with ...
{ "auto_model": "AutoModelForCausalLM", "custom_class": null, "pipeline_tag": "text-generation", "processor": "AutoTokenizer" }
{"base_model": null, "datasets": null, "eval_results": null, "language": ["en", "zh"], "library_name": "transformers", "license": "mit", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "text-generation", "tags": null}
# GLM-4.5-Air <div align="center"> <img src=https://raw.githubusercontent.com/zai-org/GLM-4.5/refs/heads/main/resources/logo.svg width="15%"/> </div> <p align="center"> 👋 Join our <a href="https://discord.gg/QR7SARHRxK" target="_blank">Discord</a> community. <br> 📖 Check out the GLM-4.5 <a href="https://...
null
[ "mit" ]
null
[ "en", "zh" ]
110,468,824,832
null
null
[ "AutoModelForCausalLM", "Glm4MoeForCausalLM", "glm4_moe" ]
[ "text-generation" ]
[ "text" ]
[ "text" ]
[ "text" ]
6889ec2029577a514ed82d21
rednote-hilab/dots.ocr
rednote-hilab
null
257,194
4,511,440
False
2025-07-30T09:55:44Z
2025-10-31T08:49:31Z
dots_ocr
1,281
4
null
image-text-to-text
{"parameters": {"BF16": 3039179264}, "total": 3039179264}
[ ".gitattributes", "NOTICE", "README.md", "chat_template.json", "config.json", "configuration_dots.py", "dots.ocr LICENSE AGREEMENT", "generation_config.json", "merges.txt", "model-00001-of-00002.safetensors", "model-00002-of-00002.safetensors", "model.safetensors.index.json", "modeling_dots_...
c0111ce6bc07803dbc267932ffef0ae3a51dc951
[ "dots_ocr", "safetensors", "text-generation", "image-to-text", "ocr", "document-parse", "layout", "table", "formula", "transformers", "custom_code", "image-text-to-text", "conversational", "en", "zh", "multilingual", "license:mit", "eval-results", "region:us" ]
null
{"architectures": ["DotsOCRForCausalLM"], "model_type": "dots_ocr", "auto_map": {"AutoConfig": "configuration_dots.DotsOCRConfig", "AutoModelForCausalLM": "modeling_dots_ocr.DotsOCRForCausalLM"}, "processor_config": {"chat_template": "{% set image_count = namespace(value=0) %}{% set video_count = namespace(value=0) %}{...
{ "auto_model": "AutoModelForCausalLM", "custom_class": "modeling_dots_ocr.DotsOCRForCausalLM", "pipeline_tag": "text-generation", "processor": null }
{"base_model": null, "datasets": null, "eval_results": null, "language": ["en", "zh", "multilingual"], "library_name": "dots_ocr", "license": "mit", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "image-text-to-text", "tags": ["image-to-text", "ocr", "document-parse", "...
<div align="center"> <p align="center"> <img src="https://raw.githubusercontent.com/rednote-hilab/dots.ocr/master/assets/logo.png" width="300"/> <p> <h1 align="center"> dots.ocr: Multilingual Document Layout Parsing in a Single Vision-Language Model </h1> [![Blog](https://img.shields.io/badge/Blog-View_on_GitHub...
null
[ "mit" ]
null
[ "en", "zh", "multilingual" ]
3,039,179,264
null
null
[ "AutoModelForCausalLM", "DotsOCRForCausalLM", "dots_ocr", "modeling_dots_ocr.DotsOCRForCausalLM" ]
[ "image-text-to-text", "image-to-text", "text-generation" ]
[ "text", "multimodal" ]
[ "text", "image" ]
[ "text" ]
6891e4c34b616cd1bb82f50e
Qwen/Qwen3-4B-Thinking-2507
Qwen
null
1,214,957
4,177,619
False
2025-08-05T11:02:27Z
2025-08-06T11:08:25Z
transformers
573
4
null
text-generation
{"parameters": {"BF16": 4022468096}, "total": 4022468096}
[ ".gitattributes", "LICENSE", "README.md", "config.json", "generation_config.json", "merges.txt", "model-00001-of-00003.safetensors", "model-00002-of-00003.safetensors", "model-00003-of-00003.safetensors", "model.safetensors.index.json", "tokenizer.json", "tokenizer_config.json", "vocab.json"...
768f209d9ea81521153ed38c47d515654e938aea
[ "transformers", "safetensors", "qwen3", "text-generation", "conversational", "arxiv:2505.09388", "license:apache-2.0", "eval-results", "text-generation-inference", "endpoints_compatible", "deploy:azure", "region:us" ]
null
{"architectures": ["Qwen3ForCausalLM"], "model_type": "qwen3", "tokenizer_config": {"bos_token": null, "chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0].role == 'system' %}\n {{- messages[0].content + '\\n\\n' }}\n {%- endif %}\n {{- \"# Tools\\n\\nYou may call o...
{ "auto_model": "AutoModelForCausalLM", "custom_class": null, "pipeline_tag": "text-generation", "processor": "AutoTokenizer" }
{"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "apache-2.0", "license_name": null, "license_link": "https://huggingface.co/Qwen/Qwen3-4B-Thinking-2507/blob/main/LICENSE", "metrics": null, "model_name": null, "pipeline_tag": "text-generation", "t...
# Qwen3-4B-Thinking-2507 <a href="https://chat.qwen.ai/" target="_blank" style="margin: 2px;"> <img alt="Chat" src="https://img.shields.io/badge/%F0%9F%92%9C%EF%B8%8F%20Qwen%20Chat%20-536af5" style="display: inline-block; vertical-align: middle;"/> </a> ## Highlights Over the past three months, we have continued ...
null
[ "apache-2.0", "https://huggingface.co/Qwen/Qwen3-4B-Thinking-2507/blob/main/LICENSE" ]
null
null
4,022,468,096
null
null
[ "AutoModelForCausalLM", "Qwen3ForCausalLM", "qwen3" ]
[ "text-generation" ]
[ "text" ]
[ "text" ]
[ "text" ]
68a34925f057ffe5051b5b6b
NousResearch/Hermes-4-70B
NousResearch
{ "models": [ { "_id": "66944fd095c7fa6e68c314ae", "id": "meta-llama/Llama-3.1-70B" } ], "relation": "finetune" }
707
39,998
False
2025-08-18T15:39:17Z
2025-09-02T06:40:36Z
transformers
179
4
[{"name": "Hermes-4-Llama-3.1-70B", "results": []}]
text-generation
null
[ ".gitattributes", "README.md", "chat_template.jinja", "config.json", "generation_config.json", "model-00001-of-00030.safetensors", "model-00002-of-00030.safetensors", "model-00003-of-00030.safetensors", "model-00004-of-00030.safetensors", "model-00005-of-00030.safetensors", "model-00006-of-00030...
d5dec2bd6b3930a09ddefd0b7fc6523fe0720d09
[ "transformers", "safetensors", "llama", "text-generation", "Llama-3.1", "instruct", "finetune", "reasoning", "hybrid-mode", "chatml", "function calling", "tool use", "json mode", "structured outputs", "atropos", "dataforge", "long context", "roleplaying", "chat", "conversationa...
null
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "eos_token": "<|eot_id|>", "pad_token": "<|end_of_text|>"}, "chat_template_jinja": "{%- set thinking_prompt = 'You are a deep thinking AI, you may use extremely long chains of thought to deeply consider...
{ "auto_model": "AutoModelForCausalLM", "custom_class": null, "pipeline_tag": "text-generation", "processor": "AutoTokenizer" }
{"base_model": "meta-llama/Meta-Llama-3.1-70B", "datasets": null, "eval_results": [], "language": ["en"], "library_name": "transformers", "license": "llama3", "license_name": null, "license_link": null, "metrics": null, "model_name": "Hermes-4-Llama-3.1-70B", "pipeline_tag": "text-generation", "tags": ["Llama-3.1", "in...
# Hermes 4 — Llama-3.1 70B ![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/6317aade83d8d2fd903192d9/roT9o5bMYBtQziRMlaSDf.jpeg) ## Model Description Hermes 4 70B is a frontier, hybrid-mode **reasoning** model based on Llama-3.1-70B by Nous Research that is aligned to **you**. Read the Hermes 4 t...
null
[ "llama3" ]
null
[ "en" ]
null
null
null
[ "AutoModelForCausalLM", "llama", "LlamaForCausalLM" ]
[ "text-generation" ]
[ "text" ]
[ "text" ]
[ "text" ]
68c04b9ba4d1823e9b0ee2bd
Qwen/Qwen3-Next-80B-A3B-Thinking
Qwen
null
50,718
2,282,134
False
2025-09-09T15:45:31Z
2025-09-15T09:12:48Z
transformers
486
4
null
text-generation
null
[ ".gitattributes", "LICENSE", "README.md", "config.json", "generation_config.json", "merges.txt", "model-00001-of-00041.safetensors", "model-00002-of-00041.safetensors", "model-00003-of-00041.safetensors", "model-00004-of-00041.safetensors", "model-00005-of-00041.safetensors", "model-00006-of-0...
e502dd4100cc68c0de57643fd4317ec93a128670
[ "transformers", "safetensors", "qwen3_next", "text-generation", "conversational", "arxiv:2309.00071", "arxiv:2505.09388", "arxiv:2501.15383", "license:apache-2.0", "eval-results", "endpoints_compatible", "deploy:azure", "region:us" ]
null
{"architectures": ["Qwen3NextForCausalLM"], "model_type": "qwen3_next", "tokenizer_config": {"bos_token": null, "chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0].role == 'system' %}\n {{- messages[0].content + '\\n\\n' }}\n {%- endif %}\n {{- \"# Tools\\n\\nYou m...
{ "auto_model": "AutoModelForCausalLM", "custom_class": null, "pipeline_tag": "text-generation", "processor": "AutoTokenizer" }
{"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "apache-2.0", "license_name": null, "license_link": "https://huggingface.co/Qwen/Qwen3-Next-80B-A3B-Thinking/blob/main/LICENSE", "metrics": null, "model_name": null, "pipeline_tag": "text-generation...
# Qwen3-Next-80B-A3B-Thinking <a href="https://chat.qwen.ai/" target="_blank" style="margin: 2px;"> <img alt="Chat" src="https://img.shields.io/badge/%F0%9F%92%9C%EF%B8%8F%20Qwen%20Chat%20-536af5" style="display: inline-block; vertical-align: middle;"/> </a> Over the past few months, we have observed increasingly ...
null
[ "apache-2.0", "https://huggingface.co/Qwen/Qwen3-Next-80B-A3B-Thinking/blob/main/LICENSE" ]
null
null
null
null
null
[ "qwen3_next", "Qwen3NextForCausalLM", "AutoModelForCausalLM" ]
[ "text-generation" ]
[ "text" ]
[ "text" ]
[ "text" ]
68d4e10c40e8730ee32cd0c6
tencent/HunyuanImage-3.0
tencent
null
115,866
1,064,280
False
2025-09-25T06:28:28Z
2026-01-28T08:14:42Z
transformers
656
4
null
text-to-image
null
[ ".gitattributes", "Hunyuan-Image3.md", "LICENSE", "README.md", "README_zh_CN.md", "__init__.py", "assets/.DS_Store", "assets/WECHAT.md", "assets/banner.png", "assets/banner_all.jpg", "assets/demo_instruct_imgs/input_0_0.png", "assets/demo_instruct_imgs/input_1_0.png", "assets/demo_instruct_i...
36f21fe74b65614451cc50ffd8a35a5f662dac70
[ "transformers", "safetensors", "hunyuan_image_3_moe", "text-generation", "text-to-image", "custom_code", "arxiv:2509.23951", "license:other", "region:us" ]
null
{"architectures": ["HunyuanImage3ForCausalMM"], "auto_map": {"AutoConfig": "configuration_hunyuan.HunyuanImage3Config", "AutoModel": "hunyuan.HunyuanImage3Model", "AutoModelForCausalLM": "hunyuan.HunyuanImage3ForCausalMM"}, "model_type": "hunyuan_image_3_moe", "tokenizer_config": {"bos_token": "<|startoftext|>", "eos_t...
{ "auto_model": "AutoModelForCausalLM", "custom_class": "hunyuan.HunyuanImage3ForCausalMM", "pipeline_tag": "text-generation", "processor": null }
{"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "other", "license_name": "tencent-hunyuan-community", "license_link": "LICENSE", "metrics": null, "model_name": null, "pipeline_tag": "text-to-image", "tags": null}
null
null
[ "other", "tencent-hunyuan-community", "LICENSE" ]
null
null
null
null
null
[ "AutoModelForCausalLM", "HunyuanImage3ForCausalMM", "hunyuan.HunyuanImage3ForCausalMM", "hunyuan_image_3_moe" ]
[ "text-to-image", "text-generation" ]
[ "text", "vision" ]
[ "text" ]
[ "text", "image" ]
68f17551daf20cc0cd8d9c40
nvidia/llama-nemotron-embed-1b-v2
nvidia
null
58,549
152,621
False
2025-10-16T22:44:33Z
2026-03-11T15:56:39Z
sentence-transformers
49
4
null
feature-extraction
{"parameters": {"BF16": 1235814400}, "total": 1235814400}
[ ".gitattributes", "1_Pooling/config.json", "LICENSE", "README.md", "config.json", "config_sentence_transformers.json", "llama_bidirectional_model.py", "model.safetensors", "modules.json", "pooling.py", "pytorch_model.bin", "sentence_bert_config.json", "special_tokens_map.json", "tokenizer....
5ee12b8f82486746155a9db12cacea9bc4b9b8ad
[ "sentence-transformers", "pytorch", "safetensors", "llama_bidirec", "feature-extraction", "text", "text-embeddings", "retrieval", "semantic-search", "transformers", "custom_code", "multilingual", "arxiv:2407.15831", "license:other", "text-embeddings-inference", "endpoints_compatible", ...
null
{"architectures": ["LlamaBidirectionalModel"], "auto_map": {"AutoConfig": "llama_bidirectional_model.LlamaBidirectionalConfig", "AutoModel": "llama_bidirectional_model.LlamaBidirectionalModel"}, "model_type": "llama_bidirec", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "eos_token": "<|end_of_text|>", "pad_to...
{ "auto_model": "AutoModel", "custom_class": "llama_bidirectional_model.LlamaBidirectionalModel", "pipeline_tag": "feature-extraction", "processor": null }
{"base_model": null, "datasets": null, "eval_results": null, "language": ["multilingual"], "library_name": "sentence-transformers", "license": "other", "license_name": "nvidia-open-model-license", "license_link": "https://www.nvidia.com/en-us/agreements/enterprise-software/nvidia-open-model-license/", "metrics": null, ...
null
null
[ "other", "nvidia-open-model-license", "https://www.nvidia.com/en-us/agreements/enterprise-software/nvidia-open-model-license/" ]
null
[ "multilingual" ]
1,235,814,400
null
null
[ "AutoModel", "llama_bidirec", "llama_bidirectional_model.LlamaBidirectionalModel", "LlamaBidirectionalModel" ]
[ "feature-extraction" ]
[ "multimodal" ]
[ "text" ]
[ "embeddings" ]
68f4e3f4f82416217114139c
Qwen/Qwen3-VL-2B-Instruct
Qwen
null
2,528,954
18,035,950
False
2025-10-19T13:13:24Z
2025-10-23T11:30:44Z
transformers
355
4
null
image-text-to-text
{"parameters": {"BF16": 2127532032}, "total": 2127532032}
[ ".gitattributes", "README.md", "chat_template.json", "config.json", "generation_config.json", "merges.txt", "model.safetensors", "preprocessor_config.json", "tokenizer.json", "tokenizer_config.json", "video_preprocessor_config.json", "vocab.json" ]
89644892e4d85e24eaac8bacfd4f463576704203
[ "transformers", "safetensors", "qwen3_vl", "image-text-to-text", "conversational", "arxiv:2505.09388", "arxiv:2502.13923", "arxiv:2409.12191", "arxiv:2308.12966", "license:apache-2.0", "endpoints_compatible", "deploy:azure", "region:us" ]
null
{"architectures": ["Qwen3VLForConditionalGeneration"], "model_type": "qwen3_vl", "processor_config": {"chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0].role == 'system' %}\n {%- if messages[0].content is string %}\n {{- messages[0].content }}\n {%- el...
{ "auto_model": "AutoModelForImageTextToText", "custom_class": null, "pipeline_tag": "image-text-to-text", "processor": "AutoProcessor" }
{"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "apache-2.0", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "image-text-to-text", "tags": null}
null
null
[ "apache-2.0" ]
null
null
2,127,532,032
null
null
[ "AutoModelForImageTextToText", "qwen3_vl", "Qwen3VLForConditionalGeneration" ]
[ "image-text-to-text" ]
[ "multimodal" ]
[ "text", "image" ]
[ "text" ]
6901435c6b9d023f25308d41
ByteDance/Ouro-2.6B-Thinking
ByteDance
null
7,921
22,747
False
2025-10-28T22:27:40Z
2026-02-26T18:39:17Z
transformers
100
4
null
text-generation
null
[ ".gitattributes", "README.md", "assets/logo.png", "assets/ouro_thinking.png", "config.json", "configuration_ouro.py", "merges.txt", "model.safetensors", "modeling_ouro.py", "special_tokens_map.json", "tokenizer.json", "tokenizer_config.json", "vocab.json" ]
f1edd81e7ac41355db670500ceaf204e0f73af68
[ "transformers", "safetensors", "ouro", "text-generation", "looped-language-model", "reasoning", "recurrent-depth", "thinking", "chain-of-thought", "conversational", "custom_code", "arxiv:2510.25741", "license:apache-2.0", "region:us" ]
null
{"architectures": ["OuroForCausalLM"], "auto_map": {"AutoConfig": "configuration_ouro.OuroConfig", "AutoModel": "modeling_ouro.OuroModel", "AutoModelForCausalLM": "modeling_ouro.OuroForCausalLM"}, "model_type": "ouro", "tokenizer_config": {"bos_token": "<|im_start|>", "chat_template": "{%- if messages[0]['role'] == 'sy...
{ "auto_model": "AutoModelForCausalLM", "custom_class": "modeling_ouro.OuroForCausalLM", "pipeline_tag": "text-generation", "processor": null }
{"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "apache-2.0", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "text-generation", "tags": ["looped-language-model", "reasoning", "recurrent-depth", "t...
null
null
[ "apache-2.0" ]
null
null
null
null
null
[ "ouro", "AutoModelForCausalLM", "OuroForCausalLM", "modeling_ouro.OuroForCausalLM" ]
[ "text-generation" ]
[ "text" ]
[ "text" ]
[ "text" ]
621ffdc036468d709f174332
google-bert/bert-base-chinese
google-bert
null
3,273,190
139,639,994
False
2022-03-02T23:29:04Z
2025-07-03T11:58:48Z
transformers
1,403
3
null
fill-mask
null
[ ".gitattributes", "README.md", "config.json", "flax_model.msgpack", "model.safetensors", "pytorch_model.bin", "tf_model.h5", "tokenizer.json", "tokenizer_config.json", "vocab.txt" ]
8f23c25b06e129b6c986331a13d8d025a92cf0ea
[ "transformers", "pytorch", "tf", "jax", "safetensors", "bert", "fill-mask", "zh", "arxiv:1810.04805", "license:apache-2.0", "endpoints_compatible", "deploy:azure", "region:us" ]
null
{"architectures": ["BertForMaskedLM"], "model_type": "bert", "tokenizer_config": {}}
{ "auto_model": "AutoModelForMaskedLM", "custom_class": null, "pipeline_tag": "fill-mask", "processor": "AutoTokenizer" }
{"base_model": null, "datasets": null, "eval_results": null, "language": "zh", "library_name": null, "license": "apache-2.0", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": null, "tags": null}
null
null
[ "apache-2.0" ]
null
[ "zh" ]
null
null
null
[ "AutoModelForMaskedLM", "bert", "BertForMaskedLM" ]
[ "fill-mask" ]
[ "text" ]
[ "text" ]
[ "logits" ]
621ffdc036468d709f174358
google-t5/t5-small
google-t5
null
1,813,634
174,084,827
False
2022-03-02T23:29:04Z
2023-06-30T02:31:26Z
transformers
536
3
null
translation
{"parameters": {"F32": 60506880}, "total": 60506880}
[ ".gitattributes", "README.md", "config.json", "flax_model.msgpack", "generation_config.json", "model.safetensors", "onnx/decoder_model.onnx", "onnx/decoder_model_merged.onnx", "onnx/decoder_model_merged_quantized.onnx", "onnx/decoder_model_quantized.onnx", "onnx/decoder_with_past_model.onnx", ...
df1b051c49625cf57a3d0d8d3863ed4d13564fe4
[ "transformers", "pytorch", "tf", "jax", "rust", "onnx", "safetensors", "t5", "text2text-generation", "summarization", "translation", "en", "fr", "ro", "de", "multilingual", "dataset:c4", "arxiv:1805.12471", "arxiv:1708.00055", "arxiv:1704.05426", "arxiv:1606.05250", "arxiv:...
null
{"architectures": ["T5ForConditionalGeneration"], "model_type": "t5", "tokenizer_config": {"eos_token": "</s>", "pad_token": "<pad>", "unk_token": "<unk>"}}
{ "auto_model": "AutoModelForSeq2SeqLM", "custom_class": null, "pipeline_tag": "text2text-generation", "processor": "AutoTokenizer" }
{"base_model": null, "datasets": ["c4"], "eval_results": null, "language": ["en", "fr", "ro", "de", "multilingual"], "library_name": null, "license": "apache-2.0", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": null, "tags": ["summarization", "translation"]}
# Model Card for T5 Small ![model image](https://camo.githubusercontent.com/623b4dea0b653f2ad3f36c71ebfe749a677ac0a1/68747470733a2f2f6d69726f2e6d656469756d2e636f6d2f6d61782f343030362f312a44304a31674e51663876727255704b657944387750412e706e67) # Table of Contents 1. [Model Details](#model-details) 2. [Uses](#uses) 3. ...
null
[ "apache-2.0" ]
[ "c4" ]
[ "en", "fr", "ro", "de", "multilingual" ]
60,506,880
null
null
[ "t5", "T5ForConditionalGeneration", "AutoModelForSeq2SeqLM" ]
[ "text2text-generation", "translation", "summarization" ]
[ "text" ]
[ "text" ]
[ "text" ]
621ffdc136468d709f17a149
cointegrated/rubert-tiny2
cointegrated
null
1,135,308
17,129,341
False
2022-03-02T23:29:05Z
2025-12-30T08:33:24Z
sentence-transformers
166
3
null
sentence-similarity
{"parameters": {"I64": 2048, "F32": 29376502}, "total": 29378550}
[ ".gitattributes", ".gitignore", "1_Pooling/config.json", "README.md", "config.json", "model.safetensors", "modules.json", "pytorch_model.bin", "sentence_bert_config.json", "special_tokens_map.json", "tinybert-ru-labse-adapter-v2.pt", "tokenizer.json", "tokenizer_config.json", "vocab.txt" ]
e8ed3b0c8bbf4fb6984c3de043bf7d2f4e5969ae
[ "sentence-transformers", "pytorch", "safetensors", "bert", "pretraining", "russian", "fill-mask", "embeddings", "masked-lm", "tiny", "feature-extraction", "sentence-similarity", "transformers", "ru", "license:mit", "text-embeddings-inference", "endpoints_compatible", "region:us" ]
null
{"architectures": ["BertForPreTraining"], "model_type": "bert", "tokenizer_config": {"unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]"}}
{ "auto_model": "AutoModelForPreTraining", "custom_class": null, "pipeline_tag": "pretraining", "processor": "AutoTokenizer" }
{"base_model": null, "datasets": null, "eval_results": null, "language": ["ru"], "library_name": null, "license": "mit", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "sentence-similarity", "tags": ["russian", "fill-mask", "pretraining", "embeddings", "masked-lm", "tin...
This is an updated version of [cointegrated/rubert-tiny](https://huggingface.co/cointegrated/rubert-tiny): a small Russian BERT-based encoder with high-quality sentence embeddings. This [post in Russian](https://habr.com/ru/post/669674/) gives more details. The differences from the previous version include: - a larger...
null
[ "mit" ]
null
[ "ru" ]
29,378,550
null
null
[ "bert", "AutoModelForPreTraining", "BertForPreTraining" ]
[ "fill-mask", "sentence-similarity", "feature-extraction", "pretraining" ]
[ "text", "multimodal" ]
[ "text" ]
[ "logits", "embeddings" ]
621ffdc136468d709f17adb6
facebook/bart-large-cnn
facebook
null
2,055,150
138,785,394
False
2022-03-02T23:29:05Z
2024-02-13T18:02:05Z
transformers
1,552
3
[{"name": "facebook/bart-large-cnn", "results": [{"task": {"type": "summarization", "name": "Summarization"}, "dataset": {"name": "cnn_dailymail", "type": "cnn_dailymail", "config": "3.0.0", "split": "train"}, "metrics": [{"name": "ROUGE-1", "type": "rouge", "value": 42.9486, "verified": false}, {"name": "ROUGE-2", "ty...
summarization
{"parameters": {"F32": 406290432}, "total": 406290432}
[ ".gitattributes", "README.md", "config.json", "flax_model.msgpack", "generation_config.json", "generation_config_for_summarization.json", "merges.txt", "model.safetensors", "pytorch_model.bin", "rust_model.ot", "tf_model.h5", "tokenizer.json", "vocab.json" ]
37f520fa929c961707657b28798b30c003dd100b
[ "transformers", "pytorch", "tf", "jax", "rust", "safetensors", "bart", "text2text-generation", "summarization", "en", "dataset:cnn_dailymail", "arxiv:1910.13461", "license:mit", "model-index", "endpoints_compatible", "deploy:azure", "region:us" ]
null
{"architectures": ["BartForConditionalGeneration"], "model_type": "bart"}
{ "auto_model": "AutoModelForSeq2SeqLM", "custom_class": null, "pipeline_tag": "text2text-generation", "processor": "AutoTokenizer" }
{"datasets": ["cnn_dailymail"], "language": ["en"], "license": "mit", "pipeline_tag": "summarization", "thumbnail": "https://huggingface.co/front/thumbnails/facebook.png", "model-index": [{"name": "facebook/bart-large-cnn", "results": [{"task": {"type": "summarization", "name": "Summarization"}, "dataset": {"name": "cn...
# BART (large-sized model), fine-tuned on CNN Daily Mail BART model pre-trained on English language, and fine-tuned on [CNN Daily Mail](https://huggingface.co/datasets/cnn_dailymail). It was introduced in the paper [BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Co...
null
[ "mit" ]
[ "cnn_dailymail" ]
[ "en" ]
406,290,432
null
null
[ "bart", "BartForConditionalGeneration", "AutoModelForSeq2SeqLM" ]
[ "text2text-generation", "summarization" ]
[ "text" ]
[ "text" ]
[ "text" ]
62c8018851eb740327229910
facebook/nllb-200-3.3B
facebook
null
58,098
2,270,668
False
2022-07-08T10:06:00Z
2023-02-11T20:19:13Z
transformers
434
3
null
translation
null
[ ".gitattributes", "README.md", "config.json", "generation_config.json", "pytorch_model-00001-of-00003.bin", "pytorch_model-00002-of-00003.bin", "pytorch_model-00003-of-00003.bin", "pytorch_model.bin.index.json", "sentencepiece.bpe.model", "special_tokens_map.json", "tokenizer.json", "tokenizer...
1a07f7d195896b2114afcb79b7b57ab512e7b43e
[ "transformers", "pytorch", "m2m_100", "text2text-generation", "nllb", "translation", "ace", "acm", "acq", "aeb", "af", "ajp", "ak", "als", "am", "apc", "ar", "ars", "ary", "arz", "as", "ast", "awa", "ayr", "azb", "azj", "ba", "bm", "ban", "be", "bem", "b...
null
{"architectures": ["M2M100ForConditionalGeneration"], "model_type": "m2m_100", "tokenizer_config": {"bos_token": "<s>", "cls_token": "<s>", "eos_token": "</s>", "mask_token": {"__type": "AddedToken", "content": "<mask>", "lstrip": true, "normalized": true, "rstrip": false, "single_word": false}, "pad_token": "<pad>", "...
{ "auto_model": "AutoModelForSeq2SeqLM", "custom_class": null, "pipeline_tag": "text2text-generation", "processor": "AutoTokenizer" }
{"base_model": null, "datasets": ["flores-200"], "eval_results": null, "language": ["ace", "acm", "acq", "aeb", "af", "ajp", "ak", "als", "am", "apc", "ar", "ars", "ary", "arz", "as", "ast", "awa", "ayr", "azb", "azj", "ba", "bm", "ban", "be", "bem", "bn", "bho", "bjn", "bo", "bs", "bug", "bg", "ca", "ceb", "cs", "cjk"...
# NLLB-200 This is the model card of NLLB-200's 3.3B variant. Here are the [metrics](https://tinyurl.com/nllb200dense3bmetrics) for that particular checkpoint. - Information about training algorithms, parameters, fairness constraints or other applied approaches, and features. The exact training algorithm, data and t...
null
[ "cc-by-nc-4.0" ]
[ "flores-200" ]
[ "ace", "acm", "acq", "aeb", "af", "ajp", "ak", "als", "am", "apc", "ar", "ars", "ary", "arz", "as", "ast", "awa", "ayr", "azb", "azj", "ba", "bm", "ban", "be", "bem", "bn", "bho", "bjn", "bo", "bs", "bug", "bg", "ca", "ceb", "cs", "cjk", "ckb",...
null
null
[ "bleu", "spbleu", "chrf++" ]
[ "AutoModelForSeq2SeqLM", "m2m_100", "M2M100ForConditionalGeneration" ]
[ "text2text-generation", "translation" ]
[ "text" ]
[ "text" ]
[ "text" ]
63314bb6acb6472115aa55a9
openai/whisper-tiny
openai
null
776,748
14,064,428
False
2022-09-26T06:50:30Z
2024-02-29T10:57:33Z
transformers
420
3
[{"name": "whisper-tiny", "results": [{"task": {"name": "Automatic Speech Recognition", "type": "automatic-speech-recognition"}, "dataset": {"name": "LibriSpeech (clean)", "type": "librispeech_asr", "config": "clean", "split": "test", "args": {"language": "en"}}, "metrics": [{"name": "Test WER", "type": "wer", "value":...
automatic-speech-recognition
null
[ ".gitattributes", "README.md", "added_tokens.json", "config.json", "flax_model.msgpack", "generation_config.json", "merges.txt", "model.safetensors", "normalizer.json", "preprocessor_config.json", "pytorch_model.bin", "special_tokens_map.json", "tf_model.h5", "tokenizer.json", "tokenizer...
169d4a4341b33bc18d8881c4b69c2e104e1cc0af
[ "transformers", "pytorch", "tf", "jax", "safetensors", "whisper", "automatic-speech-recognition", "audio", "hf-asr-leaderboard", "en", "zh", "de", "es", "ru", "ko", "fr", "ja", "pt", "tr", "pl", "ca", "nl", "ar", "sv", "it", "id", "hi", "fi", "vi", "he", "...
null
{"architectures": ["WhisperForConditionalGeneration"], "model_type": "whisper", "tokenizer_config": {"bos_token": "<|endoftext|>", "eos_token": "<|endoftext|>", "pad_token": "<|endoftext|>", "unk_token": "<|endoftext|>"}}
{ "auto_model": "AutoModelForSpeechSeq2Seq", "custom_class": null, "pipeline_tag": "automatic-speech-recognition", "processor": "AutoProcessor" }
{"language": ["en", "zh", "de", "es", "ru", "ko", "fr", "ja", "pt", "tr", "pl", "ca", "nl", "ar", "sv", "it", "id", "hi", "fi", "vi", "he", "uk", "el", "ms", "cs", "ro", "da", "hu", "ta", "no", "th", "ur", "hr", "bg", "lt", "la", "mi", "ml", "cy", "sk", "te", "fa", "lv", "bn", "sr", "az", "sl", "kn", "et", "mk", "br", ...
# Whisper Whisper is a pre-trained model for automatic speech recognition (ASR) and speech translation. Trained on 680k hours of labelled data, Whisper models demonstrate a strong ability to generalise to many datasets and domains **without** the need for fine-tuning. Whisper was proposed in the paper [Robust Speec...
null
[ "apache-2.0" ]
null
[ "en", "zh", "de", "es", "ru", "ko", "fr", "ja", "pt", "tr", "pl", "ca", "nl", "ar", "sv", "it", "id", "hi", "fi", "vi", "he", "uk", "el", "ms", "cs", "ro", "da", "hu", "ta", "no", "th", "ur", "hr", "bg", "lt", "la", "mi", "ml", "cy", "sk"...
null
null
null
[ "WhisperForConditionalGeneration", "whisper", "AutoModelForSpeechSeq2Seq" ]
[ "automatic-speech-recognition" ]
[ "multimodal" ]
[ "audio" ]
[ "text" ]
639861ac02d4b462c93d29f5
Salesforce/blip-image-captioning-large
Salesforce
null
1,572,616
45,731,127
False
2022-12-13T11:27:40Z
2025-02-03T06:42:42Z
transformers
1,460
3
null
image-to-text
{"parameters": {"I64": 512, "F32": 469732924}, "total": 469733436}
[ ".gitattributes", "README.md", "config.json", "model.safetensors", "preprocessor_config.json", "pytorch_model.bin", "special_tokens_map.json", "tf_model.h5", "tokenizer.json", "tokenizer_config.json", "vocab.txt" ]
353689b859fcf0523410b1806dace5fb46ecdf41
[ "transformers", "pytorch", "tf", "safetensors", "blip", "image-text-to-text", "image-captioning", "image-to-text", "arxiv:2201.12086", "license:bsd-3-clause", "endpoints_compatible", "region:us" ]
null
{"architectures": ["BlipForConditionalGeneration"], "model_type": "blip", "tokenizer_config": {"cls_token": "[CLS]", "mask_token": "[MASK]", "pad_token": "[PAD]", "sep_token": "[SEP]", "unk_token": "[UNK]"}}
{ "auto_model": "AutoModelForImageTextToText", "custom_class": null, "pipeline_tag": "image-text-to-text", "processor": "AutoProcessor" }
{"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": null, "license": "bsd-3-clause", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "image-to-text", "tags": ["image-captioning"], "languages": ["en"]}
# BLIP: Bootstrapping Language-Image Pre-training for Unified Vision-Language Understanding and Generation Model card for image captioning pretrained on COCO dataset - base architecture (with ViT large backbone). | ![BLIP.gif](https://cdn-uploads.huggingface.co/production/uploads/1670928184033-62441d1d9fdefb55a0b7d12...
null
[ "bsd-3-clause" ]
null
null
469,733,436
null
null
[ "blip", "AutoModelForImageTextToText", "BlipForConditionalGeneration" ]
[ "image-text-to-text", "image-to-text" ]
[ "multimodal" ]
[ "text", "image" ]
[ "text" ]
644677e41173e85ac7f10745
bigcode/starcoder
bigcode
null
9,760
674,568
auto
2023-04-24T12:36:52Z
2024-10-08T20:53:18Z
transformers
2,929
3
[{"name": "StarCoder", "results": [{"task": {"type": "text-generation"}, "dataset": {"type": "openai_humaneval", "name": "HumanEval (Prompted)"}, "metrics": [{"name": "pass@1", "type": "pass@1", "value": 0.408, "verified": false}]}, {"task": {"type": "text-generation"}, "dataset": {"type": "openai_humaneval", "name": "...
text-generation
{"parameters": {"F32": 15819446272}, "total": 15819446272}
[ ".gitattributes", "README.md", "config.json", "generation_config.json", "merges.txt", "model-00001-of-00007.safetensors", "model-00002-of-00007.safetensors", "model-00003-of-00007.safetensors", "model-00004-of-00007.safetensors", "model-00005-of-00007.safetensors", "model-00006-of-00007.safetens...
827cd7721231872c152e3faef5cae1b22964ed58
[ "transformers", "pytorch", "safetensors", "gpt_bigcode", "text-generation", "code", "dataset:bigcode/the-stack-dedup", "arxiv:1911.02150", "arxiv:2205.14135", "arxiv:2207.14255", "arxiv:2305.06161", "license:bigcode-openrail-m", "model-index", "text-generation-inference", "endpoints_comp...
null
{"architectures": ["GPTBigCodeForCausalLM"], "model_type": "gpt_bigcode", "tokenizer_config": {"bos_token": "<|endoftext|>", "eos_token": "<|endoftext|>", "unk_token": "<|endoftext|>"}}
{ "auto_model": "AutoModelForCausalLM", "custom_class": null, "pipeline_tag": "text-generation", "processor": "AutoTokenizer" }
{"datasets": ["bigcode/the-stack-dedup"], "library_name": "transformers", "license": "bigcode-openrail-m", "metrics": ["code_eval"], "pipeline_tag": "text-generation", "tags": ["code"], "inference": true, "widget": [{"text": "def print_hello_world():", "example_title": "Hello world", "group": "Python"}], "extra_gated_p...
null
null
[ "bigcode-openrail-m" ]
[ "bigcode/the-stack-dedup" ]
null
15,819,446,272
null
[ "code_eval" ]
[ "gpt_bigcode", "AutoModelForCausalLM", "GPTBigCodeForCausalLM" ]
[ "text-generation" ]
[ "text" ]
[ "text" ]
[ "text" ]
650aedb6238a644cb93a52c3
mistralai/Mistral-7B-v0.1
mistralai
null
454,618
23,222,227
False
2023-09-20T13:03:50Z
2025-07-24T16:44:02Z
transformers
4,057
3
null
text-generation
{"parameters": {"BF16": 7241732096}, "total": 7241732096}
[ ".gitattributes", "README.md", "config.json", "generation_config.json", "model-00001-of-00002.safetensors", "model-00002-of-00002.safetensors", "model.safetensors.index.json", "pytorch_model-00001-of-00002.bin", "pytorch_model-00002-of-00002.bin", "pytorch_model.bin.index.json", "special_tokens_...
27d67f1b5f57dc0953326b2601d68371d40ea8da
[ "transformers", "pytorch", "safetensors", "mistral", "text-generation", "pretrained", "mistral-common", "en", "arxiv:2310.06825", "license:apache-2.0", "text-generation-inference", "region:us" ]
null
{"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "eos_token": "</s>", "pad_token": null, "unk_token": "<unk>", "use_default_system_prompt": false}}
{ "auto_model": "AutoModelForCausalLM", "custom_class": null, "pipeline_tag": "text-generation", "processor": "AutoTokenizer" }
{"base_model": null, "datasets": null, "eval_results": null, "language": ["en"], "library_name": "transformers", "license": "apache-2.0", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": null, "tags": ["pretrained", "mistral-common"], "inference": false, "extra_gated_desc...
null
null
[ "apache-2.0" ]
null
[ "en" ]
7,241,732,096
null
null
[ "AutoModelForCausalLM", "mistral", "MistralForCausalLM" ]
[ "text-generation" ]
[ "text" ]
[ "text" ]
[ "text" ]
65d4e83bc2e85b8b98af701f
bigcode/starcoder2-15b
bigcode
null
10,053
517,419
False
2024-02-20T17:58:19Z
2024-06-05T19:52:45Z
transformers
664
3
[{"name": "starcoder2-15b", "results": [{"task": {"type": "text-generation"}, "dataset": {"name": "CruxEval-I", "type": "cruxeval-i"}, "metrics": [{"type": "pass@1", "value": 48.1, "verified": false}]}, {"task": {"type": "text-generation"}, "dataset": {"name": "DS-1000", "type": "ds-1000"}, "metrics": [{"type": "pass@1...
text-generation
null
[ ".gitattributes", "README.md", "config.json", "generation_config.json", "merges.txt", "model-00001-of-00014.safetensors", "model-00002-of-00014.safetensors", "model-00003-of-00014.safetensors", "model-00004-of-00014.safetensors", "model-00005-of-00014.safetensors", "model-00006-of-00014.safetens...
46d44742909c03ac8cee08eb03fdebce02e193ec
[ "transformers", "safetensors", "starcoder2", "text-generation", "code", "dataset:bigcode/the-stack-v2-train", "arxiv:2305.13245", "arxiv:2205.14135", "arxiv:2004.05150", "arxiv:2207.14255", "arxiv:2402.19173", "license:bigcode-openrail-m", "model-index", "text-generation-inference", "end...
null
{"architectures": ["Starcoder2ForCausalLM"], "model_type": "starcoder2", "tokenizer_config": {"bos_token": "<|endoftext|>", "eos_token": "<|endoftext|>", "unk_token": "<|endoftext|>"}}
{ "auto_model": "AutoModelForCausalLM", "custom_class": null, "pipeline_tag": "text-generation", "processor": "AutoTokenizer" }
{"datasets": ["bigcode/the-stack-v2-train"], "library_name": "transformers", "license": "bigcode-openrail-m", "pipeline_tag": "text-generation", "tags": ["code"], "inference": {"parameters": {"temperature": 0.2, "top_p": 0.95}}, "widget": [{"text": "def print_hello_world():", "example_title": "Hello world", "group": "P...
# StarCoder2 <center> <img src="https://huggingface.co/datasets/bigcode/admin_private/resolve/main/starcoder2_banner.png" alt="SC2" width="900" height="600"> </center> ## Table of Contents 1. [Model Summary](#model-summary) 2. [Use](#use) 3. [Limitations](#limitations) 4. [Training](#training) 5. [License](#lic...
null
[ "bigcode-openrail-m" ]
[ "bigcode/the-stack-v2-train" ]
null
null
null
null
[ "AutoModelForCausalLM", "starcoder2", "Starcoder2ForCausalLM" ]
[ "text-generation" ]
[ "text" ]
[ "text" ]
[ "text" ]
65fc66d12c813664be7731ae
google/codegemma-7b
google
null
1,183
401,887
manual
2024-03-21T16:56:49Z
2024-08-07T18:27:09Z
transformers
215
3
null
text-generation
null
[ ".gitattributes", "README.md", "codegemma_nl_benchmarks.png", "config.json", "generation_config.json", "model-00001-of-00004.safetensors", "model-00002-of-00004.safetensors", "model-00003-of-00004.safetensors", "model-00004-of-00004.safetensors", "model.safetensors.index.json", "special_tokens_m...
2ec970089fe8f083d10d1ce6bac30538ac4badaa
[ "transformers", "safetensors", "gemma", "text-generation", "license:gemma", "text-generation-inference", "endpoints_compatible", "region:us" ]
null
{"architectures": ["GemmaForCausalLM"], "model_type": "gemma", "tokenizer_config": {"bos_token": "<bos>", "eos_token": "<eos>", "pad_token": "<pad>", "unk_token": "<unk>", "use_default_system_prompt": false}}
{ "auto_model": "AutoModelForCausalLM", "custom_class": null, "pipeline_tag": "text-generation", "processor": "AutoTokenizer" }
{"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "gemma", "license_name": null, "license_link": "https://ai.google.dev/gemma/terms", "metrics": null, "model_name": null, "pipeline_tag": null, "tags": null, "extra_gated_heading": "Access CodeGemma ...
null
null
[ "gemma", "https://ai.google.dev/gemma/terms" ]
null
null
null
null
null
[ "AutoModelForCausalLM", "gemma", "GemmaForCausalLM" ]
[ "text-generation" ]
[ "text" ]
[ "text" ]
[ "text" ]
6623f2c68fee5ba324a5fecc
protectai/deberta-v3-base-prompt-injection-v2
protectai
{ "models": [ { "_id": "621ffdc136468d709f17dee4", "id": "microsoft/deberta-v3-base" } ], "relation": "quantized" }
436,329
4,305,262
False
2024-04-20T16:52:22Z
2024-05-28T07:07:49Z
transformers
104
3
[{"name": "deberta-v3-base-prompt-injection-v2", "results": []}]
text-classification
{"parameters": {"F32": 184423682}, "total": 184423682}
[ ".gitattributes", "LICENSE", "README.md", "added_tokens.json", "config.json", "deberta-v3-base-prompt-injection-v2_emissions.csv", "model.safetensors", "onnx/added_tokens.json", "onnx/config.json", "onnx/model.onnx", "onnx/special_tokens_map.json", "onnx/spm.model", "onnx/tokenizer.json", ...
e6535ca4ce3ba852083e75ec585d7c8aeb4be4c5
[ "transformers", "onnx", "safetensors", "deberta-v2", "text-classification", "prompt-injection", "injection", "security", "llm-security", "generated_from_trainer", "en", "dataset:natolambert/xstest-v2-copy", "dataset:VMware/open-instruct", "dataset:alespalla/chatbot_instruction_prompts", ...
null
{"architectures": ["DebertaV2ForSequenceClassification"], "model_type": "deberta-v2", "tokenizer_config": {"bos_token": "[CLS]", "cls_token": "[CLS]", "eos_token": "[SEP]", "mask_token": "[MASK]", "pad_token": "[PAD]", "sep_token": "[SEP]", "unk_token": "[UNK]"}}
{ "auto_model": "AutoModelForSequenceClassification", "custom_class": null, "pipeline_tag": "text-classification", "processor": "AutoTokenizer" }
{"base_model": "microsoft/deberta-v3-base", "datasets": ["natolambert/xstest-v2-copy", "VMware/open-instruct", "alespalla/chatbot_instruction_prompts", "HuggingFaceH4/grok-conversation-harmless", "Harelix/Prompt-Injection-Mixed-Techniques-2024", "OpenSafetyLab/Salad-Data", "jackhhao/jailbreak-classification"], "eval_re...
# Model Card for deberta-v3-base-prompt-injection-v2 This model is a fine-tuned version of [microsoft/deberta-v3-base](https://huggingface.co/microsoft/deberta-v3-base) specifically developed to detect and classify prompt injection attacks which can manipulate language models into producing unintended outputs. ## Int...
null
[ "apache-2.0" ]
[ "natolambert/xstest-v2-copy", "VMware/open-instruct", "alespalla/chatbot_instruction_prompts", "HuggingFaceH4/grok-conversation-harmless", "Harelix/Prompt-Injection-Mixed-Techniques-2024", "OpenSafetyLab/Salad-Data", "jackhhao/jailbreak-classification" ]
[ "en" ]
184,423,682
null
[ "accuracy", "recall", "precision", "f1" ]
[ "AutoModelForSequenceClassification", "deberta-v2", "DebertaV2ForSequenceClassification" ]
[ "text-classification" ]
[ "text" ]
[ "text" ]
[ "logits" ]
66268dc9cd54f5fffb4e2e24
microsoft/Phi-3-mini-4k-instruct
microsoft
null
747,020
27,675,955
False
2024-04-22T16:18:17Z
2025-12-10T20:22:18Z
transformers
1,403
3
null
text-generation
{"parameters": {"BF16": 3821079552}, "total": 3821079552}
[ ".gitattributes", "CODE_OF_CONDUCT.md", "LICENSE", "NOTICE.md", "README.md", "SECURITY.md", "added_tokens.json", "config.json", "configuration_phi3.py", "data_summary_card.md", "generation_config.json", "model-00001-of-00002.safetensors", "model-00002-of-00002.safetensors", "model.safetens...
f39ac1d28e925b323eae81227eaba4464caced4e
[ "transformers", "safetensors", "phi3", "text-generation", "nlp", "code", "conversational", "custom_code", "en", "fr", "license:mit", "eval-results", "text-generation-inference", "endpoints_compatible", "region:us" ]
null
{"architectures": ["Phi3ForCausalLM"], "auto_map": {"AutoConfig": "configuration_phi3.Phi3Config", "AutoModelForCausalLM": "modeling_phi3.Phi3ForCausalLM"}, "model_type": "phi3", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{% for message in messages %}{% if message['role'] == 'system' %}{{'<|system|>\n' ...
{ "auto_model": "AutoModelForCausalLM", "custom_class": null, "pipeline_tag": "text-generation", "processor": "AutoTokenizer" }
{"base_model": null, "datasets": null, "eval_results": null, "language": ["en", "fr"], "library_name": null, "license": "mit", "license_name": null, "license_link": "https://huggingface.co/microsoft/Phi-3-mini-4k-instruct/resolve/main/LICENSE", "metrics": null, "model_name": null, "pipeline_tag": "text-generation", "ta...
🎉 **Phi-3.5**: [[mini-instruct]](https://huggingface.co/microsoft/Phi-3.5-mini-instruct); [[MoE-instruct]](https://huggingface.co/microsoft/Phi-3.5-MoE-instruct) ; [[vision-instruct]](https://huggingface.co/microsoft/Phi-3.5-vision-instruct) ## Model Summary The Phi-3-Mini-4K-Instruct is a 3.8B parameters, lightweig...
null
[ "mit", "https://huggingface.co/microsoft/Phi-3-mini-4k-instruct/resolve/main/LICENSE" ]
null
[ "en", "fr" ]
3,821,079,552
null
null
[ "AutoModelForCausalLM", "phi3", "Phi3ForCausalLM" ]
[ "text-generation" ]
[ "text" ]
[ "text" ]
[ "text" ]
66286c9842709fe826603f58
aaditya/Llama3-OpenBioLLM-70B
aaditya
{ "models": [ { "_id": "661f97be9cae93c02393e8ef", "id": "meta-llama/Meta-Llama-3-70B-Instruct" } ], "relation": "finetune" }
3,282
336,928
False
2024-04-24T02:21:12Z
2025-01-18T07:39:15Z
transformers
503
3
[{"name": "OpenBioLLM-70B", "results": []}]
text-generation
null
[ ".gitattributes", "LICENSE", "README.md", "config.json", "generation_config.json", "pytorch_model-00001-of-00030.bin", "pytorch_model-00002-of-00030.bin", "pytorch_model-00003-of-00030.bin", "pytorch_model-00004-of-00030.bin", "pytorch_model-00005-of-00030.bin", "pytorch_model-00006-of-00030.bin...
7ad17ef0d2185811f731f89d20885b2f99b1e994
[ "transformers", "pytorch", "llama", "text-generation", "llama-3", "Mixtral", "instruct", "finetune", "chatml", "DPO", "RLHF", "gpt4", "distillation", "heathcare", "medical", "clinical", "med", "lifescience", "Pharmaceutical", "Pharma", "conversational", "en", "arxiv:2305....
null
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|e...
{ "auto_model": "AutoModelForCausalLM", "custom_class": null, "pipeline_tag": "text-generation", "processor": "AutoTokenizer" }
{"base_model": "meta-llama/Meta-Llama-3-70B-Instruct", "datasets": null, "eval_results": [], "language": ["en"], "library_name": null, "license": "llama3", "license_name": null, "license_link": null, "metrics": null, "model_name": "OpenBioLLM-70B", "pipeline_tag": null, "tags": ["llama-3", "llama", "Mixtral", "instruct...
<div align="center"> <img width="260px" src="https://cdn-uploads.huggingface.co/production/uploads/5f3fe13d79c1ba4c353d0c19/BrQCb95lmEIFz79QAmoNA.png"></div> ![image/png](https://cdn-uploads.huggingface.co/production/uploads/5f3fe13d79c1ba4c353d0c19/fJIOPJnY6Ff6fUiSIuMEt.png) <div align="center"> <h1>Advancin...
null
[ "llama3" ]
null
[ "en" ]
null
null
null
[ "AutoModelForCausalLM", "llama", "LlamaForCausalLM" ]
[ "text-generation" ]
[ "text" ]
[ "text" ]
[ "text" ]
666ce6f4067382b3e98317c6
microsoft/Florence-2-base
microsoft
null
811,173
8,822,257
False
2024-06-15T00:57:24Z
2025-08-04T17:36:57Z
transformers
357
3
null
image-text-to-text
{"parameters": {"F16": 231567705}, "total": 231567705}
[ ".gitattributes", "CODE_OF_CONDUCT.md", "LICENSE", "README.md", "SECURITY.md", "SUPPORT.md", "config.json", "configuration_florence2.py", "model.safetensors", "modeling_florence2.py", "preprocessor_config.json", "processing_florence2.py", "pytorch_model.bin", "tokenizer.json", "tokenizer...
5ca5edf5bd017b9919c05d08aebef5e4c7ac3bac
[ "transformers", "pytorch", "safetensors", "florence2", "image-text-to-text", "vision", "custom_code", "arxiv:2311.06242", "license:mit", "endpoints_compatible", "region:us" ]
null
{"architectures": ["Florence2ForConditionalGeneration"], "auto_map": {"AutoConfig": "configuration_florence2.Florence2Config", "AutoModelForCausalLM": "modeling_florence2.Florence2ForConditionalGeneration"}, "model_type": "florence2", "tokenizer_config": {}}
{ "auto_model": "AutoModelForImageTextToText", "custom_class": null, "pipeline_tag": "image-text-to-text", "processor": "AutoProcessor" }
{"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": null, "license": "mit", "license_name": null, "license_link": "https://huggingface.co/microsoft/Florence-2-base/resolve/main/LICENSE", "metrics": null, "model_name": null, "pipeline_tag": "image-text-to-text", "tags": ["visio...
null
null
[ "mit", "https://huggingface.co/microsoft/Florence-2-base/resolve/main/LICENSE" ]
null
null
231,567,705
null
null
[ "florence2", "AutoModelForImageTextToText", "Florence2ForConditionalGeneration" ]
[ "image-text-to-text" ]
[ "multimodal" ]
[ "text", "image" ]
[ "text" ]
669b774854729d168aeb6207
Alibaba-NLP/gte-multilingual-base
Alibaba-NLP
null
960,722
22,385,282
False
2024-07-20T08:37:28Z
2025-07-05T01:38:16Z
sentence-transformers
350
3
[{"name": "gte-multilingual-base (dense)", "results": [{"task": {"type": "Clustering"}, "dataset": {"type": "PL-MTEB/8tags-clustering", "name": "MTEB 8TagsClustering", "config": "default", "split": "test", "revision": "None"}, "metrics": [{"type": "v_measure", "value": 33.66681726329994, "verified": false}]}, {"task": ...
sentence-similarity
{"parameters": {"F16": 305369089}, "total": 305369089}
[ ".gitattributes", "1_Pooling/config.json", "README.md", "config.json", "images/mgte-loco.png", "images/mgte-mldr.png", "images/mgte-mteb.png", "images/mgte-reranker.png", "images/mgte-retrieval.png", "model.safetensors", "modules.json", "scripts/gte_embedding.py", "sentence_bert_config.json"...
9bbca17d9273fd0d03d5725c7a4b0f6b45142062
[ "sentence-transformers", "safetensors", "new", "feature-extraction", "mteb", "transformers", "multilingual", "sentence-similarity", "text-embeddings-inference", "custom_code", "af", "ar", "az", "be", "bg", "bn", "ca", "ceb", "cs", "cy", "da", "de", "el", "en", "es", ...
null
{"architectures": ["NewModel", "NewForTokenClassification"], "auto_map": {"AutoConfig": "Alibaba-NLP/new-impl--configuration.NewConfig", "AutoModelForMaskedLM": "Alibaba-NLP/new-impl--modeling.NewForMaskedLM", "AutoModel": "Alibaba-NLP/new-impl--modeling.NewModel", "AutoModelForMultipleChoice": "Alibaba-NLP/new-impl--m...
{ "auto_model": "AutoModel", "custom_class": "Alibaba-NLP/new-impl--modeling.NewModel", "pipeline_tag": "feature-extraction", "processor": null }
{"language": ["af", "ar", "az", "be", "bg", "bn", "ca", "ceb", "cs", "cy", "da", "de", "el", "en", "es", "et", "eu", "fa", "fi", "fr", "gl", "gu", "he", "hi", "hr", "ht", "hu", "hy", "id", "is", "it", "ja", "jv", "ka", "kk", "km", "kn", "ko", "ky", "lo", "lt", "lv", "mk", "ml", "mn", "mr", "ms", "my", "ne", "nl", "no",...
null
null
[ "apache-2.0" ]
null
[ "af", "ar", "az", "be", "bg", "bn", "ca", "ceb", "cs", "cy", "da", "de", "el", "en", "es", "et", "eu", "fa", "fi", "fr", "gl", "gu", "he", "hi", "hr", "ht", "hu", "hy", "id", "is", "it", "ja", "jv", "ka", "kk", "km", "kn", "ko", "ky", "lo...
305,369,089
null
null
[ "NewForTokenClassification", "Alibaba-NLP/new-impl--modeling.NewModel", "AutoModel", "new", "NewModel" ]
[ "sentence-similarity", "feature-extraction" ]
[ "text", "multimodal" ]
[ "text" ]
[ "logits", "embeddings" ]
66e6cfeb6998c3d86c14e163
Qwen/Qwen2.5-0.5B
Qwen
null
1,824,056
17,297,091
False
2024-09-15T12:15:39Z
2024-09-25T12:32:36Z
transformers
388
3
null
text-generation
{"parameters": {"BF16": 494032768}, "total": 494032768}
[ ".gitattributes", "LICENSE", "README.md", "config.json", "generation_config.json", "merges.txt", "model.safetensors", "tokenizer.json", "tokenizer_config.json", "vocab.json" ]
060db6499f32faf8b98477b0a26969ef7d8b9987
[ "transformers", "safetensors", "qwen2", "text-generation", "conversational", "en", "arxiv:2407.10671", "license:apache-2.0", "text-generation-inference", "endpoints_compatible", "deploy:azure", "region:us" ]
null
{"architectures": ["Qwen2ForCausalLM"], "model_type": "qwen2", "tokenizer_config": {"bos_token": null, "chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0]['role'] == 'system' %}\n {{- messages[0]['content'] }}\n {%- else %}\n {{- 'You are a helpful assistant.' ...
{ "auto_model": "AutoModelForCausalLM", "custom_class": null, "pipeline_tag": "text-generation", "processor": "AutoTokenizer" }
{"base_model": null, "datasets": null, "eval_results": null, "language": ["en"], "library_name": "transformers", "license": "apache-2.0", "license_name": null, "license_link": "https://huggingface.co/Qwen/Qwen2.5-0.5B/blob/main/LICENSE", "metrics": null, "model_name": null, "pipeline_tag": "text-generation", "tags": nu...
# Qwen2.5-0.5B ## Introduction Qwen2.5 is the latest series of Qwen large language models. For Qwen2.5, we release a number of base language models and instruction-tuned language models ranging from 0.5 to 72 billion parameters. Qwen2.5 brings the following improvements upon Qwen2: - Significantly **more knowledge**...
null
[ "apache-2.0", "https://huggingface.co/Qwen/Qwen2.5-0.5B/blob/main/LICENSE" ]
null
[ "en" ]
494,032,768
null
null
[ "AutoModelForCausalLM", "Qwen2ForCausalLM", "qwen2" ]
[ "text-generation" ]
[ "text" ]
[ "text" ]
[ "text" ]
66e6d032663c452ad2400439
Qwen/Qwen2.5-1.5B
Qwen
null
704,321
8,715,182
False
2024-09-15T12:16:50Z
2024-10-08T10:56:50Z
transformers
171
3
null
text-generation
{"parameters": {"BF16": 1543714304}, "total": 1543714304}
[ ".gitattributes", "LICENSE", "README.md", "config.json", "generation_config.json", "merges.txt", "model.safetensors", "tokenizer.json", "tokenizer_config.json", "vocab.json" ]
8faed761d45a263340a0528343f099c05c9a4323
[ "transformers", "safetensors", "qwen2", "text-generation", "conversational", "en", "arxiv:2407.10671", "license:apache-2.0", "text-generation-inference", "endpoints_compatible", "deploy:azure", "region:us" ]
null
{"architectures": ["Qwen2ForCausalLM"], "model_type": "qwen2", "tokenizer_config": {"bos_token": null, "chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0]['role'] == 'system' %}\n {{- messages[0]['content'] }}\n {%- else %}\n {{- 'You are a helpful assistant.' ...
{ "auto_model": "AutoModelForCausalLM", "custom_class": null, "pipeline_tag": "text-generation", "processor": "AutoTokenizer" }
{"base_model": null, "datasets": null, "eval_results": null, "language": ["en"], "library_name": "transformers", "license": "apache-2.0", "license_name": null, "license_link": "https://huggingface.co/Qwen/Qwen2.5-1.5B/blob/main/LICENSE", "metrics": null, "model_name": null, "pipeline_tag": "text-generation", "tags": nu...
# Qwen2.5-1.5B ## Introduction Qwen2.5 is the latest series of Qwen large language models. For Qwen2.5, we release a number of base language models and instruction-tuned language models ranging from 0.5 to 72 billion parameters. Qwen2.5 brings the following improvements upon Qwen2: - Significantly **more knowledge**...
null
[ "apache-2.0", "https://huggingface.co/Qwen/Qwen2.5-1.5B/blob/main/LICENSE" ]
null
[ "en" ]
1,543,714,304
null
null
[ "AutoModelForCausalLM", "Qwen2ForCausalLM", "qwen2" ]
[ "text-generation" ]
[ "text" ]
[ "text" ]
[ "text" ]
66e902f34eeaced24a2787f8
Qwen/Qwen2.5-32B-Instruct
Qwen
{ "models": [ { "_id": "66e6d09952865ca1e41c6566", "id": "Qwen/Qwen2.5-32B" } ], "relation": "finetune" }
4,098,183
19,629,113
False
2024-09-17T04:17:55Z
2024-09-25T12:33:09Z
transformers
341
3
null
text-generation
null
[ ".gitattributes", "LICENSE", "README.md", "config.json", "generation_config.json", "merges.txt", "model-00001-of-00017.safetensors", "model-00002-of-00017.safetensors", "model-00003-of-00017.safetensors", "model-00004-of-00017.safetensors", "model-00005-of-00017.safetensors", "model-00006-of-0...
5ede1c97bbab6ce5cda5812749b4c0bdf79b18dd
[ "transformers", "safetensors", "qwen2", "text-generation", "chat", "conversational", "en", "arxiv:2309.00071", "arxiv:2407.10671", "base_model:Qwen/Qwen2.5-32B", "base_model:finetune:Qwen/Qwen2.5-32B", "license:apache-2.0", "text-generation-inference", "endpoints_compatible", "deploy:azu...
null
{"architectures": ["Qwen2ForCausalLM"], "model_type": "qwen2", "tokenizer_config": {"bos_token": null, "chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0]['role'] == 'system' %}\n {{- messages[0]['content'] }}\n {%- else %}\n {{- 'You are Qwen, created by Aliba...
{ "auto_model": "AutoModelForCausalLM", "custom_class": null, "pipeline_tag": "text-generation", "processor": "AutoTokenizer" }
{"base_model": "Qwen/Qwen2.5-32B", "datasets": null, "eval_results": null, "language": ["en"], "library_name": "transformers", "license": "apache-2.0", "license_name": null, "license_link": "https://huggingface.co/Qwen/Qwen2.5-32B-Instruct/blob/main/LICENSE", "metrics": null, "model_name": null, "pipeline_tag": "text-g...
# Qwen2.5-32B-Instruct ## Introduction Qwen2.5 is the latest series of Qwen large language models. For Qwen2.5, we release a number of base language models and instruction-tuned language models ranging from 0.5 to 72 billion parameters. Qwen2.5 brings the following improvements upon Qwen2: - Significantly **more kno...
null
[ "apache-2.0", "https://huggingface.co/Qwen/Qwen2.5-32B-Instruct/blob/main/LICENSE" ]
null
[ "en" ]
null
null
null
[ "AutoModelForCausalLM", "Qwen2ForCausalLM", "qwen2" ]
[ "text-generation" ]
[ "text" ]
[ "text" ]
[ "text" ]