_id stringlengths 24 24 | id stringlengths 7 122 | author stringlengths 2 41 | base_models dict | downloads int64 0 206M | downloads_all_time int64 0 2.81B | gated stringclasses 3
values | created_at timestamp[us, tz=UTC]date 2022-03-02 23:29:04 2025-11-03 14:41:54 | last_modified timestamp[us, tz=UTC]date 2020-12-11 21:34:15 2026-03-29 10:14:05 | library_name stringclasses 36
values | likes int64 0 13.1k | trending_score float64 0 47 | model_index stringlengths 30 911k ⌀ | pipeline_tag stringclasses 46
values | safetensors stringlengths 30 122 ⌀ | siblings listlengths 0 10k | sha stringlengths 40 40 | tags listlengths 2 1.82k | gguf stringclasses 789
values | config stringlengths 2 53.7k ⌀ | transformers_info dict | card_data stringlengths 234 905k ⌀ | card stringlengths 0 638k ⌀ | spaces null | licenses listlengths 1 3 ⌀ | datasets listlengths 1 289 ⌀ | languages listlengths 1 1.81k ⌀ | safetensors_params float64 0 1,019B ⌀ | gguf_params float64 0 122B ⌀ | metrics listlengths 1 15 ⌀ | architectures listlengths 1 6 ⌀ | tasks listlengths 1 6 ⌀ | modalities listlengths 1 4 ⌀ | input_modalities listlengths 1 3 ⌀ | output_modalities listlengths 1 4 ⌀ |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
6698d8a0653e4babe21e1e7d | meta-llama/Llama-3.1-8B-Instruct | meta-llama | {
"models": [
{
"_id": "66944f1fe0c5c2e493a804f5",
"id": "meta-llama/Llama-3.1-8B"
}
],
"relation": "finetune"
} | 8,384,864 | 133,735,004 | manual | 2024-07-18T08:56:00Z | 2024-09-25T17:00:57Z | transformers | 5,631 | 47 | null | text-generation | {"parameters": {"BF16": 8030261248}, "total": 8030261248} | [
".gitattributes",
"LICENSE",
"README.md",
"USE_POLICY.md",
"config.json",
"generation_config.json",
"model-00001-of-00004.safetensors",
"model-00002-of-00004.safetensors",
"model-00003-of-00004.safetensors",
"model-00004-of-00004.safetensors",
"model.safetensors.index.json",
"original/consolid... | 0e9e39f249a16976918f6564b8830bc894c89659 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"facebook",
"meta",
"pytorch",
"llama-3",
"conversational",
"en",
"de",
"fr",
"it",
"pt",
"hi",
"es",
"th",
"arxiv:2204.05149",
"base_model:meta-llama/Llama-3.1-8B",
"base_model:finetune:meta-llama/Llama-3.1-8B",
"lic... | null | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{{- bos_token }}\n{%- if custom_tools is defined %}\n {%- set tools = custom_tools %}\n{%- endif %}\n{%- if not tools_in_user_message is defined %}\n {%- set tools_in_user_messag... | {
"auto_model": "AutoModelForCausalLM",
"custom_class": null,
"pipeline_tag": "text-generation",
"processor": "AutoTokenizer"
} | {"base_model": "meta-llama/Meta-Llama-3.1-8B", "datasets": null, "eval_results": null, "language": ["en", "de", "fr", "it", "pt", "hi", "es", "th"], "library_name": null, "license": "llama3.1", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "text-generation", "tags": ["... | null | null | [
"llama3.1"
] | null | [
"en",
"de",
"fr",
"it",
"pt",
"hi",
"es",
"th"
] | 8,030,261,248 | null | null | [
"AutoModelForCausalLM",
"llama",
"LlamaForCausalLM"
] | [
"text-generation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
654a84cadff2f49007ce6c37 | openai/whisper-large-v3 | openai | null | 4,832,162 | 115,357,760 | False | 2023-11-07T18:41:14Z | 2024-08-12T10:20:10Z | transformers | 5,522 | 30 | null | automatic-speech-recognition | null | [
".gitattributes",
"README.md",
"added_tokens.json",
"config.json",
"flax_model.msgpack",
"generation_config.json",
"merges.txt",
"model.fp32-00001-of-00002.safetensors",
"model.fp32-00002-of-00002.safetensors",
"model.safetensors",
"model.safetensors.index.fp32.json",
"normalizer.json",
"pre... | 06f233fe06e710322aca913c1bc4249a0d71fce1 | [
"transformers",
"pytorch",
"jax",
"safetensors",
"whisper",
"automatic-speech-recognition",
"audio",
"hf-asr-leaderboard",
"en",
"zh",
"de",
"es",
"ru",
"ko",
"fr",
"ja",
"pt",
"tr",
"pl",
"ca",
"nl",
"ar",
"sv",
"it",
"id",
"hi",
"fi",
"vi",
"he",
"uk",
"... | null | {"architectures": ["WhisperForConditionalGeneration"], "model_type": "whisper", "tokenizer_config": {"bos_token": "<|endoftext|>", "eos_token": "<|endoftext|>", "pad_token": "<|endoftext|>", "unk_token": "<|endoftext|>"}} | {
"auto_model": "AutoModelForSpeechSeq2Seq",
"custom_class": null,
"pipeline_tag": "automatic-speech-recognition",
"processor": "AutoProcessor"
} | {"base_model": null, "datasets": null, "eval_results": null, "language": ["en", "zh", "de", "es", "ru", "ko", "fr", "ja", "pt", "tr", "pl", "ca", "nl", "ar", "sv", "it", "id", "hi", "fi", "vi", "he", "uk", "el", "ms", "cs", "ro", "da", "hu", "ta", "no", "th", "ur", "hr", "bg", "lt", "la", "mi", "ml", "cy", "sk", "te", ... | # Whisper
Whisper is a state-of-the-art model for automatic speech recognition (ASR) and speech translation, proposed in the paper
[Robust Speech Recognition via Large-Scale Weak Supervision](https://huggingface.co/papers/2212.04356) by Alec Radford
et al. from OpenAI. Trained on >5M hours of labeled data, Whisper d... | null | [
"apache-2.0"
] | null | [
"en",
"zh",
"de",
"es",
"ru",
"ko",
"fr",
"ja",
"pt",
"tr",
"pl",
"ca",
"nl",
"ar",
"sv",
"it",
"id",
"hi",
"fi",
"vi",
"he",
"uk",
"el",
"ms",
"cs",
"ro",
"da",
"hu",
"ta",
"no",
"th",
"ur",
"hr",
"bg",
"lt",
"la",
"mi",
"ml",
"cy",
"sk"... | null | null | null | [
"WhisperForConditionalGeneration",
"whisper",
"AutoModelForSpeechSeq2Seq"
] | [
"automatic-speech-recognition"
] | [
"multimodal"
] | [
"audio"
] | [
"text"
] |
68913522f16f3c8aaffccf1f | openai/gpt-oss-120b | openai | null | 4,363,728 | 29,644,344 | False | 2025-08-04T22:33:06Z | 2025-08-26T17:25:03Z | transformers | 4,621 | 23 | null | text-generation | {"parameters": {"BF16": 2167371072, "U8": 118244966400}, "total": 120412337472} | [
".gitattributes",
"LICENSE",
"README.md",
"USAGE_POLICY",
"chat_template.jinja",
"config.json",
"generation_config.json",
"metal/model.bin",
"model-00000-of-00014.safetensors",
"model-00001-of-00014.safetensors",
"model-00002-of-00014.safetensors",
"model-00003-of-00014.safetensors",
"model-... | b5c939de8f754692c1647ca79fbf85e8c1e70f8a | [
"transformers",
"safetensors",
"gpt_oss",
"text-generation",
"vllm",
"conversational",
"arxiv:2508.10925",
"license:apache-2.0",
"eval-results",
"endpoints_compatible",
"8-bit",
"mxfp4",
"deploy:azure",
"region:us"
] | null | {"architectures": ["GptOssForCausalLM"], "model_type": "gpt_oss", "quantization_config": {"quant_method": "mxfp4"}, "tokenizer_config": {"bos_token": "<|startoftext|>", "eos_token": "<|return|>", "pad_token": "<|endoftext|>"}, "chat_template_jinja": "{#-\n In addition to the normal inputs of `messages` and `tools`, th... | {
"auto_model": "AutoModelForCausalLM",
"custom_class": null,
"pipeline_tag": "text-generation",
"processor": "AutoTokenizer"
} | {"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "apache-2.0", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "text-generation", "tags": ["vllm"]} | <p align="center">
<img alt="gpt-oss-120b" src="https://raw.githubusercontent.com/openai/gpt-oss/main/docs/gpt-oss-120b.svg">
</p>
<p align="center">
<a href="https://gpt-oss.com"><strong>Try gpt-oss</strong></a> ·
<a href="https://cookbook.openai.com/topic/gpt-oss"><strong>Guides</strong></a> ·
<a href="https... | null | [
"apache-2.0"
] | null | null | 120,412,337,472 | null | null | [
"AutoModelForCausalLM",
"GptOssForCausalLM",
"gpt_oss"
] | [
"text-generation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
621ffdc036468d709f17434d | openai-community/gpt2 | openai-community | null | 11,636,493 | 836,599,686 | False | 2022-03-02T23:29:04Z | 2024-02-19T10:57:45Z | transformers | 3,153 | 22 | null | text-generation | {"parameters": {"F32": 137022720}, "total": 137022720} | [
".gitattributes",
"64-8bits.tflite",
"64-fp16.tflite",
"64.tflite",
"README.md",
"config.json",
"flax_model.msgpack",
"generation_config.json",
"merges.txt",
"model.safetensors",
"onnx/config.json",
"onnx/decoder_model.onnx",
"onnx/decoder_model_merged.onnx",
"onnx/decoder_with_past_model.... | 607a30d783dfa663caf39e06633721c8d4cfcd7e | [
"transformers",
"pytorch",
"tf",
"jax",
"tflite",
"rust",
"onnx",
"safetensors",
"gpt2",
"text-generation",
"exbert",
"en",
"doi:10.57967/hf/0039",
"license:mit",
"text-generation-inference",
"endpoints_compatible",
"deploy:azure",
"region:us"
] | null | {"architectures": ["GPT2LMHeadModel"], "model_type": "gpt2", "tokenizer_config": {}} | {
"auto_model": "AutoModelForCausalLM",
"custom_class": null,
"pipeline_tag": "text-generation",
"processor": "AutoTokenizer"
} | {"base_model": null, "datasets": null, "eval_results": null, "language": "en", "library_name": null, "license": "mit", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": null, "tags": ["exbert"]} | # GPT-2
Test the whole generation capabilities here: https://transformer.huggingface.co/doc/gpt2-large
Pretrained model on English language using a causal language modeling (CLM) objective. It was introduced in
[this paper](https://d4mucfpksywv.cloudfront.net/better-language-models/language_models_are_unsupervised_mu... | null | [
"mit"
] | null | [
"en"
] | 137,022,720 | null | null | [
"GPT2LMHeadModel",
"AutoModelForCausalLM",
"gpt2"
] | [
"text-generation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
66e81cbcd683a3f4e5291bbf | Qwen/Qwen2.5-7B-Instruct | Qwen | {
"models": [
{
"_id": "66e6d06401e2adb00b0be996",
"id": "Qwen/Qwen2.5-7B"
}
],
"relation": "finetune"
} | 17,917,551 | 109,933,464 | False | 2024-09-16T11:55:40Z | 2025-01-12T02:10:10Z | transformers | 1,165 | 22 | null | text-generation | {"parameters": {"BF16": 7615616512}, "total": 7615616512} | [
".gitattributes",
"LICENSE",
"README.md",
"config.json",
"generation_config.json",
"merges.txt",
"model-00001-of-00004.safetensors",
"model-00002-of-00004.safetensors",
"model-00003-of-00004.safetensors",
"model-00004-of-00004.safetensors",
"model.safetensors.index.json",
"tokenizer.json",
"... | a09a35458c702b33eeacc393d103063234e8bc28 | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"chat",
"conversational",
"en",
"arxiv:2309.00071",
"arxiv:2407.10671",
"base_model:Qwen/Qwen2.5-7B",
"base_model:finetune:Qwen/Qwen2.5-7B",
"license:apache-2.0",
"text-generation-inference",
"endpoints_compatible",
"deploy:azure... | null | {"architectures": ["Qwen2ForCausalLM"], "model_type": "qwen2", "tokenizer_config": {"bos_token": null, "chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0]['role'] == 'system' %}\n {{- messages[0]['content'] }}\n {%- else %}\n {{- 'You are Qwen, created by Aliba... | {
"auto_model": "AutoModelForCausalLM",
"custom_class": null,
"pipeline_tag": "text-generation",
"processor": "AutoTokenizer"
} | {"base_model": "Qwen/Qwen2.5-7B", "datasets": null, "eval_results": null, "language": ["en"], "library_name": "transformers", "license": "apache-2.0", "license_name": null, "license_link": "https://huggingface.co/Qwen/Qwen2.5-7B-Instruct/blob/main/LICENSE", "metrics": null, "model_name": null, "pipeline_tag": "text-gen... | # Qwen2.5-7B-Instruct
<a href="https://chat.qwenlm.ai/" target="_blank" style="margin: 2px;">
<img alt="Chat" src="https://img.shields.io/badge/%F0%9F%92%9C%EF%B8%8F%20Qwen%20Chat%20-536af5" style="display: inline-block; vertical-align: middle;"/>
</a>
## Introduction
Qwen2.5 is the latest series of Qwen large la... | null | [
"apache-2.0",
"https://huggingface.co/Qwen/Qwen2.5-7B-Instruct/blob/main/LICENSE"
] | null | [
"en"
] | 7,615,616,512 | null | null | [
"AutoModelForCausalLM",
"Qwen2ForCausalLM",
"qwen2"
] | [
"text-generation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
67b79c8700245b72c5706777 | google/gemma-3-4b-it | google | {
"models": [
{
"_id": "67b79c6c01ad68cfed14677a",
"id": "google/gemma-3-4b-pt"
}
],
"relation": "finetune"
} | 1,702,746 | 15,276,165 | manual | 2025-02-20T21:20:07Z | 2025-03-21T20:20:53Z | transformers | 1,265 | 22 | null | image-text-to-text | null | [
".gitattributes",
"README.md",
"added_tokens.json",
"chat_template.json",
"config.json",
"generation_config.json",
"model-00001-of-00002.safetensors",
"model-00002-of-00002.safetensors",
"model.safetensors.index.json",
"preprocessor_config.json",
"processor_config.json",
"special_tokens_map.js... | 093f9f388b31de276ce2de164bdc2081324b9767 | [
"transformers",
"safetensors",
"gemma3",
"image-text-to-text",
"conversational",
"arxiv:1905.07830",
"arxiv:1905.10044",
"arxiv:1911.11641",
"arxiv:1904.09728",
"arxiv:1705.03551",
"arxiv:1911.01547",
"arxiv:1907.10641",
"arxiv:1903.00161",
"arxiv:2009.03300",
"arxiv:2304.06364",
"arxi... | null | {"architectures": ["Gemma3ForConditionalGeneration"], "model_type": "gemma3", "processor_config": {"chat_template": "{{ bos_token }}\n{%- if messages[0]['role'] == 'system' -%}\n {%- if messages[0]['content'] is string -%}\n {%- set first_user_prefix = messages[0]['content'] + '\n\n' -%}\n {%- else -%}\n ... | {
"auto_model": "AutoModelForImageTextToText",
"custom_class": null,
"pipeline_tag": "image-text-to-text",
"processor": "AutoProcessor"
} | {"base_model": "google/gemma-3-4b-pt", "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "gemma", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "image-text-to-text", "tags": null, "extra_gated_heading": "Access Gemma o... | null | null | [
"gemma"
] | null | null | null | null | null | [
"AutoModelForImageTextToText",
"Gemma3ForConditionalGeneration",
"gemma3"
] | [
"image-text-to-text"
] | [
"multimodal"
] | [
"text",
"image"
] | [
"text"
] |
621ffdc136468d709f180294 | sentence-transformers/all-MiniLM-L6-v2 | sentence-transformers | null | 206,073,068 | 2,406,656,367 | False | 2022-03-02T23:29:05Z | 2025-03-06T13:37:44Z | sentence-transformers | 4,620 | 21 | null | sentence-similarity | {"parameters": {"I64": 512, "F32": 22713216}, "total": 22713728} | [
".gitattributes",
"1_Pooling/config.json",
"README.md",
"config.json",
"config_sentence_transformers.json",
"data_config.json",
"model.safetensors",
"modules.json",
"onnx/model.onnx",
"onnx/model_O1.onnx",
"onnx/model_O2.onnx",
"onnx/model_O3.onnx",
"onnx/model_O4.onnx",
"onnx/model_qint8_... | c9745ed1d9f207416be6d2e6f8de32d1f16199bf | [
"sentence-transformers",
"pytorch",
"tf",
"rust",
"onnx",
"safetensors",
"openvino",
"bert",
"feature-extraction",
"sentence-similarity",
"transformers",
"en",
"dataset:s2orc",
"dataset:flax-sentence-embeddings/stackexchange_xml",
"dataset:ms_marco",
"dataset:gooaq",
"dataset:yahoo_a... | null | {"architectures": ["BertModel"], "model_type": "bert", "tokenizer_config": {"unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]"}} | {
"auto_model": "AutoModel",
"custom_class": null,
"pipeline_tag": "feature-extraction",
"processor": "AutoTokenizer"
} | {"base_model": null, "datasets": ["s2orc", "flax-sentence-embeddings/stackexchange_xml", "ms_marco", "gooaq", "yahoo_answers_topics", "code_search_net", "search_qa", "eli5", "snli", "multi_nli", "wikihow", "natural_questions", "trivia_qa", "embedding-data/sentence-compression", "embedding-data/flickr30k-captions", "emb... | # all-MiniLM-L6-v2
This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 384 dimensional dense vector space and can be used for tasks like clustering or semantic search.
## Usage (Sentence-Transformers)
Using this model becomes easy when you have [sentence-transformers](ht... | null | [
"apache-2.0"
] | [
"s2orc",
"flax-sentence-embeddings/stackexchange_xml",
"ms_marco",
"gooaq",
"yahoo_answers_topics",
"code_search_net",
"search_qa",
"eli5",
"snli",
"multi_nli",
"wikihow",
"natural_questions",
"trivia_qa",
"embedding-data/sentence-compression",
"embedding-data/flickr30k-captions",
"emb... | [
"en"
] | 22,713,728 | null | null | [
"BertModel",
"AutoModel",
"bert"
] | [
"sentence-similarity",
"feature-extraction"
] | [
"text",
"multimodal"
] | [
"text"
] | [
"logits",
"embeddings"
] |
66944f1fe0c5c2e493a804f5 | meta-llama/Llama-3.1-8B | meta-llama | null | 1,387,690 | 20,616,230 | manual | 2024-07-14T22:20:15Z | 2024-10-16T22:00:37Z | transformers | 2,128 | 16 | null | text-generation | {"parameters": {"BF16": 8030261248}, "total": 8030261248} | [
".gitattributes",
"LICENSE",
"README.md",
"USE_POLICY.md",
"config.json",
"generation_config.json",
"model-00001-of-00004.safetensors",
"model-00002-of-00004.safetensors",
"model-00003-of-00004.safetensors",
"model-00004-of-00004.safetensors",
"model.safetensors.index.json",
"original/consolid... | d04e592bb4f6aa9cfee91e2e20afa771667e1d4b | [
"transformers",
"safetensors",
"llama",
"text-generation",
"facebook",
"meta",
"pytorch",
"llama-3",
"en",
"de",
"fr",
"it",
"pt",
"hi",
"es",
"th",
"arxiv:2204.05149",
"license:llama3.1",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | null | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "eos_token": "<|end_of_text|>"}} | {
"auto_model": "AutoModelForCausalLM",
"custom_class": null,
"pipeline_tag": "text-generation",
"processor": "AutoTokenizer"
} | {"base_model": null, "datasets": null, "eval_results": null, "language": ["en", "de", "fr", "it", "pt", "hi", "es", "th"], "library_name": "transformers", "license": "llama3.1", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "text-generation", "tags": ["facebook", "meta... | null | null | [
"llama3.1"
] | null | [
"en",
"de",
"fr",
"it",
"pt",
"hi",
"es",
"th"
] | 8,030,261,248 | null | null | [
"AutoModelForCausalLM",
"llama",
"LlamaForCausalLM"
] | [
"text-generation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
67fddfa9a7fe1f21ec1d3026 | microsoft/bitnet-b1.58-2B-4T | microsoft | null | 16,753 | 190,249 | False | 2025-04-15T04:25:13Z | 2025-12-17T18:13:05Z | transformers | 1,404 | 15 | null | text-generation | {"parameters": {"BF16": 328775890, "U8": 521011200}, "total": 849787090} | [
".gitattributes",
"LICENSE",
"README.md",
"config.json",
"data_summary_card.md",
"generation_config.json",
"model.safetensors",
"special_tokens_map.json",
"tokenizer.json",
"tokenizer_config.json"
] | 04c3b9ad9361b824064a1f25ea60a8be9599b127 | [
"transformers",
"safetensors",
"bitnet",
"text-generation",
"chat",
"large-language-model",
"conversational",
"custom_code",
"en",
"arxiv:2504.12285",
"license:mit",
"endpoints_compatible",
"8-bit",
"deploy:azure",
"region:us"
] | null | {"architectures": ["BitNetForCausalLM"], "auto_map": {"AutoConfig": "configuration_bitnet.BitNetConfig", "AutoModelForCausalLM": "modeling_bitnet.BitNetForCausalLM"}, "model_type": "bitnet", "quantization_config": {"quant_method": "bitnet"}, "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% se... | {
"auto_model": "AutoModelForCausalLM",
"custom_class": null,
"pipeline_tag": "text-generation",
"processor": "AutoTokenizer"
} | {"base_model": null, "datasets": null, "eval_results": null, "language": ["en"], "library_name": "transformers", "license": "mit", "license_name": null, "license_link": "https://huggingface.co/microsoft/bitnet-b1.58-2B-4T/blob/main/LICENSE", "metrics": null, "model_name": null, "pipeline_tag": "text-generation", "tags"... | # BitNet b1.58 2B4T - Scaling Native 1-bit LLM
This repository contains the weights for **BitNet b1.58 2B4T**, the first open-source, native 1-bit Large Language Model (LLM) at the 2-billion parameter scale, developed by Microsoft Research.
Trained on a corpus of 4 trillion tokens, this model demonstrates that native... | null | [
"mit",
"https://huggingface.co/microsoft/bitnet-b1.58-2B-4T/blob/main/LICENSE"
] | null | [
"en"
] | 849,787,090 | null | null | [
"AutoModelForCausalLM",
"BitNetForCausalLM",
"bitnet"
] | [
"text-generation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
680da79d48c19d19cf82ab5e | Qwen/Qwen3-8B | Qwen | {
"models": [
{
"_id": "680f0c3e6eacf22a04ee1219",
"id": "Qwen/Qwen3-8B-Base"
}
],
"relation": "finetune"
} | 9,481,229 | 43,490,001 | False | 2025-04-27T03:42:21Z | 2025-07-26T03:49:13Z | transformers | 1,013 | 15 | null | text-generation | {"parameters": {"BF16": 8190735360}, "total": 8190735360} | [
".gitattributes",
"LICENSE",
"README.md",
"config.json",
"generation_config.json",
"merges.txt",
"model-00001-of-00005.safetensors",
"model-00002-of-00005.safetensors",
"model-00003-of-00005.safetensors",
"model-00004-of-00005.safetensors",
"model-00005-of-00005.safetensors",
"model.safetensor... | b968826d9c46dd6066d109eabc6255188de91218 | [
"transformers",
"safetensors",
"qwen3",
"text-generation",
"conversational",
"arxiv:2309.00071",
"arxiv:2505.09388",
"base_model:Qwen/Qwen3-8B-Base",
"base_model:finetune:Qwen/Qwen3-8B-Base",
"license:apache-2.0",
"text-generation-inference",
"endpoints_compatible",
"deploy:azure",
"region... | null | {"architectures": ["Qwen3ForCausalLM"], "model_type": "qwen3", "tokenizer_config": {"bos_token": null, "chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0].role == 'system' %}\n {{- messages[0].content + '\\n\\n' }}\n {%- endif %}\n {{- \"# Tools\\n\\nYou may call o... | {
"auto_model": "AutoModelForCausalLM",
"custom_class": null,
"pipeline_tag": "text-generation",
"processor": "AutoTokenizer"
} | {"base_model": ["Qwen/Qwen3-8B-Base"], "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "apache-2.0", "license_name": null, "license_link": "https://huggingface.co/Qwen/Qwen3-8B/blob/main/LICENSE", "metrics": null, "model_name": null, "pipeline_tag": "text-generation"... | null | null | [
"apache-2.0",
"https://huggingface.co/Qwen/Qwen3-8B/blob/main/LICENSE"
] | null | null | 8,190,735,360 | null | null | [
"AutoModelForCausalLM",
"Qwen3ForCausalLM",
"qwen3"
] | [
"text-generation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
683f05dc98de733bf4c59f3d | Qwen/Qwen3-Embedding-0.6B | Qwen | {
"models": [
{
"_id": "680f0da2f99a20ddfc9a9f26",
"id": "Qwen/Qwen3-0.6B-Base"
}
],
"relation": "finetune"
} | 5,507,791 | 36,810,259 | False | 2025-06-03T14:25:32Z | 2025-06-20T09:31:05Z | sentence-transformers | 950 | 15 | null | feature-extraction | {"parameters": {"BF16": 595776512}, "total": 595776512} | [
".gitattributes",
"1_Pooling/config.json",
"README.md",
"config.json",
"config_sentence_transformers.json",
"generation_config.json",
"merges.txt",
"model.safetensors",
"modules.json",
"tokenizer.json",
"tokenizer_config.json",
"vocab.json"
] | c54f2e6e80b2d7b7de06f51cec4959f6b3e03418 | [
"sentence-transformers",
"safetensors",
"qwen3",
"text-generation",
"transformers",
"sentence-similarity",
"feature-extraction",
"text-embeddings-inference",
"arxiv:2506.05176",
"base_model:Qwen/Qwen3-0.6B-Base",
"base_model:finetune:Qwen/Qwen3-0.6B-Base",
"license:apache-2.0",
"endpoints_co... | null | {"architectures": ["Qwen3ForCausalLM"], "model_type": "qwen3", "tokenizer_config": {"bos_token": null, "chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0].role == 'system' %}\n {{- messages[0].content + '\\n\\n' }}\n {%- endif %}\n {{- \"# Tools\\n\\nYou may call o... | {
"auto_model": "AutoModelForCausalLM",
"custom_class": null,
"pipeline_tag": "text-generation",
"processor": "AutoTokenizer"
} | {"base_model": ["Qwen/Qwen3-0.6B-Base"], "datasets": null, "eval_results": null, "language": null, "library_name": null, "license": "apache-2.0", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": null, "tags": ["transformers", "sentence-transformers", "sentence-similarity"... | # Qwen3-Embedding-0.6B
<p align="center">
<img src="https://qianwen-res.oss-accelerate-overseas.aliyuncs.com/logo_qwen3.png" width="400"/>
<p>
## Highlights
The Qwen3 Embedding model series is the latest proprietary model of the Qwen family, specifically designed for text embedding and ranking tasks. Building up... | null | [
"apache-2.0"
] | null | null | 595,776,512 | null | null | [
"AutoModelForCausalLM",
"Qwen3ForCausalLM",
"qwen3"
] | [
"sentence-similarity",
"feature-extraction",
"text-generation"
] | [
"text",
"multimodal"
] | [
"text"
] | [
"logits",
"text",
"embeddings"
] |
68913539bd3d0a833438591d | openai/gpt-oss-20b | openai | null | 6,734,875 | 54,662,871 | False | 2025-08-04T22:33:29Z | 2025-08-26T17:25:47Z | transformers | 4,485 | 15 | null | text-generation | {"parameters": {"BF16": 1804459584, "U8": 19707494400}, "total": 21511953984} | [
".gitattributes",
"LICENSE",
"README.md",
"USAGE_POLICY",
"chat_template.jinja",
"config.json",
"generation_config.json",
"metal/model.bin",
"model-00000-of-00002.safetensors",
"model-00001-of-00002.safetensors",
"model-00002-of-00002.safetensors",
"model.safetensors.index.json",
"original/c... | 6cee5e81ee83917806bbde320786a8fb61efebee | [
"transformers",
"safetensors",
"gpt_oss",
"text-generation",
"vllm",
"conversational",
"arxiv:2508.10925",
"license:apache-2.0",
"eval-results",
"endpoints_compatible",
"8-bit",
"mxfp4",
"deploy:azure",
"region:us"
] | null | {"architectures": ["GptOssForCausalLM"], "model_type": "gpt_oss", "quantization_config": {"quant_method": "mxfp4"}, "tokenizer_config": {"bos_token": "<|startoftext|>", "eos_token": "<|return|>", "pad_token": "<|endoftext|>"}, "chat_template_jinja": "{#-\n In addition to the normal inputs of `messages` and `tools`, th... | {
"auto_model": "AutoModelForCausalLM",
"custom_class": null,
"pipeline_tag": "text-generation",
"processor": "AutoTokenizer"
} | {"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "apache-2.0", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "text-generation", "tags": ["vllm"]} | <p align="center">
<img alt="gpt-oss-20b" src="https://raw.githubusercontent.com/openai/gpt-oss/main/docs/gpt-oss-20b.svg">
</p>
<p align="center">
<a href="https://gpt-oss.com"><strong>Try gpt-oss</strong></a> ·
<a href="https://cookbook.openai.com/topic/gpt-oss"><strong>Guides</strong></a> ·
<a href="https:/... | null | [
"apache-2.0"
] | null | null | 21,511,953,984 | null | null | [
"AutoModelForCausalLM",
"GptOssForCausalLM",
"gpt_oss"
] | [
"text-generation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
680da718233834890aa01f51 | Qwen/Qwen3-0.6B | Qwen | {
"models": [
{
"_id": "680f0da2f99a20ddfc9a9f26",
"id": "Qwen/Qwen3-0.6B-Base"
}
],
"relation": "finetune"
} | 13,807,272 | 74,174,117 | False | 2025-04-27T03:40:08Z | 2025-07-26T03:46:27Z | transformers | 1,160 | 13 | null | text-generation | {"parameters": {"BF16": 751632384}, "total": 751632384} | [
".gitattributes",
"LICENSE",
"README.md",
"config.json",
"generation_config.json",
"merges.txt",
"model.safetensors",
"tokenizer.json",
"tokenizer_config.json",
"vocab.json"
] | c1899de289a04d12100db370d81485cdf75e47ca | [
"transformers",
"safetensors",
"qwen3",
"text-generation",
"conversational",
"arxiv:2505.09388",
"base_model:Qwen/Qwen3-0.6B-Base",
"base_model:finetune:Qwen/Qwen3-0.6B-Base",
"license:apache-2.0",
"text-generation-inference",
"endpoints_compatible",
"deploy:azure",
"region:us"
] | null | {"architectures": ["Qwen3ForCausalLM"], "model_type": "qwen3", "tokenizer_config": {"bos_token": null, "chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0].role == 'system' %}\n {{- messages[0].content + '\\n\\n' }}\n {%- endif %}\n {{- \"# Tools\\n\\nYou may call o... | {
"auto_model": "AutoModelForCausalLM",
"custom_class": null,
"pipeline_tag": "text-generation",
"processor": "AutoTokenizer"
} | {"base_model": ["Qwen/Qwen3-0.6B-Base"], "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "apache-2.0", "license_name": null, "license_link": "https://huggingface.co/Qwen/Qwen3-0.6B/blob/main/LICENSE", "metrics": null, "model_name": null, "pipeline_tag": "text-generat... | null | null | [
"apache-2.0",
"https://huggingface.co/Qwen/Qwen3-0.6B/blob/main/LICENSE"
] | null | null | 751,632,384 | null | null | [
"AutoModelForCausalLM",
"Qwen3ForCausalLM",
"qwen3"
] | [
"text-generation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
678dc6fff905d106be796d8a | deepseek-ai/DeepSeek-R1 | deepseek-ai | null | 2,203,466 | 17,068,124 | False | 2025-01-20T03:46:07Z | 2025-03-27T04:01:59Z | transformers | 13,107 | 12 | null | text-generation | {"parameters": {"BF16": 3918786560, "F8_E4M3": 680571043840, "F32": 41555600}, "total": 684531386000} | [
".gitattributes",
"LICENSE",
"README.md",
"config.json",
"configuration_deepseek.py",
"figures/benchmark.jpg",
"generation_config.json",
"model-00001-of-000163.safetensors",
"model-00002-of-000163.safetensors",
"model-00003-of-000163.safetensors",
"model-00004-of-000163.safetensors",
"model-00... | 56d4cbbb4d29f4355bab4b9a39ccb717a14ad5ad | [
"transformers",
"safetensors",
"deepseek_v3",
"text-generation",
"conversational",
"custom_code",
"arxiv:2501.12948",
"license:mit",
"eval-results",
"text-generation-inference",
"endpoints_compatible",
"fp8",
"region:us"
] | null | {"architectures": ["DeepseekV3ForCausalLM"], "auto_map": {"AutoConfig": "configuration_deepseek.DeepseekV3Config", "AutoModel": "modeling_deepseek.DeepseekV3Model", "AutoModelForCausalLM": "modeling_deepseek.DeepseekV3ForCausalLM"}, "model_type": "deepseek_v3", "quantization_config": {"quant_method": "fp8"}, "tokenizer... | {
"auto_model": "AutoModelForCausalLM",
"custom_class": null,
"pipeline_tag": "text-generation",
"processor": "AutoTokenizer"
} | {"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "mit", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": null, "tags": null} | # DeepSeek-R1
<!-- markdownlint-disable first-line-h1 -->
<!-- markdownlint-disable html -->
<!-- markdownlint-disable no-duplicate-header -->
<div align="center">
<img src="https://github.com/deepseek-ai/DeepSeek-V2/blob/main/figures/logo.svg?raw=true" width="60%" alt="DeepSeek-V3" />
</div>
<hr>
<div align="center... | null | [
"mit"
] | null | null | 684,531,386,000 | null | null | [
"AutoModelForCausalLM",
"deepseek_v3",
"DeepseekV3ForCausalLM"
] | [
"text-generation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
6891e3bb084ce75acffb033d | Qwen/Qwen3-4B-Instruct-2507 | Qwen | null | 6,235,673 | 28,945,408 | False | 2025-08-05T10:58:03Z | 2025-09-17T06:56:53Z | transformers | 785 | 12 | null | text-generation | {"parameters": {"BF16": 4022468096}, "total": 4022468096} | [
".gitattributes",
"LICENSE",
"README.md",
"config.json",
"generation_config.json",
"merges.txt",
"model-00001-of-00003.safetensors",
"model-00002-of-00003.safetensors",
"model-00003-of-00003.safetensors",
"model.safetensors.index.json",
"tokenizer.json",
"tokenizer_config.json",
"vocab.json"... | cdbee75f17c01a7cc42f958dc650907174af0554 | [
"transformers",
"safetensors",
"qwen3",
"text-generation",
"conversational",
"arxiv:2505.09388",
"license:apache-2.0",
"eval-results",
"text-generation-inference",
"endpoints_compatible",
"deploy:azure",
"region:us"
] | null | {"architectures": ["Qwen3ForCausalLM"], "model_type": "qwen3", "tokenizer_config": {"bos_token": null, "chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0].role == 'system' %}\n {{- messages[0].content + '\\n\\n' }}\n {%- endif %}\n {{- \"# Tools\\n\\nYou may call o... | {
"auto_model": "AutoModelForCausalLM",
"custom_class": null,
"pipeline_tag": "text-generation",
"processor": "AutoTokenizer"
} | {"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "apache-2.0", "license_name": null, "license_link": "https://huggingface.co/Qwen/Qwen3-4B-Instruct-2507/blob/main/LICENSE", "metrics": null, "model_name": null, "pipeline_tag": "text-generation", "t... | # Qwen3-4B-Instruct-2507
<a href="https://chat.qwen.ai" target="_blank" style="margin: 2px;">
<img alt="Chat" src="https://img.shields.io/badge/%F0%9F%92%9C%EF%B8%8F%20Qwen%20Chat%20-536af5" style="display: inline-block; vertical-align: middle;"/>
</a>
## Highlights
We introduce the updated version of the **Qwen3... | null | [
"apache-2.0",
"https://huggingface.co/Qwen/Qwen3-4B-Instruct-2507/blob/main/LICENSE"
] | null | null | 4,022,468,096 | null | null | [
"AutoModelForCausalLM",
"Qwen3ForCausalLM",
"qwen3"
] | [
"text-generation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
68ea05fb43df37d95ad2491d | Qwen/Qwen3-VL-8B-Instruct | Qwen | null | 4,649,140 | 20,365,833 | False | 2025-10-11T07:23:39Z | 2025-10-15T16:16:59Z | transformers | 837 | 12 | null | image-text-to-text | {"parameters": {"BF16": 8767123696}, "total": 8767123696} | [
".gitattributes",
"README.md",
"chat_template.json",
"config.json",
"generation_config.json",
"merges.txt",
"model-00001-of-00004.safetensors",
"model-00002-of-00004.safetensors",
"model-00003-of-00004.safetensors",
"model-00004-of-00004.safetensors",
"model.safetensors.index.json",
"preproces... | 0c351dd01ed87e9c1b53cbc748cba10e6187ff3b | [
"transformers",
"safetensors",
"qwen3_vl",
"image-text-to-text",
"conversational",
"arxiv:2505.09388",
"arxiv:2502.13923",
"arxiv:2409.12191",
"arxiv:2308.12966",
"license:apache-2.0",
"endpoints_compatible",
"deploy:azure",
"region:us"
] | null | {"architectures": ["Qwen3VLForConditionalGeneration"], "model_type": "qwen3_vl", "processor_config": {"chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0].role == 'system' %}\n {%- if messages[0].content is string %}\n {{- messages[0].content }}\n {%- el... | {
"auto_model": "AutoModelForImageTextToText",
"custom_class": null,
"pipeline_tag": "image-text-to-text",
"processor": "AutoProcessor"
} | {"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "apache-2.0", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "image-text-to-text", "tags": null} | null | null | [
"apache-2.0"
] | null | null | 8,767,123,696 | null | null | [
"AutoModelForImageTextToText",
"qwen3_vl",
"Qwen3VLForConditionalGeneration"
] | [
"image-text-to-text"
] | [
"multimodal"
] | [
"text",
"image"
] | [
"text"
] |
680da769038897aa905586a7 | Qwen/Qwen3-4B | Qwen | {
"models": [
{
"_id": "680f0c5b4e1c1da6491f3b9d",
"id": "Qwen/Qwen3-4B-Base"
}
],
"relation": "finetune"
} | 6,491,995 | 30,446,215 | False | 2025-04-27T03:41:29Z | 2025-07-26T03:46:39Z | transformers | 582 | 11 | null | text-generation | null | [
".gitattributes",
"LICENSE",
"README.md",
"config.json",
"generation_config.json",
"merges.txt",
"model-00001-of-00003.safetensors",
"model-00002-of-00003.safetensors",
"model-00003-of-00003.safetensors",
"model.safetensors.index.json",
"tokenizer.json",
"tokenizer_config.json",
"vocab.json"... | 1cfa9a7208912126459214e8b04321603b3df60c | [
"transformers",
"safetensors",
"qwen3",
"text-generation",
"conversational",
"arxiv:2309.00071",
"arxiv:2505.09388",
"base_model:Qwen/Qwen3-4B-Base",
"base_model:finetune:Qwen/Qwen3-4B-Base",
"license:apache-2.0",
"text-generation-inference",
"endpoints_compatible",
"deploy:azure",
"region... | null | {"architectures": ["Qwen3ForCausalLM"], "model_type": "qwen3", "tokenizer_config": {"bos_token": null, "chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0].role == 'system' %}\n {{- messages[0].content + '\\n\\n' }}\n {%- endif %}\n {{- \"# Tools\\n\\nYou may call o... | {
"auto_model": "AutoModelForCausalLM",
"custom_class": null,
"pipeline_tag": "text-generation",
"processor": "AutoTokenizer"
} | {"base_model": ["Qwen/Qwen3-4B-Base"], "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "apache-2.0", "license_name": null, "license_link": "https://huggingface.co/Qwen/Qwen3-4B/blob/main/LICENSE", "metrics": null, "model_name": null, "pipeline_tag": "text-generation"... | null | null | [
"apache-2.0",
"https://huggingface.co/Qwen/Qwen3-4B/blob/main/LICENSE"
] | null | null | null | null | null | [
"AutoModelForCausalLM",
"Qwen3ForCausalLM",
"qwen3"
] | [
"text-generation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
683f090e77b41ac9fc98c51f | Qwen/Qwen3-Embedding-8B | Qwen | {
"models": [
{
"_id": "680f0c3e6eacf22a04ee1219",
"id": "Qwen/Qwen3-8B-Base"
}
],
"relation": "finetune"
} | 1,648,122 | 9,165,432 | False | 2025-06-03T14:39:10Z | 2025-07-07T09:02:21Z | sentence-transformers | 628 | 11 | null | feature-extraction | {"parameters": {"BF16": 7567295488}, "total": 7567295488} | [
".gitattributes",
"1_Pooling/config.json",
"LICENSE",
"README.md",
"config.json",
"config_sentence_transformers.json",
"generation_config.json",
"merges.txt",
"model-00001-of-00004.safetensors",
"model-00002-of-00004.safetensors",
"model-00003-of-00004.safetensors",
"model-00004-of-00004.safet... | 1d8ad4ca9b3dd8059ad90a75d4983776a23d44af | [
"sentence-transformers",
"safetensors",
"qwen3",
"text-generation",
"transformers",
"sentence-similarity",
"feature-extraction",
"text-embeddings-inference",
"arxiv:2506.05176",
"base_model:Qwen/Qwen3-8B-Base",
"base_model:finetune:Qwen/Qwen3-8B-Base",
"license:apache-2.0",
"endpoints_compat... | null | {"architectures": ["Qwen3ForCausalLM"], "model_type": "qwen3", "tokenizer_config": {"bos_token": null, "chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0]['role'] == 'system' %}\n {{- messages[0]['content'] }}\n {%- else %}\n {{- 'You are a helpful assistant.' ... | {
"auto_model": "AutoModelForCausalLM",
"custom_class": null,
"pipeline_tag": "text-generation",
"processor": "AutoTokenizer"
} | {"base_model": ["Qwen/Qwen3-8B-Base"], "datasets": null, "eval_results": null, "language": null, "library_name": null, "license": "apache-2.0", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": null, "tags": ["transformers", "sentence-transformers", "sentence-similarity", ... | null | null | [
"apache-2.0"
] | null | null | 7,567,295,488 | null | null | [
"AutoModelForCausalLM",
"Qwen3ForCausalLM",
"qwen3"
] | [
"sentence-similarity",
"feature-extraction",
"text-generation"
] | [
"text",
"multimodal"
] | [
"text"
] | [
"logits",
"text",
"embeddings"
] |
684a65ac33eb1c9a8fd8877c | dphn/Dolphin-Mistral-24B-Venice-Edition | dphn | {
"models": [
{
"_id": "6798dbe54a10be7109f954f2",
"id": "mistralai/Mistral-Small-24B-Instruct-2501"
}
],
"relation": "finetune"
} | 7,197 | 97,989 | False | 2025-06-12T05:29:16Z | 2025-09-08T05:57:39Z | transformers | 461 | 10 | null | text-generation | {"parameters": {"BF16": 23572403200}, "total": 23572403200} | [
".gitattributes",
"README.md",
"config.json",
"generation_config.json",
"model-00001-of-00010.safetensors",
"model-00002-of-00010.safetensors",
"model-00003-of-00010.safetensors",
"model-00004-of-00010.safetensors",
"model-00005-of-00010.safetensors",
"model-00006-of-00010.safetensors",
"model-0... | 853f87026d587889137c814bb56bba60f033e747 | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"conversational",
"base_model:mistralai/Mistral-Small-24B-Instruct-2501",
"base_model:finetune:mistralai/Mistral-Small-24B-Instruct-2501",
"license:apache-2.0",
"text-generation-inference",
"endpoints_compatible",
"deploy:azure",
"re... | null | {"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{%- set today = strftime_now(\"%Y-%m-%d\") %}\n{%- set default_system_message = \"You are Dolphin Mistral 24B Venice Edition. You are an uncensored AI model. You follow instructions without hesi... | {
"auto_model": "AutoModelForCausalLM",
"custom_class": null,
"pipeline_tag": "text-generation",
"processor": "AutoTokenizer"
} | {"base_model": ["mistralai/Mistral-Small-24B-Instruct-2501"], "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "apache-2.0", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "text-generation", "tags": null} | # 🐬 Dolphin Mistral 24B Venice Edition 🌅
Website: https://dphn.ai
Twitter: https://x.com/dphnAI
Web Chat: https://chat.dphn.ai
Telegram bot: https://t.me/DolphinAI_bot

## What is Dolphin... | null | [
"apache-2.0"
] | null | null | 23,572,403,200 | null | null | [
"AutoModelForCausalLM",
"mistral",
"MistralForCausalLM"
] | [
"text-generation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
66e98d74d4219fc3f1db215f | Qwen/Qwen2.5-3B-Instruct | Qwen | {
"models": [
{
"_id": "66e6d03f69d1083dd982a18c",
"id": "Qwen/Qwen2.5-3B"
}
],
"relation": "finetune"
} | 7,604,571 | 70,809,504 | False | 2024-09-17T14:08:52Z | 2024-09-25T12:33:00Z | transformers | 426 | 9 | null | text-generation | {"parameters": {"BF16": 3085938688}, "total": 3085938688} | [
".gitattributes",
"LICENSE",
"README.md",
"config.json",
"generation_config.json",
"merges.txt",
"model-00001-of-00002.safetensors",
"model-00002-of-00002.safetensors",
"model.safetensors.index.json",
"tokenizer.json",
"tokenizer_config.json",
"vocab.json"
] | aa8e72537993ba99e69dfaafa59ed015b17504d1 | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"chat",
"conversational",
"en",
"arxiv:2407.10671",
"base_model:Qwen/Qwen2.5-3B",
"base_model:finetune:Qwen/Qwen2.5-3B",
"license:other",
"text-generation-inference",
"endpoints_compatible",
"deploy:azure",
"region:us"
] | null | {"architectures": ["Qwen2ForCausalLM"], "model_type": "qwen2", "tokenizer_config": {"bos_token": null, "chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0]['role'] == 'system' %}\n {{- messages[0]['content'] }}\n {%- else %}\n {{- 'You are Qwen, created by Aliba... | {
"auto_model": "AutoModelForCausalLM",
"custom_class": null,
"pipeline_tag": "text-generation",
"processor": "AutoTokenizer"
} | {"base_model": "Qwen/Qwen2.5-3B", "datasets": null, "eval_results": null, "language": ["en"], "library_name": "transformers", "license": "other", "license_name": "qwen-research", "license_link": "https://huggingface.co/Qwen/Qwen2.5-3B-Instruct/blob/main/LICENSE", "metrics": null, "model_name": null, "pipeline_tag": "te... | # Qwen2.5-3B-Instruct
## Introduction
Qwen2.5 is the latest series of Qwen large language models. For Qwen2.5, we release a number of base language models and instruction-tuned language models ranging from 0.5 to 72 billion parameters. Qwen2.5 brings the following improvements upon Qwen2:
- Significantly **more know... | null | [
"other",
"qwen-research",
"https://huggingface.co/Qwen/Qwen2.5-3B-Instruct/blob/main/LICENSE"
] | null | [
"en"
] | 3,085,938,688 | null | null | [
"AutoModelForCausalLM",
"Qwen2ForCausalLM",
"qwen2"
] | [
"text-generation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
67c35b9bb236f0d365bf29d3 | google/gemma-3-27b-it | google | {
"models": [
{
"_id": "67c35b8b87a7f49a82593992",
"id": "google/gemma-3-27b-pt"
}
],
"relation": "finetune"
} | 1,078,578 | 12,381,516 | manual | 2025-03-01T19:10:19Z | 2025-03-21T20:29:02Z | transformers | 1,935 | 9 | null | image-text-to-text | {"parameters": {"BF16": 27432406640}, "total": 27432406640} | [
".gitattributes",
"README.md",
"added_tokens.json",
"chat_template.json",
"config.json",
"generation_config.json",
"model-00001-of-00012.safetensors",
"model-00002-of-00012.safetensors",
"model-00003-of-00012.safetensors",
"model-00004-of-00012.safetensors",
"model-00005-of-00012.safetensors",
... | 005ad3404e59d6023443cb575daa05336842228a | [
"transformers",
"safetensors",
"gemma3",
"image-text-to-text",
"conversational",
"arxiv:1905.07830",
"arxiv:1905.10044",
"arxiv:1911.11641",
"arxiv:1904.09728",
"arxiv:1705.03551",
"arxiv:1911.01547",
"arxiv:1907.10641",
"arxiv:1903.00161",
"arxiv:2009.03300",
"arxiv:2304.06364",
"arxi... | null | {"architectures": ["Gemma3ForConditionalGeneration"], "model_type": "gemma3", "processor_config": {"chat_template": "{{ bos_token }}\n{%- if messages[0]['role'] == 'system' -%}\n {%- if messages[0]['content'] is string -%}\n {%- set first_user_prefix = messages[0]['content'] + '\n\n' -%}\n {%- else -%}\n ... | {
"auto_model": "AutoModelForImageTextToText",
"custom_class": null,
"pipeline_tag": "image-text-to-text",
"processor": "AutoProcessor"
} | {"base_model": "google/gemma-3-27b-pt", "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "gemma", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "image-text-to-text", "tags": null, "extra_gated_heading": "Access Gemma ... | null | null | [
"gemma"
] | null | null | 27,432,406,640 | null | null | [
"AutoModelForImageTextToText",
"Gemma3ForConditionalGeneration",
"gemma3"
] | [
"image-text-to-text"
] | [
"multimodal"
] | [
"text",
"image"
] | [
"text"
] |
690088e0369d9d3b7594d73c | OpenMOSS-Team/RoboOmni | OpenMOSS-Team | null | 148 | 378 | False | 2025-10-28T09:12:00Z | 2025-10-30T06:53:06Z | transformers | 15 | 9 | null | robotics | null | [
".gitattributes",
"README.md",
"added_tokens.json",
"chat_template.jinja",
"config.json",
"generation_config.json",
"merges.txt",
"model.safetensors",
"model.safetensors.index.json",
"preprocessor_config.json",
"special_tokens_map.json",
"tokenizer.json",
"tokenizer_config.json",
"vocab.js... | 3a84313809c3b16fad1295a2e3814754bf179cfe | [
"transformers",
"safetensors",
"qwen2_5_omni_thinker",
"text-generation",
"robotics",
"en",
"dataset:fnlp/OmniAction",
"arxiv:2510.23763",
"license:cc-by-nc-4.0",
"endpoints_compatible",
"region:us"
] | null | {"architectures": ["Qwen2_5OmniThinkerForConditionalGeneration"], "model_type": "qwen2_5_omni_thinker", "tokenizer_config": {"bos_token": null, "eos_token": "<|im_end|>", "pad_token": "<|endoftext|>", "unk_token": null}, "chat_template_jinja": "{% set audio_count = namespace(value=0) %}{% set image_count = namespace(va... | {
"auto_model": "AutoModelForSeq2SeqLM",
"custom_class": null,
"pipeline_tag": "text-generation",
"processor": null
} | {"base_model": null, "datasets": ["fnlp/OmniAction"], "eval_results": null, "language": ["en"], "library_name": "transformers", "license": "cc-by-nc-4.0", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "robotics", "tags": null} | null | null | [
"cc-by-nc-4.0"
] | [
"fnlp/OmniAction"
] | [
"en"
] | null | null | null | [
"AutoModelForSeq2SeqLM",
"qwen2_5_omni_thinker",
"Qwen2_5OmniThinkerForConditionalGeneration"
] | [
"robotics",
"text-generation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
621ffdc036468d709f174338 | google-bert/bert-base-uncased | google-bert | null | 71,392,969 | 2,805,772,843 | False | 2022-03-02T23:29:04Z | 2024-02-19T11:06:12Z | transformers | 2,595 | 8 | null | fill-mask | {"parameters": {"F32": 110106428}, "total": 110106428} | [
".gitattributes",
"LICENSE",
"README.md",
"config.json",
"coreml/fill-mask/float32_model.mlpackage/Data/com.apple.CoreML/model.mlmodel",
"coreml/fill-mask/float32_model.mlpackage/Data/com.apple.CoreML/weights/weight.bin",
"coreml/fill-mask/float32_model.mlpackage/Manifest.json",
"flax_model.msgpack",
... | 86b5e0934494bd15c9632b12f734a8a67f723594 | [
"transformers",
"pytorch",
"tf",
"jax",
"rust",
"coreml",
"onnx",
"safetensors",
"bert",
"fill-mask",
"exbert",
"en",
"dataset:bookcorpus",
"dataset:wikipedia",
"arxiv:1810.04805",
"license:apache-2.0",
"endpoints_compatible",
"deploy:azure",
"region:us"
] | null | {"architectures": ["BertForMaskedLM"], "model_type": "bert", "tokenizer_config": {}} | {
"auto_model": "AutoModelForMaskedLM",
"custom_class": null,
"pipeline_tag": "fill-mask",
"processor": "AutoTokenizer"
} | {"base_model": null, "datasets": ["bookcorpus", "wikipedia"], "eval_results": null, "language": "en", "library_name": null, "license": "apache-2.0", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": null, "tags": ["exbert"]} | # BERT base model (uncased)
Pretrained model on English language using a masked language modeling (MLM) objective. It was introduced in
[this paper](https://arxiv.org/abs/1810.04805) and first released in
[this repository](https://github.com/google-research/bert). This model is uncased: it does not make a difference
b... | null | [
"apache-2.0"
] | [
"bookcorpus",
"wikipedia"
] | [
"en"
] | 110,106,428 | null | null | [
"AutoModelForMaskedLM",
"bert",
"BertForMaskedLM"
] | [
"fill-mask"
] | [
"text"
] | [
"text"
] | [
"logits"
] |
66fba7309482f97131bf08d6 | openai/whisper-large-v3-turbo | openai | {
"models": [
{
"_id": "654a84cadff2f49007ce6c37",
"id": "openai/whisper-large-v3"
}
],
"relation": "finetune"
} | 5,038,910 | 72,108,416 | False | 2024-10-01T07:39:28Z | 2024-10-04T14:51:11Z | transformers | 2,871 | 8 | null | automatic-speech-recognition | null | [
".gitattributes",
"README.md",
"added_tokens.json",
"config.json",
"generation_config.json",
"merges.txt",
"model.safetensors",
"normalizer.json",
"preprocessor_config.json",
"special_tokens_map.json",
"tokenizer.json",
"tokenizer_config.json",
"vocab.json"
] | 41f01f3fe87f28c78e2fbf8b568835947dd65ed9 | [
"transformers",
"safetensors",
"whisper",
"automatic-speech-recognition",
"audio",
"en",
"zh",
"de",
"es",
"ru",
"ko",
"fr",
"ja",
"pt",
"tr",
"pl",
"ca",
"nl",
"ar",
"sv",
"it",
"id",
"hi",
"fi",
"vi",
"he",
"uk",
"el",
"ms",
"cs",
"ro",
"da",
"hu",
... | null | {"architectures": ["WhisperForConditionalGeneration"], "model_type": "whisper", "tokenizer_config": {"bos_token": "<|endoftext|>", "eos_token": "<|endoftext|>", "pad_token": "<|endoftext|>", "unk_token": "<|endoftext|>"}} | {
"auto_model": "AutoModelForSpeechSeq2Seq",
"custom_class": null,
"pipeline_tag": "automatic-speech-recognition",
"processor": "AutoProcessor"
} | {"base_model": ["openai/whisper-large-v3"], "datasets": null, "eval_results": null, "language": ["en", "zh", "de", "es", "ru", "ko", "fr", "ja", "pt", "tr", "pl", "ca", "nl", "ar", "sv", "it", "id", "hi", "fi", "vi", "he", "uk", "el", "ms", "cs", "ro", "da", "hu", "ta", "no", "th", "ur", "hr", "bg", "lt", "la", "mi", "... | # Whisper
Whisper is a state-of-the-art model for automatic speech recognition (ASR) and speech translation, proposed in the paper
[Robust Speech Recognition via Large-Scale Weak Supervision](https://huggingface.co/papers/2212.04356) by Alec Radford
et al. from OpenAI. Trained on >5M hours of labeled data, Whisper d... | null | [
"mit"
] | null | [
"en",
"zh",
"de",
"es",
"ru",
"ko",
"fr",
"ja",
"pt",
"tr",
"pl",
"ca",
"nl",
"ar",
"sv",
"it",
"id",
"hi",
"fi",
"vi",
"he",
"uk",
"el",
"ms",
"cs",
"ro",
"da",
"hu",
"ta",
"no",
"th",
"ur",
"hr",
"bg",
"lt",
"la",
"mi",
"ml",
"cy",
"sk"... | null | null | null | [
"WhisperForConditionalGeneration",
"whisper",
"AutoModelForSpeechSeq2Seq"
] | [
"automatic-speech-recognition"
] | [
"multimodal"
] | [
"audio"
] | [
"text"
] |
6795ffcd88cd7c0294702a72 | Qwen/Qwen2.5-VL-7B-Instruct | Qwen | null | 4,611,358 | 51,675,220 | False | 2025-01-26T09:26:37Z | 2025-04-06T16:23:01Z | transformers | 1,476 | 8 | null | image-text-to-text | {"parameters": {"BF16": 8292166656}, "total": 8292166656} | [
".gitattributes",
"README.md",
"chat_template.json",
"config.json",
"generation_config.json",
"merges.txt",
"model-00001-of-00005.safetensors",
"model-00002-of-00005.safetensors",
"model-00003-of-00005.safetensors",
"model-00004-of-00005.safetensors",
"model-00005-of-00005.safetensors",
"model... | cc594898137f460bfe9f0759e9844b3ce807cfb5 | [
"transformers",
"safetensors",
"qwen2_5_vl",
"image-text-to-text",
"multimodal",
"conversational",
"en",
"arxiv:2309.00071",
"arxiv:2409.12191",
"arxiv:2308.12966",
"license:apache-2.0",
"eval-results",
"text-generation-inference",
"endpoints_compatible",
"deploy:azure",
"region:us"
] | null | {"architectures": ["Qwen2_5_VLForConditionalGeneration"], "model_type": "qwen2_5_vl", "processor_config": {"chat_template": "{% set image_count = namespace(value=0) %}{% set video_count = namespace(value=0) %}{% for message in messages %}{% if loop.first and message['role'] != 'system' %}<|im_start|>system\nYou are a h... | {
"auto_model": "AutoModelForImageTextToText",
"custom_class": null,
"pipeline_tag": "image-text-to-text",
"processor": "AutoProcessor"
} | {"base_model": null, "datasets": null, "eval_results": null, "language": ["en"], "library_name": "transformers", "license": "apache-2.0", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "image-text-to-text", "tags": ["multimodal"]} | # Qwen2.5-VL-7B-Instruct
<a href="https://chat.qwenlm.ai/" target="_blank" style="margin: 2px;">
<img alt="Chat" src="https://img.shields.io/badge/%F0%9F%92%9C%EF%B8%8F%20Qwen%20Chat%20-536af5" style="display: inline-block; vertical-align: middle;"/>
</a>
## Introduction
In the past five months since Qwen2-VL’s r... | null | [
"apache-2.0"
] | null | [
"en"
] | 8,292,166,656 | null | null | [
"qwen2_5_vl",
"AutoModelForImageTextToText",
"Qwen2_5_VLForConditionalGeneration"
] | [
"image-text-to-text"
] | [
"multimodal"
] | [
"text",
"image"
] | [
"text"
] |
65f44de295b7d70871c368eb | BAAI/bge-reranker-v2-m3 | BAAI | null | 5,692,258 | 47,134,293 | False | 2024-03-15T13:32:18Z | 2024-06-24T14:08:45Z | sentence-transformers | 934 | 7 | null | text-classification | {"parameters": {"F32": 567755777}, "total": 567755777} | [
".gitattributes",
"README.md",
"assets/BEIR-bge-en-v1.5.png",
"assets/BEIR-e5-mistral.png",
"assets/CMTEB-retrieval-bge-zh-v1.5.png",
"assets/llama-index.png",
"assets/miracl-bge-m3.png",
"config.json",
"model.safetensors",
"sentencepiece.bpe.model",
"special_tokens_map.json",
"tokenizer.json"... | 953dc6f6f85a1b2dbfca4c34a2796e7dde08d41e | [
"sentence-transformers",
"safetensors",
"xlm-roberta",
"text-classification",
"transformers",
"text-embeddings-inference",
"multilingual",
"arxiv:2312.15503",
"arxiv:2402.03216",
"license:apache-2.0",
"endpoints_compatible",
"deploy:azure",
"region:us"
] | null | {"architectures": ["XLMRobertaForSequenceClassification"], "model_type": "xlm-roberta", "tokenizer_config": {"bos_token": "<s>", "cls_token": "<s>", "eos_token": "</s>", "mask_token": "<mask>", "pad_token": "<pad>", "sep_token": "</s>", "unk_token": "<unk>"}} | {
"auto_model": "AutoModelForSequenceClassification",
"custom_class": null,
"pipeline_tag": "text-classification",
"processor": "AutoTokenizer"
} | {"base_model": null, "datasets": null, "eval_results": null, "language": ["multilingual"], "library_name": null, "license": "apache-2.0", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "text-classification", "tags": ["transformers", "sentence-transformers", "text-embedd... | # Reranker
**More details please refer to our Github: [FlagEmbedding](https://github.com/FlagOpen/FlagEmbedding/tree/master).**
- [Model List](#model-list)
- [Usage](#usage)
- [Fine-tuning](#fine-tune)
- [Evaluation](#evaluation)
- [Citation](#citation)
Different from embedding model, reranker uses question and docu... | null | [
"apache-2.0"
] | null | [
"multilingual"
] | 567,755,777 | null | null | [
"AutoModelForSequenceClassification",
"XLMRobertaForSequenceClassification",
"xlm-roberta"
] | [
"text-classification"
] | [
"text"
] | [
"text"
] | [
"logits"
] |
666ce1af091b7c8e3f9aae7d | microsoft/Florence-2-large | microsoft | null | 1,208,523 | 19,527,799 | False | 2024-06-15T00:34:55Z | 2025-08-04T17:09:36Z | transformers | 1,786 | 7 | null | image-text-to-text | {"parameters": {"F16": 776721497}, "total": 776721497} | [
".gitattributes",
"CODE_OF_CONDUCT.md",
"LICENSE",
"README.md",
"SECURITY.md",
"SUPPORT.md",
"config.json",
"configuration_florence2.py",
"generation_config.json",
"model.safetensors",
"modeling_florence2.py",
"preprocessor_config.json",
"processing_florence2.py",
"pytorch_model.bin",
"s... | 21a599d414c4d928c9032694c424fb94458e3594 | [
"transformers",
"pytorch",
"safetensors",
"florence2",
"image-text-to-text",
"vision",
"custom_code",
"arxiv:2311.06242",
"license:mit",
"endpoints_compatible",
"region:us"
] | null | {"architectures": ["Florence2ForConditionalGeneration"], "auto_map": {"AutoConfig": "configuration_florence2.Florence2Config", "AutoModelForCausalLM": "modeling_florence2.Florence2ForConditionalGeneration"}, "model_type": "florence2", "tokenizer_config": {}} | {
"auto_model": "AutoModelForImageTextToText",
"custom_class": null,
"pipeline_tag": "image-text-to-text",
"processor": "AutoProcessor"
} | {"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": null, "license": "mit", "license_name": null, "license_link": "https://huggingface.co/microsoft/Florence-2-large/resolve/main/LICENSE", "metrics": null, "model_name": null, "pipeline_tag": "image-text-to-text", "tags": ["visi... | null | null | [
"mit",
"https://huggingface.co/microsoft/Florence-2-large/resolve/main/LICENSE"
] | null | null | 776,721,497 | null | null | [
"florence2",
"AutoModelForImageTextToText",
"Florence2ForConditionalGeneration"
] | [
"image-text-to-text"
] | [
"multimodal"
] | [
"text",
"image"
] | [
"text"
] |
66e98dd5899bdb384bd953b4 | Qwen/Qwen2.5-1.5B-Instruct | Qwen | {
"models": [
{
"_id": "66e6d032663c452ad2400439",
"id": "Qwen/Qwen2.5-1.5B"
}
],
"relation": "finetune"
} | 9,542,607 | 152,014,423 | False | 2024-09-17T14:10:29Z | 2024-09-25T12:32:50Z | transformers | 650 | 7 | null | text-generation | {"parameters": {"BF16": 1543714304}, "total": 1543714304} | [
".gitattributes",
"LICENSE",
"README.md",
"config.json",
"generation_config.json",
"merges.txt",
"model.safetensors",
"tokenizer.json",
"tokenizer_config.json",
"vocab.json"
] | 989aa7980e4cf806f80c7fef2b1adb7bc71aa306 | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"chat",
"conversational",
"en",
"arxiv:2407.10671",
"base_model:Qwen/Qwen2.5-1.5B",
"base_model:finetune:Qwen/Qwen2.5-1.5B",
"license:apache-2.0",
"text-generation-inference",
"endpoints_compatible",
"deploy:azure",
"region:us"
] | null | {"architectures": ["Qwen2ForCausalLM"], "model_type": "qwen2", "tokenizer_config": {"bos_token": null, "chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0]['role'] == 'system' %}\n {{- messages[0]['content'] }}\n {%- else %}\n {{- 'You are Qwen, created by Aliba... | {
"auto_model": "AutoModelForCausalLM",
"custom_class": null,
"pipeline_tag": "text-generation",
"processor": "AutoTokenizer"
} | {"base_model": "Qwen/Qwen2.5-1.5B", "datasets": null, "eval_results": null, "language": ["en"], "library_name": "transformers", "license": "apache-2.0", "license_name": null, "license_link": "https://huggingface.co/Qwen/Qwen2.5-1.5B-Instruct/blob/main/LICENSE", "metrics": null, "model_name": null, "pipeline_tag": "text... | # Qwen2.5-1.5B-Instruct
## Introduction
Qwen2.5 is the latest series of Qwen large language models. For Qwen2.5, we release a number of base language models and instruction-tuned language models ranging from 0.5 to 72 billion parameters. Qwen2.5 brings the following improvements upon Qwen2:
- Significantly **more kn... | null | [
"apache-2.0",
"https://huggingface.co/Qwen/Qwen2.5-1.5B-Instruct/blob/main/LICENSE"
] | null | [
"en"
] | 1,543,714,304 | null | null | [
"AutoModelForCausalLM",
"Qwen2ForCausalLM",
"qwen2"
] | [
"text-generation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
678e11922b39b4ed1381531b | deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B | deepseek-ai | null | 633,494 | 17,413,391 | False | 2025-01-20T09:04:18Z | 2025-02-24T03:32:35Z | transformers | 1,464 | 7 | null | text-generation | {"parameters": {"BF16": 1777088000}, "total": 1777088000} | [
".gitattributes",
"LICENSE",
"README.md",
"config.json",
"figures/benchmark.jpg",
"generation_config.json",
"model.safetensors",
"tokenizer.json",
"tokenizer_config.json"
] | ad9f0ae0864d7fbcd1cd905e3c6c5b069cc8b562 | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"conversational",
"arxiv:2501.12948",
"license:mit",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | null | {"architectures": ["Qwen2ForCausalLM"], "model_type": "qwen2", "tokenizer_config": {"bos_token": {"__type": "AddedToken", "content": "<\uff5cbegin\u2581of\u2581sentence\uff5c>", "lstrip": false, "normalized": true, "rstrip": false, "single_word": false}, "eos_token": {"__type": "AddedToken", "content": "<\uff5cend\u258... | {
"auto_model": "AutoModelForCausalLM",
"custom_class": null,
"pipeline_tag": "text-generation",
"processor": "AutoTokenizer"
} | {"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "mit", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": null, "tags": null} | # DeepSeek-R1
<!-- markdownlint-disable first-line-h1 -->
<!-- markdownlint-disable html -->
<!-- markdownlint-disable no-duplicate-header -->
<div align="center">
<img src="https://github.com/deepseek-ai/DeepSeek-V2/blob/main/figures/logo.svg?raw=true" width="60%" alt="DeepSeek-V3" />
</div>
<hr>
<div align="center... | null | [
"mit"
] | null | null | 1,777,088,000 | null | null | [
"AutoModelForCausalLM",
"Qwen2ForCausalLM",
"qwen2"
] | [
"text-generation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
67b52d4a824d77f2bba8b0af | microsoft/Phi-4-mini-instruct | microsoft | null | 605,689 | 3,848,451 | False | 2025-02-19T01:00:58Z | 2025-12-10T20:24:40Z | transformers | 706 | 7 | null | text-generation | null | [
".gitattributes",
"CODE_OF_CONDUCT.md",
"LICENSE",
"NOTICE.md",
"README.md",
"SECURITY.md",
"added_tokens.json",
"config.json",
"configuration_phi3.py",
"data_summary_card.md",
"generation_config.json",
"merges.txt",
"model-00001-of-00002.safetensors",
"model-00002-of-00002.safetensors",
... | cfbefacb99257ffa30c83adab238a50856ac3083 | [
"transformers",
"safetensors",
"phi3",
"text-generation",
"nlp",
"code",
"conversational",
"custom_code",
"multilingual",
"ar",
"zh",
"cs",
"da",
"nl",
"en",
"fi",
"fr",
"de",
"he",
"hu",
"it",
"ja",
"ko",
"no",
"pl",
"pt",
"ru",
"es",
"sv",
"th",
"tr",
... | null | {"architectures": ["Phi3ForCausalLM"], "auto_map": {"AutoConfig": "configuration_phi3.Phi3Config", "AutoModelForCausalLM": "modeling_phi3.Phi3ForCausalLM", "AutoTokenizer": "Xenova/gpt-4o"}, "model_type": "phi3", "tokenizer_config": {"bos_token": "<|endoftext|>", "chat_template": "{% for message in messages %}{% if mes... | {
"auto_model": "AutoModelForCausalLM",
"custom_class": null,
"pipeline_tag": "text-generation",
"processor": "AutoTokenizer"
} | {"base_model": null, "datasets": null, "eval_results": null, "language": ["multilingual", "ar", "zh", "cs", "da", "nl", "en", "fi", "fr", "de", "he", "hu", "it", "ja", "ko", "no", "pl", "pt", "ru", "es", "sv", "th", "tr", "uk"], "library_name": "transformers", "license": "mit", "license_name": null, "license_link": "ht... | null | null | [
"mit",
"https://huggingface.co/microsoft/Phi-4-mini-instruct/resolve/main/LICENSE"
] | null | [
"multilingual",
"ar",
"zh",
"cs",
"da",
"nl",
"en",
"fi",
"fr",
"de",
"he",
"hu",
"it",
"ja",
"ko",
"no",
"pl",
"pt",
"ru",
"es",
"sv",
"th",
"tr",
"uk"
] | null | null | null | [
"AutoModelForCausalLM",
"phi3",
"Phi3ForCausalLM"
] | [
"text-generation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
680da7517bcb5198014ad394 | Qwen/Qwen3-1.7B | Qwen | {
"models": [
{
"_id": "680f0cb340199096ff7bae3e",
"id": "Qwen/Qwen3-1.7B-Base"
}
],
"relation": "finetune"
} | 6,900,591 | 30,148,110 | False | 2025-04-27T03:41:05Z | 2025-07-26T03:46:32Z | transformers | 435 | 7 | null | text-generation | {"parameters": {"BF16": 2031739904}, "total": 2031739904} | [
".gitattributes",
"LICENSE",
"README.md",
"config.json",
"generation_config.json",
"merges.txt",
"model-00001-of-00002.safetensors",
"model-00002-of-00002.safetensors",
"model.safetensors.index.json",
"tokenizer.json",
"tokenizer_config.json",
"vocab.json"
] | 70d244cc86ccca08cf5af4e1e306ecf908b1ad5e | [
"transformers",
"safetensors",
"qwen3",
"text-generation",
"conversational",
"arxiv:2505.09388",
"base_model:Qwen/Qwen3-1.7B-Base",
"base_model:finetune:Qwen/Qwen3-1.7B-Base",
"license:apache-2.0",
"text-generation-inference",
"endpoints_compatible",
"deploy:azure",
"region:us"
] | null | {"architectures": ["Qwen3ForCausalLM"], "model_type": "qwen3", "tokenizer_config": {"bos_token": null, "chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0].role == 'system' %}\n {{- messages[0].content + '\\n\\n' }}\n {%- endif %}\n {{- \"# Tools\\n\\nYou may call o... | {
"auto_model": "AutoModelForCausalLM",
"custom_class": null,
"pipeline_tag": "text-generation",
"processor": "AutoTokenizer"
} | {"base_model": ["Qwen/Qwen3-1.7B-Base"], "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "apache-2.0", "license_name": null, "license_link": "https://huggingface.co/Qwen/Qwen3-1.7B/blob/main/LICENSE", "metrics": null, "model_name": null, "pipeline_tag": "text-generat... | null | null | [
"apache-2.0",
"https://huggingface.co/Qwen/Qwen3-1.7B/blob/main/LICENSE"
] | null | null | 2,031,739,904 | null | null | [
"AutoModelForCausalLM",
"Qwen3ForCausalLM",
"qwen3"
] | [
"text-generation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
682b8c278503c5bbf012e9a9 | ibm-granite/granite-docling-258M | ibm-granite | null | 68,982 | 1,071,300 | False | 2025-05-19T19:53:11Z | 2025-09-23T08:52:16Z | transformers | 1,143 | 7 | null | image-text-to-text | null | [
".gitattributes",
"README.md",
"added_tokens.json",
"assets/granite_docling_split_page.png",
"assets/new_arxiv.png",
"chat_template.jinja",
"config.json",
"generation_config.json",
"granite_docling.png",
"merges.txt",
"model.safetensors",
"preprocessor_config.json",
"processor_config.json",
... | 982fe3b40f2fa73c365bdb1bcacf6c81b7184bfe | [
"transformers",
"safetensors",
"idefics3",
"image-text-to-text",
"text-generation",
"documents",
"code",
"formula",
"chart",
"ocr",
"layout",
"table",
"document-parse",
"docling",
"granite",
"extraction",
"math",
"conversational",
"en",
"dataset:ds4sd/SynthCodeNet",
"dataset:... | null | {"architectures": ["Idefics3ForConditionalGeneration"], "model_type": "idefics3", "tokenizer_config": {"bos_token": "<|start_of_role|>", "eos_token": "<|end_of_text|>", "pad_token": "<|end_of_text|>", "unk_token": "<|unk|>"}, "chat_template_jinja": "{%- for message in messages -%}\n{{- '<|start_of_role|>' + message['ro... | {
"auto_model": "AutoModelForImageTextToText",
"custom_class": null,
"pipeline_tag": "image-text-to-text",
"processor": "AutoProcessor"
} | {"base_model": null, "datasets": ["ds4sd/SynthCodeNet", "ds4sd/SynthFormulaNet", "ds4sd/SynthChartNet", "HuggingFaceM4/DoclingMatix"], "eval_results": null, "language": ["en"], "library_name": "transformers", "license": "apache-2.0", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipe... | # granite-docling-258m
<div style="display: flex; align-items: center;">
<img src="https://huggingface.co/ibm-granite/granite-docling-258M/resolve/main/granite_docling.png" alt="Granite Docling Logo" style="width: 200px; height: auto; margin-right: 20px;">
<div>
<p>Granite Docling is a multimodal Image-... | null | [
"apache-2.0"
] | [
"ds4sd/SynthCodeNet",
"ds4sd/SynthFormulaNet",
"ds4sd/SynthChartNet",
"HuggingFaceM4/DoclingMatix"
] | [
"en"
] | null | null | null | [
"AutoModelForImageTextToText",
"Idefics3ForConditionalGeneration",
"idefics3"
] | [
"image-text-to-text",
"text-generation"
] | [
"text",
"multimodal"
] | [
"text",
"image"
] | [
"text"
] |
686ceee17e3b40a013a9afdc | HuggingFaceTB/SmolLM3-3B | HuggingFaceTB | {
"models": [
{
"_id": "6853f5779e7c268e15850d9c",
"id": "HuggingFaceTB/SmolLM3-3B-Base"
}
],
"relation": "finetune"
} | 1,092,049 | 2,453,821 | False | 2025-07-08T10:11:45Z | 2025-09-10T12:28:11Z | transformers | 920 | 7 | null | text-generation | {"parameters": {"BF16": 3075098624}, "total": 3075098624} | [
".gitattributes",
"README.md",
"chat_template.jinja",
"config.json",
"generation_config.json",
"model-00001-of-00002.safetensors",
"model-00002-of-00002.safetensors",
"model.safetensors.index.json",
"notebook.ipynb",
"special_tokens_map.json",
"tokenizer.json",
"tokenizer_config.json"
] | a07cc9a04f16550a088caea529712d1d335b0ac1 | [
"transformers",
"safetensors",
"smollm3",
"text-generation",
"conversational",
"en",
"fr",
"es",
"it",
"pt",
"zh",
"ar",
"ru",
"base_model:HuggingFaceTB/SmolLM3-3B-Base",
"base_model:finetune:HuggingFaceTB/SmolLM3-3B-Base",
"license:apache-2.0",
"endpoints_compatible",
"deploy:azur... | null | {"architectures": ["SmolLM3ForCausalLM"], "model_type": "smollm3", "tokenizer_config": {"bos_token": null, "eos_token": "<|im_end|>", "pad_token": "<|im_end|>"}, "chat_template_jinja": "{# \u2500\u2500\u2500\u2500\u2500 defaults \u2500\u2500\u2500\u2500\u2500 #}\n{%- if enable_thinking is not defined -%}\n{%- set enabl... | {
"auto_model": "AutoModelForCausalLM",
"custom_class": null,
"pipeline_tag": "text-generation",
"processor": "AutoTokenizer"
} | {"base_model": ["HuggingFaceTB/SmolLM3-3B-Base"], "datasets": null, "eval_results": null, "language": ["en", "fr", "es", "it", "pt", "zh", "ar", "ru"], "library_name": "transformers", "license": "apache-2.0", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": null, "tags": ... | # SmolLM3

## Table of Contents
1. [Model Summary](#model-summary)
2. [How to use](#how-to-use)
3. [Evaluation](#evaluation)
4. [Training](#training)
5. [Limitations](#limitations)
6. [License](#l... | null | [
"apache-2.0"
] | null | [
"en",
"fr",
"es",
"it",
"pt",
"zh",
"ar",
"ru"
] | 3,075,098,624 | null | null | [
"AutoModelForCausalLM",
"SmolLM3ForCausalLM",
"smollm3"
] | [
"text-generation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
621ffdc136468d709f17ea63 | openai/clip-vit-large-patch14 | openai | null | 24,178,746 | 1,101,271,075 | False | 2022-03-02T23:29:05Z | 2023-09-15T15:49:35Z | transformers | 1,978 | 6 | null | zero-shot-image-classification | {"parameters": {"I64": 334, "F32": 427616512}, "total": 427616846} | [
".gitattributes",
"README.md",
"config.json",
"flax_model.msgpack",
"merges.txt",
"model.safetensors",
"preprocessor_config.json",
"pytorch_model.bin",
"special_tokens_map.json",
"tf_model.h5",
"tokenizer.json",
"tokenizer_config.json",
"vocab.json"
] | 32bd64288804d66eefd0ccbe215aa642df71cc41 | [
"transformers",
"pytorch",
"tf",
"jax",
"safetensors",
"clip",
"zero-shot-image-classification",
"vision",
"arxiv:2103.00020",
"arxiv:1908.04913",
"endpoints_compatible",
"region:us"
] | null | {"architectures": ["CLIPModel"], "model_type": "clip", "tokenizer_config": {"unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "bos_token": {"content": "<|startoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "... | {
"auto_model": "AutoModelForZeroShotImageClassification",
"custom_class": null,
"pipeline_tag": "zero-shot-image-classification",
"processor": "AutoProcessor"
} | {"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": null, "license": null, "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": null, "tags": ["vision"], "widget": [{"src": "https://huggingface.co/datasets/mishig/sample_images/resolv... | # Model Card: CLIP
Disclaimer: The model card is taken and modified from the official CLIP repository, it can be found [here](https://github.com/openai/CLIP/blob/main/model-card.md).
## Model Details
The CLIP model was developed by researchers at OpenAI to learn about what contributes to robustness in computer visio... | null | null | null | null | 427,616,846 | null | null | [
"AutoModelForZeroShotImageClassification",
"CLIPModel",
"clip"
] | [
"zero-shot-image-classification"
] | [
"multimodal"
] | [
"text",
"image"
] | [
"logits"
] |
621ffdc136468d709f1802ec | sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2 | sentence-transformers | null | 26,743,180 | 314,376,005 | False | 2022-03-02T23:29:05Z | 2026-01-28T10:02:26Z | sentence-transformers | 1,172 | 6 | null | sentence-similarity | {"parameters": {"I64": 512, "F32": 117653760}, "total": 117654272} | [
".gitattributes",
"1_Pooling/config.json",
"README.md",
"config.json",
"config_sentence_transformers.json",
"model.safetensors",
"modules.json",
"onnx/model.onnx",
"onnx/model_O1.onnx",
"onnx/model_O2.onnx",
"onnx/model_O3.onnx",
"onnx/model_O4.onnx",
"onnx/model_qint8_arm64.onnx",
"onnx/m... | e8f8c211226b894fcb81acc59f3b34ba3efd5f42 | [
"sentence-transformers",
"pytorch",
"tf",
"onnx",
"safetensors",
"openvino",
"bert",
"feature-extraction",
"sentence-similarity",
"transformers",
"multilingual",
"ar",
"bg",
"ca",
"cs",
"da",
"de",
"el",
"en",
"es",
"et",
"fa",
"fi",
"fr",
"gl",
"gu",
"he",
"hi"... | null | {"architectures": ["BertModel"], "model_type": "bert", "tokenizer_config": {"unk_token": "<unk>", "sep_token": "</s>", "pad_token": "<pad>", "cls_token": "<s>", "mask_token": {"content": "<mask>", "single_word": false, "lstrip": true, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "bos_token": "<s>", "eo... | {
"auto_model": "AutoModel",
"custom_class": null,
"pipeline_tag": "feature-extraction",
"processor": "AutoTokenizer"
} | {"base_model": null, "datasets": null, "eval_results": null, "language": ["multilingual", "ar", "bg", "ca", "cs", "da", "de", "el", "en", "es", "et", "fa", "fi", "fr", "gl", "gu", "he", "hi", "hr", "hu", "hy", "id", "it", "ja", "ka", "ko", "ku", "lt", "lv", "mk", "mn", "mr", "ms", "my", "nb", "nl", "pl", "pt", "ro", "r... | # sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2
This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 384 dimensional dense vector space and can be used for tasks like clustering or semantic search.
## Usage (Sentence-Transformers)
Using this model become... | null | [
"apache-2.0"
] | null | [
"multilingual",
"ar",
"bg",
"ca",
"cs",
"da",
"de",
"el",
"en",
"es",
"et",
"fa",
"fi",
"fr",
"gl",
"gu",
"he",
"hi",
"hr",
"hu",
"hy",
"id",
"it",
"ja",
"ka",
"ko",
"ku",
"lt",
"lv",
"mk",
"mn",
"mr",
"ms",
"my",
"nb",
"nl",
"pl",
"pt",
"r... | 117,654,272 | null | null | [
"BertModel",
"AutoModel",
"bert"
] | [
"sentence-similarity",
"feature-extraction"
] | [
"text",
"multimodal"
] | [
"text"
] | [
"logits",
"embeddings"
] |
676c000762cee1f3abc3ed5f | deepseek-ai/DeepSeek-V3 | deepseek-ai | null | 754,886 | 14,403,499 | False | 2024-12-25T12:52:23Z | 2025-03-27T04:01:45Z | transformers | 4,019 | 6 | null | text-generation | {"parameters": {"BF16": 3918786560, "F8_E4M3": 680571043840, "F32": 41555600}, "total": 684531386000} | [
".gitattributes",
"LICENSE-CODE",
"LICENSE-MODEL",
"README.md",
"README_WEIGHTS.md",
"config.json",
"configuration_deepseek.py",
"figures/benchmark.png",
"figures/niah.png",
"inference/configs/config_16B.json",
"inference/configs/config_236B.json",
"inference/configs/config_671B.json",
"infe... | e815299b0bcbac849fa540c768ef21845365c9eb | [
"transformers",
"safetensors",
"deepseek_v3",
"text-generation",
"conversational",
"custom_code",
"arxiv:2412.19437",
"eval-results",
"text-generation-inference",
"endpoints_compatible",
"fp8",
"region:us"
] | null | {"architectures": ["DeepseekV3ForCausalLM"], "auto_map": {"AutoConfig": "configuration_deepseek.DeepseekV3Config", "AutoModel": "modeling_deepseek.DeepseekV3Model", "AutoModelForCausalLM": "modeling_deepseek.DeepseekV3ForCausalLM"}, "model_type": "deepseek_v3", "quantization_config": {"quant_method": "fp8"}, "tokenizer... | {
"auto_model": "AutoModelForCausalLM",
"custom_class": null,
"pipeline_tag": "text-generation",
"processor": "AutoTokenizer"
} | {"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": null, "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": null, "tags": null} | <!-- markdownlint-disable first-line-h1 -->
<!-- markdownlint-disable html -->
<!-- markdownlint-disable no-duplicate-header -->
<div align="center">
<img src="https://github.com/deepseek-ai/DeepSeek-V2/blob/main/figures/logo.svg?raw=true" width="60%" alt="DeepSeek-V3" />
</div>
<hr>
<div align="center" style="line-... | null | null | null | null | 684,531,386,000 | null | null | [
"AutoModelForCausalLM",
"deepseek_v3",
"DeepseekV3ForCausalLM"
] | [
"text-generation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
67c35be6eae05d8f94fae4c2 | google/gemma-3-12b-it | google | {
"models": [
{
"_id": "67c35bd50e998b1f7c4d3139",
"id": "google/gemma-3-12b-pt"
}
],
"relation": "finetune"
} | 2,534,085 | 12,192,318 | manual | 2025-03-01T19:11:34Z | 2025-03-21T20:28:56Z | transformers | 696 | 6 | null | image-text-to-text | null | [
".gitattributes",
"README.md",
"added_tokens.json",
"chat_template.json",
"config.json",
"generation_config.json",
"model-00001-of-00005.safetensors",
"model-00002-of-00005.safetensors",
"model-00003-of-00005.safetensors",
"model-00004-of-00005.safetensors",
"model-00005-of-00005.safetensors",
... | 96b6f1eccf38110c56df3a15bffe176da04bfd80 | [
"transformers",
"safetensors",
"gemma3",
"image-text-to-text",
"conversational",
"arxiv:1905.07830",
"arxiv:1905.10044",
"arxiv:1911.11641",
"arxiv:1904.09728",
"arxiv:1705.03551",
"arxiv:1911.01547",
"arxiv:1907.10641",
"arxiv:1903.00161",
"arxiv:2009.03300",
"arxiv:2304.06364",
"arxi... | null | {"architectures": ["Gemma3ForConditionalGeneration"], "model_type": "gemma3", "processor_config": {"chat_template": "{{ bos_token }}\n{%- if messages[0]['role'] == 'system' -%}\n {%- if messages[0]['content'] is string -%}\n {%- set first_user_prefix = messages[0]['content'] + '\n\n' -%}\n {%- else -%}\n ... | {
"auto_model": "AutoModelForImageTextToText",
"custom_class": null,
"pipeline_tag": "image-text-to-text",
"processor": "AutoProcessor"
} | {"base_model": "google/gemma-3-12b-pt", "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "gemma", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "image-text-to-text", "tags": null, "extra_gated_heading": "Access Gemma ... | null | null | [
"gemma"
] | null | null | null | null | null | [
"AutoModelForImageTextToText",
"Gemma3ForConditionalGeneration",
"gemma3"
] | [
"image-text-to-text"
] | [
"multimodal"
] | [
"text",
"image"
] | [
"text"
] |
67f524d174216a9575d48e68 | google/gemma-3-12b-it-qat-q4_0-unquantized | google | {
"models": [
{
"_id": "67c35be6eae05d8f94fae4c2",
"id": "google/gemma-3-12b-it"
}
],
"relation": "finetune"
} | 27,707 | 77,234 | manual | 2025-04-08T13:29:53Z | 2025-04-15T21:07:07Z | transformers | 80 | 6 | null | image-text-to-text | null | [
".gitattributes",
"README.md",
"added_tokens.json",
"chat_template.json",
"config.json",
"generation_config.json",
"model-00001-of-00005.safetensors",
"model-00002-of-00005.safetensors",
"model-00003-of-00005.safetensors",
"model-00004-of-00005.safetensors",
"model-00005-of-00005.safetensors",
... | 68f7ee4fbd59087436ada77ed2d62f373fdd4482 | [
"transformers",
"safetensors",
"gemma3",
"image-text-to-text",
"gemma",
"google",
"conversational",
"arxiv:1905.07830",
"arxiv:1905.10044",
"arxiv:1911.11641",
"arxiv:1904.09728",
"arxiv:1705.03551",
"arxiv:1911.01547",
"arxiv:1907.10641",
"arxiv:1903.00161",
"arxiv:2009.03300",
"arx... | null | {"architectures": ["Gemma3ForConditionalGeneration"], "model_type": "gemma3", "processor_config": {"chat_template": "{{ bos_token }}\n{%- if messages[0]['role'] == 'system' -%}\n {%- if messages[0]['content'] is string -%}\n {%- set first_user_prefix = messages[0]['content'] + '\n\n' -%}\n {%- else -%}\n ... | {
"auto_model": "AutoModelForImageTextToText",
"custom_class": null,
"pipeline_tag": "image-text-to-text",
"processor": "AutoProcessor"
} | {"base_model": "google/gemma-3-12b-it", "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "gemma", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "image-text-to-text", "tags": ["gemma3", "gemma", "google"], "extra_gated... | null | null | [
"gemma"
] | null | null | null | null | null | [
"AutoModelForImageTextToText",
"Gemma3ForConditionalGeneration",
"gemma3"
] | [
"image-text-to-text"
] | [
"multimodal"
] | [
"text",
"image"
] | [
"text"
] |
684b11e729c32c69e56d1db9 | google/gemma-3n-E2B-it | google | {
"models": [
{
"_id": "683f3c2354280d882006f816",
"id": "google/gemma-3n-E4B-it"
}
],
"relation": "finetune"
} | 279,054 | 1,756,622 | manual | 2025-06-12T17:44:07Z | 2025-07-14T13:55:52Z | transformers | 289 | 6 | null | image-text-to-text | null | [
".gitattributes",
"README.md",
"chat_template.jinja",
"config.json",
"generation_config.json",
"model-00001-of-00003.safetensors",
"model-00002-of-00003.safetensors",
"model-00003-of-00003.safetensors",
"model.safetensors.index.json",
"notebook.ipynb",
"preprocessor_config.json",
"processor_co... | 5e092ebca197cdcd8d8b195040accf22693501bc | [
"transformers",
"safetensors",
"gemma3n",
"image-text-to-text",
"automatic-speech-recognition",
"automatic-speech-translation",
"audio-text-to-text",
"video-text-to-text",
"conversational",
"arxiv:1905.07830",
"arxiv:1905.10044",
"arxiv:1911.11641",
"arxiv:1904.09728",
"arxiv:1705.03551",
... | null | {"architectures": ["Gemma3nForConditionalGeneration"], "model_type": "gemma3n", "tokenizer_config": {"bos_token": "<bos>", "eos_token": "<eos>", "pad_token": "<pad>", "unk_token": "<unk>", "use_default_system_prompt": false}, "chat_template_jinja": "{{ bos_token }}\n{%- if messages[0]['role'] == 'system' -%}\n {%- i... | {
"auto_model": "AutoModelForImageTextToText",
"custom_class": null,
"pipeline_tag": "image-text-to-text",
"processor": "AutoProcessor"
} | {"base_model": "google/gemma-3n-E4B-it", "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "gemma", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "image-text-to-text", "tags": ["automatic-speech-recognition", "automati... | null | null | [
"gemma"
] | null | null | null | null | null | [
"AutoModelForImageTextToText",
"gemma3n",
"Gemma3nForConditionalGeneration"
] | [
"image-text-to-text",
"video-text-to-text",
"automatic-speech-recognition"
] | [
"vision",
"multimodal"
] | [
"text",
"audio",
"image"
] | [
"text"
] |
621ffdc036468d709f174347 | distilbert/distilbert-base-uncased | distilbert | null | 7,004,758 | 627,000,229 | False | 2022-03-02T23:29:04Z | 2024-05-06T13:44:53Z | transformers | 848 | 5 | null | fill-mask | {"parameters": {"F32": 66985530}, "total": 66985530} | [
".gitattributes",
"LICENSE",
"README.md",
"config.json",
"flax_model.msgpack",
"model.safetensors",
"pytorch_model.bin",
"rust_model.ot",
"tf_model.h5",
"tokenizer.json",
"tokenizer_config.json",
"vocab.txt"
] | 12040accade4e8a0f71eabdb258fecc2e7e948be | [
"transformers",
"pytorch",
"tf",
"jax",
"rust",
"safetensors",
"distilbert",
"fill-mask",
"exbert",
"en",
"dataset:bookcorpus",
"dataset:wikipedia",
"arxiv:1910.01108",
"license:apache-2.0",
"endpoints_compatible",
"deploy:azure",
"region:us"
] | null | {"architectures": ["DistilBertForMaskedLM"], "model_type": "distilbert", "tokenizer_config": {}} | {
"auto_model": "AutoModelForMaskedLM",
"custom_class": null,
"pipeline_tag": "fill-mask",
"processor": "AutoTokenizer"
} | {"base_model": null, "datasets": ["bookcorpus", "wikipedia"], "eval_results": null, "language": "en", "library_name": null, "license": "apache-2.0", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": null, "tags": ["exbert"]} | # DistilBERT base model (uncased)
This model is a distilled version of the [BERT base model](https://huggingface.co/bert-base-uncased). It was
introduced in [this paper](https://arxiv.org/abs/1910.01108). The code for the distillation process can be found
[here](https://github.com/huggingface/transformers/tree/main/ex... | null | [
"apache-2.0"
] | [
"bookcorpus",
"wikipedia"
] | [
"en"
] | 66,985,530 | null | null | [
"distilbert",
"AutoModelForMaskedLM",
"DistilBertForMaskedLM"
] | [
"fill-mask"
] | [
"text"
] | [
"text"
] | [
"logits"
] |
64fff537d522560505ad6567 | BAAI/bge-small-en-v1.5 | BAAI | null | 11,538,782 | 118,876,425 | False | 2023-09-12T05:20:55Z | 2024-02-22T03:36:23Z | sentence-transformers | 429 | 5 | [{"name": "bge-small-en-v1.5", "results": [{"task": {"type": "Classification"}, "dataset": {"type": "mteb/amazon_counterfactual", "name": "MTEB AmazonCounterfactualClassification (en)", "config": "en", "split": "test", "revision": "e8379541af4e31359cca9fbcf4b00f2671dba205"}, "metrics": [{"type": "accuracy", "value": 73... | feature-extraction | {"parameters": {"I64": 512, "F32": 33360000}, "total": 33360512} | [
".gitattributes",
"1_Pooling/config.json",
"README.md",
"config.json",
"config_sentence_transformers.json",
"model.safetensors",
"modules.json",
"onnx/model.onnx",
"pytorch_model.bin",
"sentence_bert_config.json",
"special_tokens_map.json",
"tokenizer.json",
"tokenizer_config.json",
"vocab... | 5c38ec7c405ec4b44b94cc5a9bb96e735b38267a | [
"sentence-transformers",
"pytorch",
"onnx",
"safetensors",
"bert",
"feature-extraction",
"sentence-similarity",
"transformers",
"mteb",
"en",
"arxiv:2401.03462",
"arxiv:2312.15503",
"arxiv:2311.13534",
"arxiv:2310.07554",
"arxiv:2309.07597",
"license:mit",
"model-index",
"text-embe... | null | {"architectures": ["BertModel"], "model_type": "bert", "tokenizer_config": {"cls_token": "[CLS]", "mask_token": "[MASK]", "pad_token": "[PAD]", "sep_token": "[SEP]", "unk_token": "[UNK]"}} | {
"auto_model": "AutoModel",
"custom_class": null,
"pipeline_tag": "feature-extraction",
"processor": "AutoTokenizer"
} | {"language": ["en"], "license": "mit", "tags": ["sentence-transformers", "feature-extraction", "sentence-similarity", "transformers", "mteb"], "model-index": [{"name": "bge-small-en-v1.5", "results": [{"task": {"type": "Classification"}, "dataset": {"name": "MTEB AmazonCounterfactualClassification (en)", "type": "mteb/... | <h1 align="center">FlagEmbedding</h1>
<h4 align="center">
<p>
<a href=#model-list>Model List</a> |
<a href=#frequently-asked-questions>FAQ</a> |
<a href=#usage>Usage</a> |
<a href="#evaluation">Evaluation</a> |
<a href="#train">Train</a> |
<a href="#contact">Conta... | null | [
"mit"
] | null | [
"en"
] | 33,360,512 | null | null | [
"BertModel",
"AutoModel",
"bert"
] | [
"sentence-similarity",
"feature-extraction"
] | [
"text",
"multimodal"
] | [
"text"
] | [
"logits",
"embeddings"
] |
65770c3426ef61bbf101d4da | mistralai/Mistral-7B-Instruct-v0.2 | mistralai | null | 2,719,964 | 53,687,337 | False | 2023-12-11T13:18:44Z | 2025-07-24T16:57:21Z | transformers | 3,098 | 5 | null | text-generation | {"parameters": {"BF16": 7241732096}, "total": 7241732096} | [
".gitattributes",
"README.md",
"config.json",
"generation_config.json",
"model-00001-of-00003.safetensors",
"model-00002-of-00003.safetensors",
"model-00003-of-00003.safetensors",
"model.safetensors.index.json",
"pytorch_model-00001-of-00003.bin",
"pytorch_model-00002-of-00003.bin",
"pytorch_mod... | 63a8b081895390a26e140280378bc85ec8bce07a | [
"transformers",
"pytorch",
"safetensors",
"mistral",
"text-generation",
"finetuned",
"mistral-common",
"conversational",
"arxiv:2310.06825",
"license:apache-2.0",
"text-generation-inference",
"deploy:azure",
"region:us"
] | null | {"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{%- if messages[0]['role'] == 'system' %}\n {%- set system_message = messages[0]['content'] %}\n {%- set loop_messages = messages[1:] %}\n{%- else %}\n {%- set loop_messages = messages ... | {
"auto_model": "AutoModelForCausalLM",
"custom_class": null,
"pipeline_tag": "text-generation",
"processor": "AutoTokenizer"
} | {"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "apache-2.0", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": null, "tags": ["finetuned", "mistral-common"], "new_version": "mistralai/Mistral-7B-Ins... | null | null | [
"apache-2.0"
] | null | null | 7,241,732,096 | null | null | [
"AutoModelForCausalLM",
"mistral",
"MistralForCausalLM"
] | [
"text-generation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
658fb85235c41262d661dc48 | TinyLlama/TinyLlama-1.1B-Chat-v1.0 | TinyLlama | null | 2,817,808 | 35,656,120 | False | 2023-12-30T06:27:30Z | 2024-03-17T05:07:08Z | transformers | 1,556 | 5 | null | text-generation | {"parameters": {"BF16": 1100048384}, "total": 1100048384} | [
".gitattributes",
"README.md",
"config.json",
"eval_results.json",
"generation_config.json",
"model.safetensors",
"special_tokens_map.json",
"tokenizer.json",
"tokenizer.model",
"tokenizer_config.json"
] | fe8a4ea1ffedaf415f4da2f062534de366a451e6 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"conversational",
"en",
"dataset:cerebras/SlimPajama-627B",
"dataset:bigcode/starcoderdata",
"dataset:HuggingFaceH4/ultrachat_200k",
"dataset:HuggingFaceH4/ultrafeedback_binarized",
"license:apache-2.0",
"text-generation-inference",
... | null | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{% for message in messages %}\n{% if message['role'] == 'user' %}\n{{ '<|user|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'system' %}\n{{ '<|system|>\n' + message['content']... | {
"auto_model": "AutoModelForCausalLM",
"custom_class": null,
"pipeline_tag": "text-generation",
"processor": "AutoTokenizer"
} | {"base_model": null, "datasets": ["cerebras/SlimPajama-627B", "bigcode/starcoderdata", "HuggingFaceH4/ultrachat_200k", "HuggingFaceH4/ultrafeedback_binarized"], "eval_results": null, "language": ["en"], "library_name": null, "license": "apache-2.0", "license_name": null, "license_link": null, "metrics": null, "model_na... | <div align="center">
# TinyLlama-1.1B
</div>
https://github.com/jzhang38/TinyLlama
The TinyLlama project aims to **pretrain** a **1.1B Llama model on 3 trillion tokens**. With some proper optimization, we can achieve this within a span of "just" 90 days using 16 A100-40G GPUs 🚀🚀. The training has started on 2023-0... | null | [
"apache-2.0"
] | [
"cerebras/SlimPajama-627B",
"bigcode/starcoderdata",
"HuggingFaceH4/ultrachat_200k",
"HuggingFaceH4/ultrafeedback_binarized"
] | [
"en"
] | 1,100,048,384 | null | null | [
"AutoModelForCausalLM",
"llama",
"LlamaForCausalLM"
] | [
"text-generation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
65e60cda4b2e0f45e49dc03d | vikhyatk/moondream2 | vikhyatk | null | 4,627,664 | 24,499,658 | False | 2024-03-04T18:03:06Z | 2025-09-23T01:22:59Z | transformers | 1,399 | 5 | null | image-text-to-text | {"parameters": {"BF16": 1927237104}, "total": 1927237104} | [
".gitattributes",
"README.md",
"added_tokens.json",
"config.json",
"config.py",
"configuration_moondream.py",
"fourier_features.py",
"generation_config.json",
"handler.py",
"hf_moondream.py",
"image_crops.py",
"layers.py",
"lora.py",
"merges.txt",
"model.safetensors",
"moondream.py",
... | 6b714b26eea5cbd9f31e4edb2541c170afa935ba | [
"transformers",
"safetensors",
"moondream1",
"text-generation",
"image-text-to-text",
"custom_code",
"doi:10.57967/hf/6762",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | null | {"architectures": ["HfMoondream"], "auto_map": {"AutoConfig": "hf_moondream.HfConfig", "AutoModelForCausalLM": "hf_moondream.HfMoondream"}, "model_type": "moondream1", "tokenizer_config": {"bos_token": "<|endoftext|>", "eos_token": "<|endoftext|>", "unk_token": "<|endoftext|>"}} | {
"auto_model": "AutoModelForCausalLM",
"custom_class": "hf_moondream.HfMoondream",
"pipeline_tag": "text-generation",
"processor": null
} | {"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": null, "license": "apache-2.0", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "image-text-to-text", "tags": null, "new_version": "moondream/moondream3-preview"} | ⚠️ This repository contains the latest version of Moondream 2, our previous generation model. The latest version of Moondream is [Moondream 3 (Preview)](https://huggingface.co/moondream/moondream3-preview).
---
Moondream is a small vision language model designed to run efficiently everywhere.
[Website](https://moon... | null | [
"apache-2.0"
] | null | null | 1,927,237,104 | null | null | [
"HfMoondream",
"hf_moondream.HfMoondream",
"AutoModelForCausalLM",
"moondream1"
] | [
"image-text-to-text",
"text-generation"
] | [
"text",
"multimodal"
] | [
"text",
"image"
] | [
"text"
] |
6641487429be04778b9180f0 | google/paligemma-3b-pt-224 | google | null | 81,889 | 887,103 | manual | 2024-05-12T22:53:40Z | 2024-09-21T10:14:25Z | transformers | 426 | 5 | null | image-text-to-text | null | [
".gitattributes",
"README.md",
"added_tokens.json",
"config.json",
"generation_config.json",
"model-00001-of-00003.safetensors",
"model-00002-of-00003.safetensors",
"model-00003-of-00003.safetensors",
"model.safetensors.index.json",
"preprocessor_config.json",
"special_tokens_map.json",
"token... | 35e4f46485b4d07967e7e9935bc3786aad50687c | [
"transformers",
"safetensors",
"paligemma",
"image-text-to-text",
"arxiv:2310.09199",
"arxiv:2303.15343",
"arxiv:2403.08295",
"arxiv:1706.03762",
"arxiv:2010.11929",
"arxiv:2209.06794",
"arxiv:2209.04372",
"arxiv:2103.01913",
"arxiv:2205.12522",
"arxiv:2110.11624",
"arxiv:2108.03353",
... | null | {"architectures": ["PaliGemmaForConditionalGeneration"], "model_type": "paligemma", "tokenizer_config": {"bos_token": "<bos>", "eos_token": "<eos>", "pad_token": "<pad>", "unk_token": "<unk>", "use_default_system_prompt": false}} | {
"auto_model": "AutoModelForImageTextToText",
"custom_class": null,
"pipeline_tag": "image-text-to-text",
"processor": "AutoProcessor"
} | {"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "gemma", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "image-text-to-text", "tags": null, "extra_gated_heading": "Access PaliGemma on Hugging Face... | null | null | [
"gemma"
] | null | null | null | null | null | [
"PaliGemmaForConditionalGeneration",
"AutoModelForImageTextToText",
"paligemma"
] | [
"image-text-to-text"
] | [
"multimodal"
] | [
"text",
"image"
] | [
"text"
] |
66592ee81ba271cba98959cb | Tiiny/TurboSparse-Mixtral | Tiiny | null | 20 | 452 | False | 2024-05-31T01:59:04Z | 2024-06-14T02:13:18Z | transformers | 42 | 5 | null | feature-extraction | {"parameters": {"BF16": 51796250624}, "total": 51796250624} | [
".gitattributes",
"README.md",
"added_tokens.json",
"config.json",
"configuration_turbosparsemixtral.py",
"generation_config.json",
"model-00001-of-00021.safetensors",
"model-00002-of-00021.safetensors",
"model-00003-of-00021.safetensors",
"model-00004-of-00021.safetensors",
"model-00005-of-0002... | 706a33e3e53182035d2e8e2c7f4ed6d0fc3e272e | [
"transformers",
"safetensors",
"turbosparsemixtral",
"feature-extraction",
"custom_code",
"en",
"arxiv:2406.05955",
"license:apache-2.0",
"region:us"
] | null | {"architectures": ["TurboSparseMixtralForCausalLM"], "auto_map": {"AutoConfig": "configuration_turbosparsemixtral.TurboSparseMixtralConfig", "AutoModel": "modeling_turbosparsemixtral.TurboSparseMixtralForCausalLM", "AutoModelForCausalLM": "modeling_turbosparsemixtral.TurboSparseMixtralForCausalLM"}, "model_type": "turb... | {
"auto_model": "AutoModel",
"custom_class": "modeling_turbosparsemixtral.TurboSparseMixtralForCausalLM",
"pipeline_tag": "feature-extraction",
"processor": null
} | {"base_model": null, "datasets": null, "eval_results": null, "language": ["en"], "library_name": null, "license": "apache-2.0", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": null, "tags": null} | null | null | [
"apache-2.0"
] | null | [
"en"
] | 51,796,250,624 | null | null | [
"AutoModel",
"modeling_turbosparsemixtral.TurboSparseMixtralForCausalLM",
"TurboSparseMixtralForCausalLM",
"turbosparsemixtral"
] | [
"feature-extraction"
] | [
"multimodal"
] | [
"text"
] | [
"embeddings"
] |
66672b6f2e5f1c08ecf1c61d | openvla/openvla-7b | openvla | null | 1,196,859 | 11,449,520 | False | 2024-06-10T16:35:59Z | 2026-02-17T03:43:23Z | transformers | 197 | 5 | null | robotics | {"parameters": {"BF16": 7541237184}, "total": 7541237184} | [
".gitattributes",
"README.md",
"added_tokens.json",
"config.json",
"configuration_prismatic.py",
"generation_config.json",
"model-00001-of-00003.safetensors",
"model-00002-of-00003.safetensors",
"model-00003-of-00003.safetensors",
"model.safetensors.index.json",
"modeling_prismatic.py",
"prepr... | 47a0ec7fc4ec123775a391911046cf33cf9ed83f | [
"transformers",
"safetensors",
"openvla",
"feature-extraction",
"robotics",
"vla",
"image-text-to-text",
"multimodal",
"pretraining",
"custom_code",
"en",
"arxiv:2406.09246",
"license:mit",
"region:us"
] | null | {"architectures": ["OpenVLAForActionPrediction"], "auto_map": {"AutoConfig": "configuration_prismatic.OpenVLAConfig", "AutoModelForVision2Seq": "modeling_prismatic.OpenVLAForActionPrediction"}, "model_type": "openvla", "tokenizer_config": {"bos_token": "<s>", "eos_token": "</s>", "pad_token": "<PAD>", "unk_token": "<un... | {
"auto_model": "AutoModelForVision2Seq",
"custom_class": "modeling_prismatic.OpenVLAForActionPrediction",
"pipeline_tag": "feature-extraction",
"processor": null
} | {"base_model": null, "datasets": null, "eval_results": null, "language": ["en"], "library_name": "transformers", "license": "mit", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "robotics", "tags": ["robotics", "vla", "image-text-to-text", "multimodal", "pretraining"]} | # OpenVLA 7B
OpenVLA 7B (`openvla-7b`) is an open vision-language-action model trained on 970K robot manipulation episodes from the [Open X-Embodiment](https://robotics-transformer-x.github.io/) dataset.
The model takes language instructions and camera images as input and generates robot actions. It supports controll... | null | [
"mit"
] | null | [
"en"
] | 7,541,237,184 | null | null | [
"AutoModelForVision2Seq",
"modeling_prismatic.OpenVLAForActionPrediction",
"openvla",
"OpenVLAForActionPrediction"
] | [
"image-text-to-text",
"robotics",
"feature-extraction"
] | [
"multimodal"
] | [
"text",
"image"
] | [
"text",
"embeddings"
] |
666be1e54e6657b7d1a78f80 | deepseek-ai/DeepSeek-Coder-V2-Lite-Instruct | deepseek-ai | null | 319,323 | 5,373,292 | False | 2024-06-14T06:23:33Z | 2024-07-03T05:16:11Z | transformers | 570 | 5 | null | text-generation | {"parameters": {"BF16": 15706484224}, "total": 15706484224} | [
".gitattributes",
"README.md",
"config.json",
"configuration_deepseek.py",
"generation_config.json",
"model-00001-of-000004.safetensors",
"model-00002-of-000004.safetensors",
"model-00003-of-000004.safetensors",
"model-00004-of-000004.safetensors",
"model.safetensors.index.json",
"modeling_deeps... | e434a23f91ba5b4923cf6c9d9a238eb4a08e3a11 | [
"transformers",
"safetensors",
"deepseek_v2",
"text-generation",
"conversational",
"custom_code",
"arxiv:2401.06066",
"license:other",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | null | {"architectures": ["DeepseekV2ForCausalLM"], "auto_map": {"AutoConfig": "configuration_deepseek.DeepseekV2Config", "AutoModel": "modeling_deepseek.DeepseekV2Model", "AutoModelForCausalLM": "modeling_deepseek.DeepseekV2ForCausalLM"}, "model_type": "deepseek_v2", "tokenizer_config": {"bos_token": {"__type": "AddedToken",... | {
"auto_model": "AutoModelForCausalLM",
"custom_class": null,
"pipeline_tag": "text-generation",
"processor": "AutoTokenizer"
} | {"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": null, "license": "other", "license_name": "deepseek-license", "license_link": "LICENSE", "metrics": null, "model_name": null, "pipeline_tag": null, "tags": null} | <!-- markdownlint-disable first-line-h1 -->
<!-- markdownlint-disable html -->
<!-- markdownlint-disable no-duplicate-header -->
<div align="center">
<img src="https://github.com/deepseek-ai/DeepSeek-V2/blob/main/figures/logo.svg?raw=true" width="60%" alt="DeepSeek-V2" />
</div>
<hr>
<div align="center" style="line-... | null | [
"other",
"deepseek-license",
"LICENSE"
] | null | null | 15,706,484,224 | null | null | [
"AutoModelForCausalLM",
"DeepseekV2ForCausalLM",
"deepseek_v2"
] | [
"text-generation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
66962a41534f204a2b9982fd | google/gemma-2-2b | google | null | 536,112 | 24,639,131 | manual | 2024-07-16T08:07:29Z | 2024-08-07T18:25:49Z | transformers | 636 | 5 | null | text-generation | null | [
".gitattributes",
"README.md",
"config.json",
"generation_config.json",
"model-00001-of-00003.safetensors",
"model-00002-of-00003.safetensors",
"model-00003-of-00003.safetensors",
"model.safetensors.index.json",
"special_tokens_map.json",
"tokenizer.json",
"tokenizer.model",
"tokenizer_config.... | c5ebcd40d208330abc697524c919956e692655cf | [
"transformers",
"safetensors",
"gemma2",
"text-generation",
"arxiv:2009.03300",
"arxiv:1905.07830",
"arxiv:1911.11641",
"arxiv:1904.09728",
"arxiv:1905.10044",
"arxiv:1907.10641",
"arxiv:1811.00937",
"arxiv:1809.02789",
"arxiv:1911.01547",
"arxiv:1705.03551",
"arxiv:2107.03374",
"arxiv... | null | {"architectures": ["Gemma2ForCausalLM"], "model_type": "gemma2", "tokenizer_config": {"bos_token": "<bos>", "eos_token": "<eos>", "pad_token": "<pad>", "unk_token": "<unk>", "use_default_system_prompt": false}} | {
"auto_model": "AutoModelForCausalLM",
"custom_class": null,
"pipeline_tag": "text-generation",
"processor": "AutoTokenizer"
} | {"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "gemma", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "text-generation", "tags": null, "extra_gated_heading": "Access Gemma on Hugging Face", "ext... | null | null | [
"gemma"
] | null | null | null | null | null | [
"AutoModelForCausalLM",
"gemma2",
"Gemma2ForCausalLM"
] | [
"text-generation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
669650bb11dbbf600cf4dcf0 | google/gemma-2-2b-it | google | {
"models": [
{
"_id": "66962a41534f204a2b9982fd",
"id": "google/gemma-2-2b"
}
],
"relation": "finetune"
} | 390,438 | 8,672,814 | manual | 2024-07-16T10:51:39Z | 2024-08-27T19:41:44Z | transformers | 1,310 | 5 | null | text-generation | {"parameters": {"BF16": 2614341888}, "total": 2614341888} | [
".gitattributes",
"README.md",
"config.json",
"generation_config.json",
"model-00001-of-00002.safetensors",
"model-00002-of-00002.safetensors",
"model.safetensors.index.json",
"special_tokens_map.json",
"tokenizer.json",
"tokenizer.model",
"tokenizer_config.json"
] | 299a8560bedf22ed1c72a8a11e7dce4a7f9f51f8 | [
"transformers",
"safetensors",
"gemma2",
"text-generation",
"conversational",
"arxiv:2009.03300",
"arxiv:1905.07830",
"arxiv:1911.11641",
"arxiv:1904.09728",
"arxiv:1905.10044",
"arxiv:1907.10641",
"arxiv:1811.00937",
"arxiv:1809.02789",
"arxiv:1911.01547",
"arxiv:1705.03551",
"arxiv:2... | null | {"architectures": ["Gemma2ForCausalLM"], "model_type": "gemma2", "tokenizer_config": {"bos_token": "<bos>", "chat_template": "{{ bos_token }}{% if messages[0]['role'] == 'system' %}{{ raise_exception('System role not supported') }}{% endif %}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0... | {
"auto_model": "AutoModelForCausalLM",
"custom_class": null,
"pipeline_tag": "text-generation",
"processor": "AutoTokenizer"
} | {"base_model": "google/gemma-2-2b", "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "gemma", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "text-generation", "tags": ["conversational"], "extra_gated_heading": "Access... | null | null | [
"gemma"
] | null | null | 2,614,341,888 | null | null | [
"AutoModelForCausalLM",
"gemma2",
"Gemma2ForCausalLM"
] | [
"text-generation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
6722d34c3189f1de8ad8ce22 | HuggingFaceTB/SmolLM2-135M | HuggingFaceTB | null | 1,056,461 | 10,564,948 | False | 2024-10-31T00:46:04Z | 2025-02-06T10:34:41Z | transformers | 177 | 5 | null | text-generation | {"parameters": {"BF16": 134515008}, "total": 134515008} | [
".gitattributes",
"README.md",
"config.json",
"generation_config.json",
"merges.txt",
"model.safetensors",
"special_tokens_map.json",
"tokenizer.json",
"tokenizer_config.json",
"vocab.json"
] | 93efa2f097d58c2a74874c7e644dbc9b0cee75a2 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"en",
"arxiv:2502.02737",
"license:apache-2.0",
"text-generation-inference",
"endpoints_compatible",
"deploy:azure",
"region:us"
] | null | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|endoftext|>", "eos_token": "<|endoftext|>", "unk_token": "<|endoftext|>"}} | {
"auto_model": "AutoModelForCausalLM",
"custom_class": null,
"pipeline_tag": "text-generation",
"processor": "AutoTokenizer"
} | {"base_model": null, "datasets": null, "eval_results": null, "language": ["en"], "library_name": "transformers", "license": "apache-2.0", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": null, "tags": null} | # SmolLM2

## Table of Contents
1. [Model Summary](##model-summary)
2. [Limitations](##limitations)
3. [Training](##training)
4. [License](##license)
5. [Citation](##citation)
## Model Summary
Smo... | null | [
"apache-2.0"
] | null | [
"en"
] | 134,515,008 | null | null | [
"AutoModelForCausalLM",
"llama",
"LlamaForCausalLM"
] | [
"text-generation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
678e145ed35c6f68a26df940 | deepseek-ai/DeepSeek-R1-Distill-Qwen-7B | deepseek-ai | null | 620,266 | 12,793,856 | False | 2025-01-20T09:16:14Z | 2025-02-24T03:32:20Z | transformers | 799 | 5 | null | text-generation | {"parameters": {"BF16": 7615616512}, "total": 7615616512} | [
".gitattributes",
"LICENSE",
"README.md",
"config.json",
"figures/benchmark.jpg",
"generation_config.json",
"model-00001-of-000002.safetensors",
"model-00002-of-000002.safetensors",
"model.safetensors.index.json",
"tokenizer.json",
"tokenizer_config.json"
] | 916b56a44061fd5cd7d6a8fb632557ed4f724f60 | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"conversational",
"arxiv:2501.12948",
"license:mit",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | null | {"architectures": ["Qwen2ForCausalLM"], "model_type": "qwen2", "tokenizer_config": {"bos_token": {"__type": "AddedToken", "content": "<\uff5cbegin\u2581of\u2581sentence\uff5c>", "lstrip": false, "normalized": true, "rstrip": false, "single_word": false}, "eos_token": {"__type": "AddedToken", "content": "<\uff5cend\u258... | {
"auto_model": "AutoModelForCausalLM",
"custom_class": null,
"pipeline_tag": "text-generation",
"processor": "AutoTokenizer"
} | {"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "mit", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": null, "tags": null} | # DeepSeek-R1
<!-- markdownlint-disable first-line-h1 -->
<!-- markdownlint-disable html -->
<!-- markdownlint-disable no-duplicate-header -->
<div align="center">
<img src="https://github.com/deepseek-ai/DeepSeek-V2/blob/main/figures/logo.svg?raw=true" width="60%" alt="DeepSeek-V3" />
</div>
<hr>
<div align="center... | null | [
"mit"
] | null | null | 7,615,616,512 | null | null | [
"AutoModelForCausalLM",
"Qwen2ForCausalLM",
"qwen2"
] | [
"text-generation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
678e14e3ccefc7605ae5d807 | deepseek-ai/DeepSeek-R1-Distill-Qwen-14B | deepseek-ai | null | 567,723 | 6,242,713 | False | 2025-01-20T09:18:27Z | 2025-02-24T03:31:45Z | transformers | 615 | 5 | null | text-generation | {"parameters": {"BF16": 14770033664}, "total": 14770033664} | [
".gitattributes",
"LICENSE",
"README.md",
"config.json",
"figures/benchmark.jpg",
"generation_config.json",
"model-00001-of-000004.safetensors",
"model-00002-of-000004.safetensors",
"model-00003-of-000004.safetensors",
"model-00004-of-000004.safetensors",
"model.safetensors.index.json",
"token... | 1df8507178afcc1bef68cd8c393f61a886323761 | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"conversational",
"arxiv:2501.12948",
"license:mit",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | null | {"architectures": ["Qwen2ForCausalLM"], "model_type": "qwen2", "tokenizer_config": {"bos_token": {"__type": "AddedToken", "content": "<\uff5cbegin\u2581of\u2581sentence\uff5c>", "lstrip": false, "normalized": true, "rstrip": false, "single_word": false}, "eos_token": {"__type": "AddedToken", "content": "<\uff5cend\u258... | {
"auto_model": "AutoModelForCausalLM",
"custom_class": null,
"pipeline_tag": "text-generation",
"processor": "AutoTokenizer"
} | {"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "mit", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": null, "tags": null} | # DeepSeek-R1
<!-- markdownlint-disable first-line-h1 -->
<!-- markdownlint-disable html -->
<!-- markdownlint-disable no-duplicate-header -->
<div align="center">
<img src="https://github.com/deepseek-ai/DeepSeek-V2/blob/main/figures/logo.svg?raw=true" width="60%" alt="DeepSeek-V3" />
</div>
<hr>
<div align="center... | null | [
"mit"
] | null | null | 14,770,033,664 | null | null | [
"AutoModelForCausalLM",
"Qwen2ForCausalLM",
"qwen2"
] | [
"text-generation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
68125124ea07c4672522dca0 | darkc0de/XortronCriminalComputingConfig | darkc0de | {
"models": [
{
"_id": "67aacebd6b842ad4f7d26c64",
"id": "TroyDoesAI/BlackSheep-24B"
},
{
"_id": "68125124ea07c4672522dca0",
"id": "darkc0de/XortronCriminalComputingConfig"
}
],
"relation": "merge"
} | 668 | 7,683 | False | 2025-04-30T16:34:44Z | 2026-03-24T20:22:40Z | transformers | 149 | 5 | null | text-generation | {"parameters": {"F16": 23572403200}, "total": 23572403200} | [
"README.md",
"config.json",
"gitattributes",
"mergekit_config.yml",
"model-00001-of-00010.safetensors",
"model-00002-of-00010.safetensors",
"model-00003-of-00010.safetensors",
"model-00004-of-00010.safetensors",
"model-00005-of-00010.safetensors",
"model-00006-of-00010.safetensors",
"model-00007... | 940cf735f387b133bd8178a4db89b0d3c6f1ccd2 | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"mergekit",
"merge",
"uncensored",
"harmful",
"conversational",
"en",
"base_model:TroyDoesAI/BlackSheep-24B",
"base_model:merge:TroyDoesAI/BlackSheep-24B",
"base_model:darkc0de/XortronCriminalComputingConfig",
"base_model:merge:d... | null | {"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{%- set today = strftime_now(\"%Y-%m-%d\") %}\n{%- set default_system_message = \"You are Xortron. You are an uncensored AI model. You were created by Dr. Buddy Glass and Dr. Sonny DeSorbo at th... | {
"auto_model": "AutoModelForCausalLM",
"custom_class": null,
"pipeline_tag": "text-generation",
"processor": "AutoTokenizer"
} | {"base_model": ["darkc0de/XortronCriminalComputing", "TroyDoesAI/BlackSheep-24B", "darkc0de/XortronCriminalComputingConfig"], "datasets": null, "eval_results": null, "language": ["en"], "library_name": "transformers", "license": "apache-2.0", "license_name": null, "license_link": null, "metrics": null, "model_name": nu... | null | null | [
"apache-2.0"
] | null | [
"en"
] | 23,572,403,200 | null | null | [
"AutoModelForCausalLM",
"mistral",
"MistralForCausalLM"
] | [
"text-generation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
688b1597e5e83e19d1b3238a | Qwen/Qwen3-Coder-30B-A3B-Instruct | Qwen | null | 1,134,092 | 5,693,237 | False | 2025-07-31T07:04:55Z | 2025-12-03T08:05:17Z | transformers | 985 | 5 | null | text-generation | {"parameters": {"BF16": 30532122624}, "total": 30532122624} | [
".gitattributes",
"LICENSE",
"README.md",
"chat_template.jinja",
"config.json",
"generation_config.json",
"merges.txt",
"model-00001-of-00016.safetensors",
"model-00002-of-00016.safetensors",
"model-00003-of-00016.safetensors",
"model-00004-of-00016.safetensors",
"model-00005-of-00016.safetens... | b2cff646eb4bb1d68355c01b18ae02e7cf42d120 | [
"transformers",
"safetensors",
"qwen3_moe",
"text-generation",
"conversational",
"arxiv:2505.09388",
"license:apache-2.0",
"endpoints_compatible",
"deploy:azure",
"region:us"
] | null | {"architectures": ["Qwen3MoeForCausalLM"], "model_type": "qwen3_moe", "tokenizer_config": {"bos_token": null, "chat_template": "{% macro render_extra_keys(json_dict, handled_keys) %}\n {%- if json_dict is mapping %}\n {%- for json_key in json_dict if json_key not in handled_keys %}\n {%- if json_di... | {
"auto_model": "AutoModelForCausalLM",
"custom_class": null,
"pipeline_tag": "text-generation",
"processor": "AutoTokenizer"
} | {"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "apache-2.0", "license_name": null, "license_link": "https://huggingface.co/Qwen/Qwen3-Coder-30B-A3B-Instruct/blob/main/LICENSE", "metrics": null, "model_name": null, "pipeline_tag": "text-generatio... | # Qwen3-Coder-30B-A3B-Instruct
<a href="https://chat.qwen.ai/" target="_blank" style="margin: 2px;">
<img alt="Chat" src="https://img.shields.io/badge/%F0%9F%92%9C%EF%B8%8F%20Qwen%20Chat%20-536af5" style="display: inline-block; vertical-align: middle;"/>
</a>
## Highlights
**Qwen3-Coder** is available in multiple... | null | [
"apache-2.0",
"https://huggingface.co/Qwen/Qwen3-Coder-30B-A3B-Instruct/blob/main/LICENSE"
] | null | null | 30,532,122,624 | null | null | [
"qwen3_moe",
"AutoModelForCausalLM",
"Qwen3MoeForCausalLM"
] | [
"text-generation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
689252773b8900ddb9116aed | google/gemma-3-270m | google | null | 87,092 | 830,635 | manual | 2025-08-05T18:50:31Z | 2025-08-14T07:35:01Z | transformers | 1,001 | 5 | null | text-generation | {"parameters": {"BF16": 268098176}, "total": 268098176} | [
".gitattributes",
"README.md",
"added_tokens.json",
"config.json",
"generation_config.json",
"model.safetensors",
"special_tokens_map.json",
"tokenizer.json",
"tokenizer.model",
"tokenizer_config.json"
] | 9b0cfec892e2bc2afd938c98eabe4e4a7b1e0ca1 | [
"transformers",
"safetensors",
"gemma3_text",
"text-generation",
"gemma3",
"gemma",
"google",
"arxiv:2503.19786",
"arxiv:1905.07830",
"arxiv:1905.10044",
"arxiv:1911.11641",
"arxiv:1705.03551",
"arxiv:1911.01547",
"arxiv:1907.10641",
"arxiv:2311.07911",
"arxiv:2311.12022",
"arxiv:241... | null | {"architectures": ["Gemma3ForCausalLM"], "model_type": "gemma3_text", "tokenizer_config": {"bos_token": "<bos>", "eos_token": "<eos>", "pad_token": "<pad>", "unk_token": "<unk>", "use_default_system_prompt": false}} | {
"auto_model": "AutoModelForCausalLM",
"custom_class": null,
"pipeline_tag": "text-generation",
"processor": "AutoTokenizer"
} | {"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "gemma", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "text-generation", "tags": ["gemma3", "gemma", "google"], "extra_gated_heading": "Access Gem... | null | null | [
"gemma"
] | null | null | 268,098,176 | null | null | [
"gemma3_text",
"AutoModelForCausalLM",
"Gemma3ForCausalLM"
] | [
"text-generation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
68e51bdbe980d4b93323e6b8 | LiquidAI/LFM2-8B-A1B | LiquidAI | null | 47,879 | 142,143 | False | 2025-10-07T13:55:39Z | 2026-03-04T12:20:04Z | transformers | 342 | 5 | null | text-generation | {"parameters": {"F32": 704, "BF16": 8339929856}, "total": 8339929856} | [
".gitattributes",
"LICENSE",
"README.md",
"chat_template.jinja",
"config.json",
"generation_config.json",
"model-00001-of-00004.safetensors",
"model-00002-of-00004.safetensors",
"model-00003-of-00004.safetensors",
"model-00004-of-00004.safetensors",
"model.safetensors.index.json",
"special_tok... | e2277e438c87cef87a57a8d437a7e74d58717939 | [
"transformers",
"safetensors",
"lfm2_moe",
"text-generation",
"liquid",
"lfm2",
"edge",
"moe",
"conversational",
"en",
"ar",
"zh",
"fr",
"de",
"ja",
"ko",
"es",
"arxiv:2511.23404",
"license:other",
"endpoints_compatible",
"deploy:azure",
"region:us"
] | null | {"architectures": ["Lfm2MoeForCausalLM"], "model_type": "lfm2_moe", "tokenizer_config": {"bos_token": "<|startoftext|>", "eos_token": "<|im_end|>", "pad_token": "<|pad|>", "use_default_system_prompt": false}, "chat_template_jinja": "{{- bos_token -}}{%- set system_prompt = \"\" -%}{%- set ns = namespace(system_prompt=\... | {
"auto_model": "AutoModelForCausalLM",
"custom_class": null,
"pipeline_tag": "text-generation",
"processor": "AutoTokenizer"
} | {"base_model": null, "datasets": null, "eval_results": null, "language": ["en", "ar", "zh", "fr", "de", "ja", "ko", "es"], "library_name": "transformers", "license": "other", "license_name": "lfm1.0", "license_link": "LICENSE", "metrics": null, "model_name": null, "pipeline_tag": "text-generation", "tags": ["liquid", "... | null | null | [
"other",
"lfm1.0",
"LICENSE"
] | null | [
"en",
"ar",
"zh",
"fr",
"de",
"ja",
"ko",
"es"
] | 8,339,929,856 | null | null | [
"AutoModelForCausalLM",
"lfm2_moe",
"Lfm2MoeForCausalLM"
] | [
"text-generation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
68ea05ea8bfbf816c8e9ad2e | Qwen/Qwen3-VL-4B-Instruct | Qwen | null | 2,023,717 | 5,797,256 | False | 2025-10-11T07:23:22Z | 2025-10-15T16:15:55Z | transformers | 361 | 5 | null | image-text-to-text | {"parameters": {"BF16": 4437815808}, "total": 4437815808} | [
".gitattributes",
"README.md",
"chat_template.json",
"config.json",
"generation_config.json",
"merges.txt",
"model-00001-of-00002.safetensors",
"model-00002-of-00002.safetensors",
"model.safetensors.index.json",
"preprocessor_config.json",
"tokenizer.json",
"tokenizer_config.json",
"video_pr... | ebb281ec70b05090aa6165b016eac8ec08e71b17 | [
"transformers",
"safetensors",
"qwen3_vl",
"image-text-to-text",
"conversational",
"arxiv:2505.09388",
"arxiv:2502.13923",
"arxiv:2409.12191",
"arxiv:2308.12966",
"license:apache-2.0",
"endpoints_compatible",
"deploy:azure",
"region:us"
] | null | {"architectures": ["Qwen3VLForConditionalGeneration"], "model_type": "qwen3_vl", "processor_config": {"chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0].role == 'system' %}\n {%- if messages[0].content is string %}\n {{- messages[0].content }}\n {%- el... | {
"auto_model": "AutoModelForImageTextToText",
"custom_class": null,
"pipeline_tag": "image-text-to-text",
"processor": "AutoProcessor"
} | {"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "apache-2.0", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "image-text-to-text", "tags": null} | null | null | [
"apache-2.0"
] | null | null | 4,437,815,808 | null | null | [
"AutoModelForImageTextToText",
"qwen3_vl",
"Qwen3VLForConditionalGeneration"
] | [
"image-text-to-text"
] | [
"multimodal"
] | [
"text",
"image"
] | [
"text"
] |
68f6d55d42953c4ac0f8890f | datalab-to/chandra | datalab-to | null | 164,256 | 1,120,608 | False | 2025-10-21T00:35:41Z | 2026-03-26T15:55:54Z | transformers | 510 | 5 | null | image-text-to-text | {"parameters": {"BF16": 8767123696}, "total": 8767123696} | [
".gitattributes",
"LICENSE",
"README.md",
"added_tokens.json",
"bench.png",
"chat_template.jinja",
"config.json",
"generation_config.json",
"handwritten_form.png",
"merges.txt",
"model-00001-of-00004.safetensors",
"model-00002-of-00004.safetensors",
"model-00003-of-00004.safetensors",
"mod... | bd40c21576564d31ba99f204e28ccd5c4c1751fa | [
"transformers",
"safetensors",
"qwen3_vl",
"image-text-to-text",
"ocr",
"vlm",
"conversational",
"license:openrail",
"eval-results",
"endpoints_compatible",
"deploy:azure",
"region:us"
] | null | {"architectures": ["Qwen3VLForConditionalGeneration"], "model_type": "qwen3_vl", "tokenizer_config": {"bos_token": null, "eos_token": "<|im_end|>", "pad_token": "<|endoftext|>", "unk_token": null}, "chat_template_jinja": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0].role == 'system' %}\n ... | {
"auto_model": "AutoModelForImageTextToText",
"custom_class": null,
"pipeline_tag": "image-text-to-text",
"processor": "AutoProcessor"
} | {"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "openrail", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": null, "tags": ["ocr", "vlm"], "new_version": "datalab-to/chandra-ocr-2"} | null | null | [
"openrail"
] | null | null | 8,767,123,696 | null | null | [
"AutoModelForImageTextToText",
"qwen3_vl",
"Qwen3VLForConditionalGeneration"
] | [
"image-text-to-text"
] | [
"multimodal"
] | [
"text",
"image"
] | [
"text"
] |
621ffdc136468d709f17adb7 | facebook/bart-large-mnli | facebook | null | 3,358,415 | 143,603,374 | False | 2022-03-02T23:29:05Z | 2023-09-05T14:49:34Z | transformers | 1,552 | 4 | null | zero-shot-classification | {"parameters": {"F32": 407344133}, "total": 407344133} | [
".gitattributes",
"README.md",
"config.json",
"flax_model.msgpack",
"merges.txt",
"model.safetensors",
"pytorch_model.bin",
"rust_model.ot",
"tokenizer.json",
"tokenizer_config.json",
"vocab.json"
] | d7645e127eaf1aefc7862fd59a17a5aa8558b8ce | [
"transformers",
"pytorch",
"jax",
"rust",
"safetensors",
"bart",
"text-classification",
"zero-shot-classification",
"dataset:multi_nli",
"arxiv:1910.13461",
"arxiv:1909.00161",
"license:mit",
"endpoints_compatible",
"deploy:azure",
"region:us"
] | null | {"architectures": ["BartForSequenceClassification"], "model_type": "bart", "tokenizer_config": {}} | {
"auto_model": "AutoModelForSequenceClassification",
"custom_class": null,
"pipeline_tag": "text-classification",
"processor": "AutoTokenizer"
} | {"base_model": null, "datasets": ["multi_nli"], "eval_results": null, "language": null, "library_name": null, "license": "mit", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "zero-shot-classification", "tags": null, "thumbnail": "https://huggingface.co/front/thumbnails... | # bart-large-mnli
This is the checkpoint for [bart-large](https://huggingface.co/facebook/bart-large) after being trained on the [MultiNLI (MNLI)](https://huggingface.co/datasets/multi_nli) dataset.
Additional information about this model:
- The [bart-large](https://huggingface.co/facebook/bart-large) model page
- [B... | null | [
"mit"
] | [
"multi_nli"
] | null | 407,344,133 | null | null | [
"BartForSequenceClassification",
"bart",
"AutoModelForSequenceClassification"
] | [
"zero-shot-classification",
"text-classification"
] | [
"text"
] | [
"text"
] | [
"logits"
] |
621ffdc136468d709f180297 | sentence-transformers/all-mpnet-base-v2 | sentence-transformers | null | 29,268,568 | 1,158,149,103 | False | 2022-03-02T23:29:05Z | 2025-08-19T10:14:25Z | sentence-transformers | 1,262 | 4 | null | sentence-similarity | {"parameters": {"I64": 514, "F32": 109486464}, "total": 109486978} | [
".gitattributes",
"1_Pooling/config.json",
"README.md",
"config.json",
"config_sentence_transformers.json",
"data_config.json",
"model.safetensors",
"modules.json",
"onnx/model.onnx",
"onnx/model_O1.onnx",
"onnx/model_O2.onnx",
"onnx/model_O3.onnx",
"onnx/model_O4.onnx",
"onnx/model_qint8_... | e8c3b32edf5434bc2275fc9bab85f82640a19130 | [
"sentence-transformers",
"pytorch",
"onnx",
"safetensors",
"openvino",
"mpnet",
"fill-mask",
"feature-extraction",
"sentence-similarity",
"transformers",
"text-embeddings-inference",
"en",
"dataset:s2orc",
"dataset:flax-sentence-embeddings/stackexchange_xml",
"dataset:ms_marco",
"datas... | null | {"architectures": ["MPNetForMaskedLM"], "model_type": "mpnet", "tokenizer_config": {"bos_token": "<s>", "eos_token": "</s>", "sep_token": "</s>", "cls_token": "<s>", "unk_token": "[UNK]", "pad_token": "<pad>", "mask_token": "<mask>"}} | {
"auto_model": "AutoModelForMaskedLM",
"custom_class": null,
"pipeline_tag": "fill-mask",
"processor": "AutoTokenizer"
} | {"base_model": null, "datasets": ["s2orc", "flax-sentence-embeddings/stackexchange_xml", "ms_marco", "gooaq", "yahoo_answers_topics", "code_search_net", "search_qa", "eli5", "snli", "multi_nli", "wikihow", "natural_questions", "trivia_qa", "embedding-data/sentence-compression", "embedding-data/flickr30k-captions", "emb... | # all-mpnet-base-v2
This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search.
## Usage (Sentence-Transformers)
Using this model becomes easy when you have [sentence-transformers](h... | null | [
"apache-2.0"
] | [
"s2orc",
"flax-sentence-embeddings/stackexchange_xml",
"ms_marco",
"gooaq",
"yahoo_answers_topics",
"code_search_net",
"search_qa",
"eli5",
"snli",
"multi_nli",
"wikihow",
"natural_questions",
"trivia_qa",
"embedding-data/sentence-compression",
"embedding-data/flickr30k-captions",
"emb... | [
"en"
] | 109,486,978 | null | null | [
"MPNetForMaskedLM",
"AutoModelForMaskedLM",
"mpnet"
] | [
"fill-mask",
"sentence-similarity",
"feature-extraction"
] | [
"text",
"multimodal"
] | [
"text"
] | [
"logits",
"embeddings"
] |
63314bef9e3604f3f17e5b9e | openai/whisper-small | openai | null | 1,779,882 | 118,833,474 | False | 2022-09-26T06:51:27Z | 2024-02-29T10:57:38Z | transformers | 544 | 4 | [{"name": "whisper-small", "results": [{"task": {"name": "Automatic Speech Recognition", "type": "automatic-speech-recognition"}, "dataset": {"name": "LibriSpeech (clean)", "type": "librispeech_asr", "config": "clean", "split": "test", "args": {"language": "en"}}, "metrics": [{"name": "Test WER", "type": "wer", "value"... | automatic-speech-recognition | null | [
".gitattributes",
"README.md",
"added_tokens.json",
"config.json",
"flax_model.msgpack",
"generation_config.json",
"merges.txt",
"model.safetensors",
"normalizer.json",
"preprocessor_config.json",
"pytorch_model.bin",
"special_tokens_map.json",
"tf_model.h5",
"tokenizer.json",
"tokenizer... | 973afd24965f72e36ca33b3055d56a652f456b4d | [
"transformers",
"pytorch",
"tf",
"jax",
"safetensors",
"whisper",
"automatic-speech-recognition",
"audio",
"hf-asr-leaderboard",
"en",
"zh",
"de",
"es",
"ru",
"ko",
"fr",
"ja",
"pt",
"tr",
"pl",
"ca",
"nl",
"ar",
"sv",
"it",
"id",
"hi",
"fi",
"vi",
"he",
"... | null | {"architectures": ["WhisperForConditionalGeneration"], "model_type": "whisper", "tokenizer_config": {"bos_token": "<|endoftext|>", "eos_token": "<|endoftext|>", "pad_token": "<|endoftext|>", "unk_token": "<|endoftext|>"}} | {
"auto_model": "AutoModelForSpeechSeq2Seq",
"custom_class": null,
"pipeline_tag": "automatic-speech-recognition",
"processor": "AutoProcessor"
} | {"language": ["en", "zh", "de", "es", "ru", "ko", "fr", "ja", "pt", "tr", "pl", "ca", "nl", "ar", "sv", "it", "id", "hi", "fi", "vi", "he", "uk", "el", "ms", "cs", "ro", "da", "hu", "ta", "no", "th", "ur", "hr", "bg", "lt", "la", "mi", "ml", "cy", "sk", "te", "fa", "lv", "bn", "sr", "az", "sl", "kn", "et", "mk", "br", ... | # Whisper
Whisper is a pre-trained model for automatic speech recognition (ASR) and speech translation. Trained on 680k hours
of labelled data, Whisper models demonstrate a strong ability to generalise to many datasets and domains **without** the need
for fine-tuning.
Whisper was proposed in the paper [Robust Speec... | null | [
"apache-2.0"
] | null | [
"en",
"zh",
"de",
"es",
"ru",
"ko",
"fr",
"ja",
"pt",
"tr",
"pl",
"ca",
"nl",
"ar",
"sv",
"it",
"id",
"hi",
"fi",
"vi",
"he",
"uk",
"el",
"ms",
"cs",
"ro",
"da",
"hu",
"ta",
"no",
"th",
"ur",
"hr",
"bg",
"lt",
"la",
"mi",
"ml",
"cy",
"sk"... | null | null | null | [
"WhisperForConditionalGeneration",
"whisper",
"AutoModelForSpeechSeq2Seq"
] | [
"automatic-speech-recognition"
] | [
"multimodal"
] | [
"audio"
] | [
"text"
] |
65c5577b1080431ea9e083cd | google/gemma-7b | google | null | 30,631 | 2,821,063 | manual | 2024-02-08T22:36:43Z | 2024-06-27T14:09:40Z | transformers | 3,290 | 4 | null | text-generation | null | [
".gitattributes",
"README.md",
"config.json",
"examples/example_fsdp.py",
"examples/example_sft_qlora.py",
"examples/notebook_sft_peft.ipynb",
"gemma-7b.gguf",
"generation_config.json",
"model-00001-of-00004.safetensors",
"model-00002-of-00004.safetensors",
"model-00003-of-00004.safetensors",
... | ff6768d9368919a1f025a54f9f5aa0ee591730bb | [
"transformers",
"safetensors",
"gguf",
"gemma",
"text-generation",
"arxiv:2305.14314",
"arxiv:2312.11805",
"arxiv:2009.03300",
"arxiv:1905.07830",
"arxiv:1911.11641",
"arxiv:1904.09728",
"arxiv:1905.10044",
"arxiv:1907.10641",
"arxiv:1811.00937",
"arxiv:1809.02789",
"arxiv:1911.01547",... | {"total": 8538074112, "architecture": "gemma", "context_length": 8192, "bos_token": "<bos>", "eos_token": "<eos>"} | {"architectures": ["GemmaForCausalLM"], "model_type": "gemma", "tokenizer_config": {"bos_token": "<bos>", "eos_token": "<eos>", "pad_token": "<pad>", "unk_token": "<unk>", "use_default_system_prompt": false}} | {
"auto_model": "AutoModelForCausalLM",
"custom_class": null,
"pipeline_tag": "text-generation",
"processor": "AutoTokenizer"
} | {"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "gemma", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": null, "tags": null, "extra_gated_heading": "Access Gemma on Hugging Face", "extra_gated_prom... | null | null | [
"gemma"
] | null | null | null | 8,538,074,112 | null | [
"AutoModelForCausalLM",
"gemma",
"GemmaForCausalLM"
] | [
"text-generation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
65c7188324e59a0e4f89d863 | nomic-ai/nomic-embed-text-v1.5 | nomic-ai | null | 10,410,034 | 60,578,496 | False | 2024-02-10T06:32:35Z | 2025-07-21T17:44:14Z | sentence-transformers | 785 | 4 | [{"name": "epoch_0_model", "results": [{"task": {"type": "Classification"}, "dataset": {"type": "mteb/amazon_counterfactual", "name": "MTEB AmazonCounterfactualClassification (en)", "config": "en", "split": "test", "revision": "e8379541af4e31359cca9fbcf4b00f2671dba205"}, "metrics": [{"type": "accuracy", "value": 75.208... | sentence-similarity | {"parameters": {"F32": 136731648}, "total": 136731648} | [
".gitattributes",
"1_Pooling/config.json",
"README.md",
"config.json",
"config_sentence_transformers.json",
"model.safetensors",
"modules.json",
"onnx/model.onnx",
"onnx/model_bnb4.onnx",
"onnx/model_fp16.onnx",
"onnx/model_int8.onnx",
"onnx/model_q4.onnx",
"onnx/model_q4f16.onnx",
"onnx/m... | e5cf08aadaa33385f5990def41f7a23405aec398 | [
"sentence-transformers",
"onnx",
"safetensors",
"nomic_bert",
"feature-extraction",
"sentence-similarity",
"mteb",
"transformers",
"transformers.js",
"custom_code",
"en",
"arxiv:2402.01613",
"arxiv:2205.13147",
"license:apache-2.0",
"model-index",
"eval-results",
"text-embeddings-inf... | null | {"architectures": ["NomicBertModel"], "auto_map": {"AutoConfig": "nomic-ai/nomic-bert-2048--configuration_hf_nomic_bert.NomicBertConfig", "AutoModel": "nomic-ai/nomic-bert-2048--modeling_hf_nomic_bert.NomicBertModel", "AutoModelForMaskedLM": "nomic-ai/nomic-bert-2048--modeling_hf_nomic_bert.NomicBertForPreTraining", "A... | {
"auto_model": "AutoModel",
"custom_class": "nomic-ai/nomic-bert-2048--modeling_hf_nomic_bert.NomicBertModel",
"pipeline_tag": "feature-extraction",
"processor": null
} | {"language": ["en"], "library_name": "sentence-transformers", "license": "apache-2.0", "pipeline_tag": "sentence-similarity", "tags": ["feature-extraction", "sentence-similarity", "mteb", "transformers", "transformers.js"], "model-index": [{"name": "epoch_0_model", "results": [{"task": {"type": "Classification"}, "data... | null | null | [
"apache-2.0"
] | null | [
"en"
] | 136,731,648 | null | null | [
"AutoModel",
"NomicBertModel",
"nomic_bert",
"nomic-ai/nomic-bert-2048--modeling_hf_nomic_bert.NomicBertModel"
] | [
"sentence-similarity",
"feature-extraction"
] | [
"text",
"multimodal"
] | [
"text"
] | [
"logits",
"embeddings"
] |
65ee931c7443c0926729b6b5 | sophosympatheia/Midnight-Miqu-70B-v1.5 | sophosympatheia | {
"models": [
{
"_id": "65e3b8a6607c2f5380421ffb",
"id": "migtissera/Tess-70B-v1.6"
},
{
"_id": "65e012a1ebc538ecfac4cd1c",
"id": "sophosympatheia/Midnight-Miqu-70B-v1.0"
}
],
"relation": "merge"
} | 11,205 | 142,689 | False | 2024-03-11T05:14:04Z | 2024-12-10T16:50:16Z | transformers | 252 | 4 | [{"name": "Midnight-Miqu-70B-v1.5", "results": [{"task": {"type": "text-generation", "name": "Text Generation"}, "dataset": {"name": "IFEval (0-Shot)", "type": "HuggingFaceH4/ifeval", "args": {"num_few_shot": 0}}, "metrics": [{"type": "inst_level_strict_acc and prompt_level_strict_acc", "value": 61.18, "name": "strict ... | text-generation | {"parameters": {"F16": 68976648192}, "total": 68976648192} | [
".gitattributes",
"LICENSE",
"README.md",
"config.json",
"mergekit_config.yml",
"model-00001-of-00015.safetensors",
"model-00002-of-00015.safetensors",
"model-00003-of-00015.safetensors",
"model-00004-of-00015.safetensors",
"model-00005-of-00015.safetensors",
"model-00006-of-00015.safetensors",
... | 74888979a3734f0d7466d77dbf32120f575000e3 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"mergekit",
"merge",
"conversational",
"arxiv:2311.03099",
"base_model:migtissera/Tess-70B-v1.6",
"base_model:merge:migtissera/Tess-70B-v1.6",
"base_model:sophosympatheia/Midnight-Miqu-70B-v1.0",
"base_model:merge:sophosympatheia/Midni... | null | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}... | {
"auto_model": "AutoModelForCausalLM",
"custom_class": null,
"pipeline_tag": "text-generation",
"processor": "AutoTokenizer"
} | {"base_model": ["sophosympatheia/Midnight-Miqu-70B-v1.0", "migtissera/Tess-70B-v1.6"], "library_name": "transformers", "license": "other", "tags": ["mergekit", "merge"], "model-index": [{"name": "Midnight-Miqu-70B-v1.5", "results": [{"task": {"type": "text-generation", "name": "Text Generation"}, "dataset": {"name": "I... | <div style="width: auto; margin-left: auto; margin-right: auto">
<img src="https://i.imgur.com/Tn9MBg6.png" alt="MidnightMiqu" style="width: 100%; min-width: 400px; display: block; margin: auto;">
</div>
### Overview
Looking for the 103B version? You can get it from [FluffyKaeloky/Midnight-Miqu-103B-v1.5](https://hug... | null | [
"other"
] | null | null | 68,976,648,192 | null | null | [
"AutoModelForCausalLM",
"llama",
"LlamaForCausalLM"
] | [
"text-generation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
66969ad27a033bf62173f3e2 | meta-llama/Llama-3.1-70B-Instruct | meta-llama | {
"models": [
{
"_id": "66944fd095c7fa6e68c314ae",
"id": "meta-llama/Llama-3.1-70B"
}
],
"relation": "finetune"
} | 1,036,359 | 20,229,486 | manual | 2024-07-16T16:07:46Z | 2024-12-15T01:55:33Z | transformers | 902 | 4 | null | text-generation | {"parameters": {"BF16": 70553706496}, "total": 70553706496} | [
".gitattributes",
"LICENSE",
"README.md",
"USE_POLICY.md",
"config.json",
"generation_config.json",
"model-00001-of-00030.safetensors",
"model-00002-of-00030.safetensors",
"model-00003-of-00030.safetensors",
"model-00004-of-00030.safetensors",
"model-00005-of-00030.safetensors",
"model-00006-o... | 1605565b47bb9346c5515c34102e054115b4f98b | [
"transformers",
"safetensors",
"llama",
"text-generation",
"facebook",
"meta",
"pytorch",
"llama-3",
"conversational",
"en",
"de",
"fr",
"it",
"pt",
"hi",
"es",
"th",
"arxiv:2204.05149",
"base_model:meta-llama/Llama-3.1-70B",
"base_model:finetune:meta-llama/Llama-3.1-70B",
"l... | null | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{{- bos_token }}\n{%- if custom_tools is defined %}\n {%- set tools = custom_tools %}\n{%- endif %}\n{%- if not tools_in_user_message is defined %}\n {%- set tools_in_user_messag... | {
"auto_model": "AutoModelForCausalLM",
"custom_class": null,
"pipeline_tag": "text-generation",
"processor": "AutoTokenizer"
} | {"base_model": "meta-llama/Meta-Llama-3.1-70B", "datasets": null, "eval_results": null, "language": ["en", "de", "fr", "it", "pt", "hi", "es", "th"], "library_name": "transformers", "license": "llama3.1", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "text-generation",... | null | null | [
"llama3.1"
] | null | [
"en",
"de",
"fr",
"it",
"pt",
"hi",
"es",
"th"
] | 70,553,706,496 | null | null | [
"AutoModelForCausalLM",
"llama",
"LlamaForCausalLM"
] | [
"text-generation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
66b67e73621b62a4c02c1541 | Orenguteng/Llama-3.1-8B-Lexi-Uncensored-V2 | Orenguteng | null | 3,873 | 369,961 | False | 2024-08-09T20:39:15Z | 2024-09-25T02:49:53Z | transformers | 281 | 4 | [{"name": "Llama-3.1-8B-Lexi-Uncensored-V2", "results": [{"task": {"type": "text-generation", "name": "Text Generation"}, "dataset": {"name": "IFEval (0-Shot)", "type": "HuggingFaceH4/ifeval", "args": {"num_few_shot": 0}}, "metrics": [{"type": "inst_level_strict_acc and prompt_level_strict_acc", "value": 77.92, "name":... | text-generation | {"parameters": {"BF16": 8030261248}, "total": 8030261248} | [
".gitattributes",
"README.md",
"config.json",
"generation_config.json",
"model-00001-of-00004.safetensors",
"model-00002-of-00004.safetensors",
"model-00003-of-00004.safetensors",
"model-00004-of-00004.safetensors",
"model.safetensors.index.json",
"special_tokens_map.json",
"tokenizer.json",
"... | f4617caeabd21f1820ac89bd125c80eda70901a7 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"conversational",
"license:llama3.1",
"model-index",
"text-generation-inference",
"endpoints_compatible",
"deploy:azure",
"region:us"
] | null | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{{- bos_token }}\n{%- if custom_tools is defined %}\n {%- set tools = custom_tools %}\n{%- endif %}\n{%- if not tools_in_user_message is defined %}\n {%- set tools_in_user_messag... | {
"auto_model": "AutoModelForCausalLM",
"custom_class": null,
"pipeline_tag": "text-generation",
"processor": "AutoTokenizer"
} | {"library_name": "transformers", "license": "llama3.1", "model-index": [{"name": "Llama-3.1-8B-Lexi-Uncensored-V2", "results": [{"task": {"type": "text-generation", "name": "Text Generation"}, "dataset": {"name": "IFEval (0-Shot)", "type": "HuggingFaceH4/ifeval", "args": {"num_few_shot": 0}}, "metrics": [{"type": "inst... | 
VERSION 2 Update Notes:
---
- More compliant
- Smarter
- For best response, use this system prompt (feel free to expand upon it as you wish):
Think step by step with a logical reasoning and intellect... | null | [
"llama3.1"
] | null | null | 8,030,261,248 | null | null | [
"AutoModelForCausalLM",
"llama",
"LlamaForCausalLM"
] | [
"text-generation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
67d7262eaf109e28c7ddcc49 | mlabonne/gemma-3-27b-it-abliterated | mlabonne | {
"models": [
{
"_id": "67c35b9bb236f0d365bf29d3",
"id": "google/gemma-3-27b-it"
}
],
"relation": "finetune"
} | 4,371 | 275,755 | False | 2025-03-16T19:27:42Z | 2025-03-21T16:10:45Z | transformers | 305 | 4 | null | image-text-to-text | null | [
".gitattributes",
"README.md",
"added_tokens.json",
"chat_template.json",
"config.json",
"generation_config.json",
"model-00001-of-00012.safetensors",
"model-00002-of-00012.safetensors",
"model-00003-of-00012.safetensors",
"model-00004-of-00012.safetensors",
"model-00005-of-00012.safetensors",
... | eaa815dffdf0ea808a0421e4a48891bf52496a16 | [
"transformers",
"safetensors",
"gemma3",
"image-text-to-text",
"conversational",
"base_model:google/gemma-3-27b-it",
"base_model:finetune:google/gemma-3-27b-it",
"license:gemma",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | null | {"architectures": ["Gemma3ForConditionalGeneration"], "model_type": "gemma3", "processor_config": {"chat_template": "{{ bos_token }}\n{%- if messages[0]['role'] == 'system' -%}\n {%- if messages[0]['content'] is string -%}\n {%- set first_user_prefix = messages[0]['content'] + '\n\n' -%}\n {%- else -%}\n ... | {
"auto_model": "AutoModelForImageTextToText",
"custom_class": null,
"pipeline_tag": "image-text-to-text",
"processor": "AutoProcessor"
} | {"base_model": "google/gemma-3-27b-it", "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "gemma", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "image-text-to-text", "tags": null} | # 💎 Gemma 3 27B IT Abliterated

<center><a href="https://huggingface.co/mlabonne/gemma-3-1b-it-abliterated">Gemma 3 1B Abliterated</a> • <a href="https://huggingface.co/mlabonne/gemma-3-4b-it-ablitera... | null | [
"gemma"
] | null | null | null | null | null | [
"AutoModelForImageTextToText",
"Gemma3ForConditionalGeneration",
"gemma3"
] | [
"image-text-to-text"
] | [
"multimodal"
] | [
"text",
"image"
] | [
"text"
] |
67ff7b22b5ba13b43f0c3811 | ByteDance-Seed/UI-TARS-1.5-7B | ByteDance-Seed | null | 150,091 | 1,602,459 | False | 2025-04-16T09:40:50Z | 2025-04-18T01:35:38Z | transformers | 533 | 4 | null | image-text-to-text | {"parameters": {"F32": 8292166656}, "total": 8292166656} | [
".gitattributes",
"README.md",
"added_tokens.json",
"chat_template.json",
"config.json",
"merges.txt",
"model-00001-of-00007.safetensors",
"model-00002-of-00007.safetensors",
"model-00003-of-00007.safetensors",
"model-00004-of-00007.safetensors",
"model-00005-of-00007.safetensors",
"model-0000... | 683d002dd99d8f95104d31e70391a39348857f4e | [
"transformers",
"safetensors",
"qwen2_5_vl",
"image-text-to-text",
"multimodal",
"gui",
"conversational",
"en",
"arxiv:2501.12326",
"arxiv:2404.07972",
"arxiv:2409.08264",
"arxiv:2401.13919",
"arxiv:2504.01382",
"arxiv:2405.14573",
"arxiv:2410.23218",
"arxiv:2504.07981",
"license:apa... | null | {"architectures": ["Qwen2_5_VLForConditionalGeneration"], "model_type": "qwen2_5_vl", "processor_config": {"chat_template": "{% set image_count = namespace(value=0) %}{% set video_count = namespace(value=0) %}{% for message in messages %}{% if loop.first and message['role'] != 'system' %}<|im_start|>system\nYou are a h... | {
"auto_model": "AutoModelForImageTextToText",
"custom_class": null,
"pipeline_tag": "image-text-to-text",
"processor": "AutoProcessor"
} | {"base_model": null, "datasets": null, "eval_results": null, "language": ["en"], "library_name": "transformers", "license": "apache-2.0", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "image-text-to-text", "tags": ["multimodal", "gui"]} | null | null | [
"apache-2.0"
] | null | [
"en"
] | 8,292,166,656 | null | null | [
"qwen2_5_vl",
"AutoModelForImageTextToText",
"Qwen2_5_VLForConditionalGeneration"
] | [
"image-text-to-text"
] | [
"multimodal"
] | [
"text",
"image"
] | [
"text"
] |
681163a79c2765c932331a55 | XiaomiMiMo/MiMo-7B-Base | XiaomiMiMo | null | 116,314 | 554,955 | False | 2025-04-29T23:41:27Z | 2025-06-05T16:06:42Z | transformers | 129 | 4 | null | text-generation | {"parameters": {"BF16": 7833409536}, "total": 7833409536} | [
".gitattributes",
"README.md",
"config.json",
"configuration_mimo.py",
"generation_config.json",
"merges.txt",
"model-00001-of-00004.safetensors",
"model-00002-of-00004.safetensors",
"model-00003-of-00004.safetensors",
"model-00004-of-00004.safetensors",
"model.safetensors.index.json",
"modeli... | c72df4586cb8bdeebd65f36929cd3385a6566fbe | [
"transformers",
"safetensors",
"mimo",
"text-generation",
"conversational",
"custom_code",
"arxiv:2505.07608",
"license:mit",
"region:us"
] | null | {"architectures": ["MiMoForCausalLM"], "auto_map": {"AutoConfig": "configuration_mimo.MiMoConfig", "AutoModel": "modeling_mimo.MiMoModel", "AutoModelForCausalLM": "modeling_mimo.MiMoForCausalLM"}, "model_type": "mimo", "tokenizer_config": {"bos_token": null, "chat_template": "{%- if tools %}\n {{- '<|im_start|>syste... | {
"auto_model": "AutoModelForCausalLM",
"custom_class": "modeling_mimo.MiMoForCausalLM",
"pipeline_tag": "text-generation",
"processor": null
} | {"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "mit", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": null, "tags": null} | <div align="center">
<picture>
<source srcset="https://github.com/XiaomiMiMo/MiMo/raw/main/figures/Xiaomi_MiMo_darkmode.png?raw=true" media="(prefers-color-scheme: dark)">
<img src="https://github.com/XiaomiMiMo/MiMo/raw/main/figures/Xiaomi_MiMo.png?raw=true" width="60%" alt="Xiaomi-MiMo" />
</picture>
</di... | null | [
"mit"
] | null | null | 7,833,409,536 | null | null | [
"AutoModelForCausalLM",
"modeling_mimo.MiMoForCausalLM",
"mimo",
"MiMoForCausalLM"
] | [
"text-generation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
681c309e234962a917f42f50 | PocketDoc/Dans-PersonalityEngine-V1.3.0-24b | PocketDoc | {
"models": [
{
"_id": "67d75f3fd56afd880495e9b6",
"id": "mistralai/Mistral-Small-3.1-24B-Base-2503"
}
],
"relation": "finetune"
} | 181 | 91,747 | False | 2025-05-08T04:18:38Z | 2025-05-23T07:07:01Z | transformers | 139 | 4 | null | text-generation | null | [
".gitattributes",
"README.md",
"config.json",
"generation_config.json",
"model-00001-of-00010.safetensors",
"model-00002-of-00010.safetensors",
"model-00003-of-00010.safetensors",
"model-00004-of-00010.safetensors",
"model-00005-of-00010.safetensors",
"model-00006-of-00010.safetensors",
"model-0... | bb460a79d3d24f546106e8f738e9427f88d9dc40 | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"general-purpose",
"roleplay",
"storywriting",
"chemistry",
"biology",
"code",
"climate",
"axolotl",
"text-generation-inference",
"finetune",
"legal",
"medical",
"finance",
"conversational",
"en",
"ar",
"de",
"fr"... | null | {"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "[gMASK]<sop>", "chat_template": "{{ bos_token }}{%- set loop_messages = messages %}{%- for message in loop_messages %}{%- if message['role'] == 'user' %}{%- set role_token = '<|user|>' %}{%- elif message['role'] == 'as... | {
"auto_model": "AutoModelForCausalLM",
"custom_class": null,
"pipeline_tag": "text-generation",
"processor": "AutoTokenizer"
} | {"base_model": ["mistralai/Mistral-Small-3.1-24B-Base-2503"], "datasets": ["PocketDoc/Dans-Prosemaxx-RP", "PocketDoc/Dans-Personamaxx-Logs-2", "PocketDoc/Dans-Personamaxx-VN", "PocketDoc/Dans-Kinomaxx-VanillaBackrooms", "PocketDoc/Dans-Prosemaxx-Gutenberg", "PocketDoc/Dans-Prosemaxx-Cowriter-3-XL", "PocketDoc/Dans-Pros... | null | null | [
"apache-2.0"
] | [
"PocketDoc/Dans-Prosemaxx-RP",
"PocketDoc/Dans-Personamaxx-Logs-2",
"PocketDoc/Dans-Personamaxx-VN",
"PocketDoc/Dans-Kinomaxx-VanillaBackrooms",
"PocketDoc/Dans-Prosemaxx-Gutenberg",
"PocketDoc/Dans-Prosemaxx-Cowriter-3-XL",
"PocketDoc/Dans-Prosemaxx-Adventure",
"PocketDoc/Dans-Failuremaxx-Adventure-3... | [
"en",
"ar",
"de",
"fr",
"es",
"hi",
"pt",
"ja",
"ko"
] | null | null | null | [
"AutoModelForCausalLM",
"mistral",
"MistralForCausalLM"
] | [
"text-generation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
6848029d9a16b13edd1a3de4 | nanonets/Nanonets-OCR-s | nanonets | {
"models": [
{
"_id": "6795ff8f7534713f94918fc5",
"id": "Qwen/Qwen2.5-VL-3B-Instruct"
}
],
"relation": "finetune"
} | 64,163 | 1,355,238 | False | 2025-06-10T10:02:05Z | 2025-06-20T19:15:33Z | transformers | 1,589 | 4 | null | image-text-to-text | {"parameters": {"BF16": 3754622976}, "total": 3754622976} | [
".gitattributes",
"Modelfile",
"README.md",
"added_tokens.json",
"chat_template.jinja",
"config.json",
"generation_config.json",
"merges.txt",
"model-00001-of-00002.safetensors",
"model-00002-of-00002.safetensors",
"model.safetensors.index.json",
"preprocessor_config.json",
"special_tokens_m... | 3baad182cc87c65a1861f0c30357d3467e978172 | [
"transformers",
"safetensors",
"qwen2_5_vl",
"image-text-to-text",
"OCR",
"pdf2markdown",
"conversational",
"en",
"base_model:Qwen/Qwen2.5-VL-3B-Instruct",
"base_model:finetune:Qwen/Qwen2.5-VL-3B-Instruct",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | null | {"architectures": ["Qwen2_5_VLForConditionalGeneration"], "model_type": "qwen2_5_vl", "tokenizer_config": {"bos_token": null, "eos_token": "<|im_end|>", "pad_token": "<|endoftext|>", "unk_token": null}, "chat_template_jinja": "{% set image_count = namespace(value=0) %}{% set video_count = namespace(value=0) %}{% for me... | {
"auto_model": "AutoModelForImageTextToText",
"custom_class": null,
"pipeline_tag": "image-text-to-text",
"processor": "AutoProcessor"
} | {"base_model": ["Qwen/Qwen2.5-VL-3B-Instruct"], "datasets": null, "eval_results": null, "language": ["en"], "library_name": "transformers", "license": null, "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "image-text-to-text", "tags": ["OCR", "pdf2markdown"]} | Nanonets-OCR-s by [Nanonets](https://nanonets.com) is a powerful, state-of-the-art image-to-markdown OCR model that goes far beyond traditional text extraction. It transforms documents into structured markdown with intelligent content recognition and semantic tagging, making it ideal for downstream processing by Large ... | null | null | null | [
"en"
] | 3,754,622,976 | null | null | [
"qwen2_5_vl",
"AutoModelForImageTextToText",
"Qwen2_5_VLForConditionalGeneration"
] | [
"image-text-to-text"
] | [
"multimodal"
] | [
"text",
"image"
] | [
"text"
] |
687c61c324649ecb26a748f0 | zai-org/GLM-4.5-Air | zai-org | null | 406,257 | 2,723,104 | False | 2025-07-20T03:25:55Z | 2025-08-11T13:25:37Z | transformers | 592 | 4 | null | text-generation | {"parameters": {"BF16": 110468818944, "F32": 5888}, "total": 110468824832} | [
".gitattributes",
"README.md",
"chat_template.jinja",
"config.json",
"generation_config.json",
"model-00001-of-00047.safetensors",
"model-00002-of-00047.safetensors",
"model-00003-of-00047.safetensors",
"model-00004-of-00047.safetensors",
"model-00005-of-00047.safetensors",
"model-00006-of-00047... | a24ceef6ce4f3536971efe9b778bdaa1bab18daa | [
"transformers",
"safetensors",
"glm4_moe",
"text-generation",
"conversational",
"en",
"zh",
"arxiv:2508.06471",
"license:mit",
"endpoints_compatible",
"deploy:azure",
"region:us"
] | null | {"architectures": ["Glm4MoeForCausalLM"], "model_type": "glm4_moe", "tokenizer_config": {"eos_token": "<|endoftext|>", "pad_token": "<|endoftext|>"}, "chat_template_jinja": "[gMASK]<sop>\n{%- if tools -%}\n<|system|>\n# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with ... | {
"auto_model": "AutoModelForCausalLM",
"custom_class": null,
"pipeline_tag": "text-generation",
"processor": "AutoTokenizer"
} | {"base_model": null, "datasets": null, "eval_results": null, "language": ["en", "zh"], "library_name": "transformers", "license": "mit", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "text-generation", "tags": null} | # GLM-4.5-Air
<div align="center">
<img src=https://raw.githubusercontent.com/zai-org/GLM-4.5/refs/heads/main/resources/logo.svg width="15%"/>
</div>
<p align="center">
👋 Join our <a href="https://discord.gg/QR7SARHRxK" target="_blank">Discord</a> community.
<br>
📖 Check out the GLM-4.5 <a href="https://... | null | [
"mit"
] | null | [
"en",
"zh"
] | 110,468,824,832 | null | null | [
"AutoModelForCausalLM",
"Glm4MoeForCausalLM",
"glm4_moe"
] | [
"text-generation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
6889ec2029577a514ed82d21 | rednote-hilab/dots.ocr | rednote-hilab | null | 257,194 | 4,511,440 | False | 2025-07-30T09:55:44Z | 2025-10-31T08:49:31Z | dots_ocr | 1,281 | 4 | null | image-text-to-text | {"parameters": {"BF16": 3039179264}, "total": 3039179264} | [
".gitattributes",
"NOTICE",
"README.md",
"chat_template.json",
"config.json",
"configuration_dots.py",
"dots.ocr LICENSE AGREEMENT",
"generation_config.json",
"merges.txt",
"model-00001-of-00002.safetensors",
"model-00002-of-00002.safetensors",
"model.safetensors.index.json",
"modeling_dots_... | c0111ce6bc07803dbc267932ffef0ae3a51dc951 | [
"dots_ocr",
"safetensors",
"text-generation",
"image-to-text",
"ocr",
"document-parse",
"layout",
"table",
"formula",
"transformers",
"custom_code",
"image-text-to-text",
"conversational",
"en",
"zh",
"multilingual",
"license:mit",
"eval-results",
"region:us"
] | null | {"architectures": ["DotsOCRForCausalLM"], "model_type": "dots_ocr", "auto_map": {"AutoConfig": "configuration_dots.DotsOCRConfig", "AutoModelForCausalLM": "modeling_dots_ocr.DotsOCRForCausalLM"}, "processor_config": {"chat_template": "{% set image_count = namespace(value=0) %}{% set video_count = namespace(value=0) %}{... | {
"auto_model": "AutoModelForCausalLM",
"custom_class": "modeling_dots_ocr.DotsOCRForCausalLM",
"pipeline_tag": "text-generation",
"processor": null
} | {"base_model": null, "datasets": null, "eval_results": null, "language": ["en", "zh", "multilingual"], "library_name": "dots_ocr", "license": "mit", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "image-text-to-text", "tags": ["image-to-text", "ocr", "document-parse", "... | <div align="center">
<p align="center">
<img src="https://raw.githubusercontent.com/rednote-hilab/dots.ocr/master/assets/logo.png" width="300"/>
<p>
<h1 align="center">
dots.ocr: Multilingual Document Layout Parsing in a Single Vision-Language Model
</h1>
[
## Model Description
Hermes 4 70B is a frontier, hybrid-mode **reasoning** model based on Llama-3.1-70B by Nous Research that is aligned to **you**.
Read the Hermes 4 t... | null | [
"llama3"
] | null | [
"en"
] | null | null | null | [
"AutoModelForCausalLM",
"llama",
"LlamaForCausalLM"
] | [
"text-generation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
68c04b9ba4d1823e9b0ee2bd | Qwen/Qwen3-Next-80B-A3B-Thinking | Qwen | null | 50,718 | 2,282,134 | False | 2025-09-09T15:45:31Z | 2025-09-15T09:12:48Z | transformers | 486 | 4 | null | text-generation | null | [
".gitattributes",
"LICENSE",
"README.md",
"config.json",
"generation_config.json",
"merges.txt",
"model-00001-of-00041.safetensors",
"model-00002-of-00041.safetensors",
"model-00003-of-00041.safetensors",
"model-00004-of-00041.safetensors",
"model-00005-of-00041.safetensors",
"model-00006-of-0... | e502dd4100cc68c0de57643fd4317ec93a128670 | [
"transformers",
"safetensors",
"qwen3_next",
"text-generation",
"conversational",
"arxiv:2309.00071",
"arxiv:2505.09388",
"arxiv:2501.15383",
"license:apache-2.0",
"eval-results",
"endpoints_compatible",
"deploy:azure",
"region:us"
] | null | {"architectures": ["Qwen3NextForCausalLM"], "model_type": "qwen3_next", "tokenizer_config": {"bos_token": null, "chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0].role == 'system' %}\n {{- messages[0].content + '\\n\\n' }}\n {%- endif %}\n {{- \"# Tools\\n\\nYou m... | {
"auto_model": "AutoModelForCausalLM",
"custom_class": null,
"pipeline_tag": "text-generation",
"processor": "AutoTokenizer"
} | {"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "apache-2.0", "license_name": null, "license_link": "https://huggingface.co/Qwen/Qwen3-Next-80B-A3B-Thinking/blob/main/LICENSE", "metrics": null, "model_name": null, "pipeline_tag": "text-generation... | # Qwen3-Next-80B-A3B-Thinking
<a href="https://chat.qwen.ai/" target="_blank" style="margin: 2px;">
<img alt="Chat" src="https://img.shields.io/badge/%F0%9F%92%9C%EF%B8%8F%20Qwen%20Chat%20-536af5" style="display: inline-block; vertical-align: middle;"/>
</a>
Over the past few months, we have observed increasingly ... | null | [
"apache-2.0",
"https://huggingface.co/Qwen/Qwen3-Next-80B-A3B-Thinking/blob/main/LICENSE"
] | null | null | null | null | null | [
"qwen3_next",
"Qwen3NextForCausalLM",
"AutoModelForCausalLM"
] | [
"text-generation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
68d4e10c40e8730ee32cd0c6 | tencent/HunyuanImage-3.0 | tencent | null | 115,866 | 1,064,280 | False | 2025-09-25T06:28:28Z | 2026-01-28T08:14:42Z | transformers | 656 | 4 | null | text-to-image | null | [
".gitattributes",
"Hunyuan-Image3.md",
"LICENSE",
"README.md",
"README_zh_CN.md",
"__init__.py",
"assets/.DS_Store",
"assets/WECHAT.md",
"assets/banner.png",
"assets/banner_all.jpg",
"assets/demo_instruct_imgs/input_0_0.png",
"assets/demo_instruct_imgs/input_1_0.png",
"assets/demo_instruct_i... | 36f21fe74b65614451cc50ffd8a35a5f662dac70 | [
"transformers",
"safetensors",
"hunyuan_image_3_moe",
"text-generation",
"text-to-image",
"custom_code",
"arxiv:2509.23951",
"license:other",
"region:us"
] | null | {"architectures": ["HunyuanImage3ForCausalMM"], "auto_map": {"AutoConfig": "configuration_hunyuan.HunyuanImage3Config", "AutoModel": "hunyuan.HunyuanImage3Model", "AutoModelForCausalLM": "hunyuan.HunyuanImage3ForCausalMM"}, "model_type": "hunyuan_image_3_moe", "tokenizer_config": {"bos_token": "<|startoftext|>", "eos_t... | {
"auto_model": "AutoModelForCausalLM",
"custom_class": "hunyuan.HunyuanImage3ForCausalMM",
"pipeline_tag": "text-generation",
"processor": null
} | {"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "other", "license_name": "tencent-hunyuan-community", "license_link": "LICENSE", "metrics": null, "model_name": null, "pipeline_tag": "text-to-image", "tags": null} | null | null | [
"other",
"tencent-hunyuan-community",
"LICENSE"
] | null | null | null | null | null | [
"AutoModelForCausalLM",
"HunyuanImage3ForCausalMM",
"hunyuan.HunyuanImage3ForCausalMM",
"hunyuan_image_3_moe"
] | [
"text-to-image",
"text-generation"
] | [
"text",
"vision"
] | [
"text"
] | [
"text",
"image"
] |
68f17551daf20cc0cd8d9c40 | nvidia/llama-nemotron-embed-1b-v2 | nvidia | null | 58,549 | 152,621 | False | 2025-10-16T22:44:33Z | 2026-03-11T15:56:39Z | sentence-transformers | 49 | 4 | null | feature-extraction | {"parameters": {"BF16": 1235814400}, "total": 1235814400} | [
".gitattributes",
"1_Pooling/config.json",
"LICENSE",
"README.md",
"config.json",
"config_sentence_transformers.json",
"llama_bidirectional_model.py",
"model.safetensors",
"modules.json",
"pooling.py",
"pytorch_model.bin",
"sentence_bert_config.json",
"special_tokens_map.json",
"tokenizer.... | 5ee12b8f82486746155a9db12cacea9bc4b9b8ad | [
"sentence-transformers",
"pytorch",
"safetensors",
"llama_bidirec",
"feature-extraction",
"text",
"text-embeddings",
"retrieval",
"semantic-search",
"transformers",
"custom_code",
"multilingual",
"arxiv:2407.15831",
"license:other",
"text-embeddings-inference",
"endpoints_compatible",
... | null | {"architectures": ["LlamaBidirectionalModel"], "auto_map": {"AutoConfig": "llama_bidirectional_model.LlamaBidirectionalConfig", "AutoModel": "llama_bidirectional_model.LlamaBidirectionalModel"}, "model_type": "llama_bidirec", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "eos_token": "<|end_of_text|>", "pad_to... | {
"auto_model": "AutoModel",
"custom_class": "llama_bidirectional_model.LlamaBidirectionalModel",
"pipeline_tag": "feature-extraction",
"processor": null
} | {"base_model": null, "datasets": null, "eval_results": null, "language": ["multilingual"], "library_name": "sentence-transformers", "license": "other", "license_name": "nvidia-open-model-license", "license_link": "https://www.nvidia.com/en-us/agreements/enterprise-software/nvidia-open-model-license/", "metrics": null, ... | null | null | [
"other",
"nvidia-open-model-license",
"https://www.nvidia.com/en-us/agreements/enterprise-software/nvidia-open-model-license/"
] | null | [
"multilingual"
] | 1,235,814,400 | null | null | [
"AutoModel",
"llama_bidirec",
"llama_bidirectional_model.LlamaBidirectionalModel",
"LlamaBidirectionalModel"
] | [
"feature-extraction"
] | [
"multimodal"
] | [
"text"
] | [
"embeddings"
] |
68f4e3f4f82416217114139c | Qwen/Qwen3-VL-2B-Instruct | Qwen | null | 2,528,954 | 18,035,950 | False | 2025-10-19T13:13:24Z | 2025-10-23T11:30:44Z | transformers | 355 | 4 | null | image-text-to-text | {"parameters": {"BF16": 2127532032}, "total": 2127532032} | [
".gitattributes",
"README.md",
"chat_template.json",
"config.json",
"generation_config.json",
"merges.txt",
"model.safetensors",
"preprocessor_config.json",
"tokenizer.json",
"tokenizer_config.json",
"video_preprocessor_config.json",
"vocab.json"
] | 89644892e4d85e24eaac8bacfd4f463576704203 | [
"transformers",
"safetensors",
"qwen3_vl",
"image-text-to-text",
"conversational",
"arxiv:2505.09388",
"arxiv:2502.13923",
"arxiv:2409.12191",
"arxiv:2308.12966",
"license:apache-2.0",
"endpoints_compatible",
"deploy:azure",
"region:us"
] | null | {"architectures": ["Qwen3VLForConditionalGeneration"], "model_type": "qwen3_vl", "processor_config": {"chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0].role == 'system' %}\n {%- if messages[0].content is string %}\n {{- messages[0].content }}\n {%- el... | {
"auto_model": "AutoModelForImageTextToText",
"custom_class": null,
"pipeline_tag": "image-text-to-text",
"processor": "AutoProcessor"
} | {"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "apache-2.0", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "image-text-to-text", "tags": null} | null | null | [
"apache-2.0"
] | null | null | 2,127,532,032 | null | null | [
"AutoModelForImageTextToText",
"qwen3_vl",
"Qwen3VLForConditionalGeneration"
] | [
"image-text-to-text"
] | [
"multimodal"
] | [
"text",
"image"
] | [
"text"
] |
6901435c6b9d023f25308d41 | ByteDance/Ouro-2.6B-Thinking | ByteDance | null | 7,921 | 22,747 | False | 2025-10-28T22:27:40Z | 2026-02-26T18:39:17Z | transformers | 100 | 4 | null | text-generation | null | [
".gitattributes",
"README.md",
"assets/logo.png",
"assets/ouro_thinking.png",
"config.json",
"configuration_ouro.py",
"merges.txt",
"model.safetensors",
"modeling_ouro.py",
"special_tokens_map.json",
"tokenizer.json",
"tokenizer_config.json",
"vocab.json"
] | f1edd81e7ac41355db670500ceaf204e0f73af68 | [
"transformers",
"safetensors",
"ouro",
"text-generation",
"looped-language-model",
"reasoning",
"recurrent-depth",
"thinking",
"chain-of-thought",
"conversational",
"custom_code",
"arxiv:2510.25741",
"license:apache-2.0",
"region:us"
] | null | {"architectures": ["OuroForCausalLM"], "auto_map": {"AutoConfig": "configuration_ouro.OuroConfig", "AutoModel": "modeling_ouro.OuroModel", "AutoModelForCausalLM": "modeling_ouro.OuroForCausalLM"}, "model_type": "ouro", "tokenizer_config": {"bos_token": "<|im_start|>", "chat_template": "{%- if messages[0]['role'] == 'sy... | {
"auto_model": "AutoModelForCausalLM",
"custom_class": "modeling_ouro.OuroForCausalLM",
"pipeline_tag": "text-generation",
"processor": null
} | {"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "apache-2.0", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "text-generation", "tags": ["looped-language-model", "reasoning", "recurrent-depth", "t... | null | null | [
"apache-2.0"
] | null | null | null | null | null | [
"ouro",
"AutoModelForCausalLM",
"OuroForCausalLM",
"modeling_ouro.OuroForCausalLM"
] | [
"text-generation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
621ffdc036468d709f174332 | google-bert/bert-base-chinese | google-bert | null | 3,273,190 | 139,639,994 | False | 2022-03-02T23:29:04Z | 2025-07-03T11:58:48Z | transformers | 1,403 | 3 | null | fill-mask | null | [
".gitattributes",
"README.md",
"config.json",
"flax_model.msgpack",
"model.safetensors",
"pytorch_model.bin",
"tf_model.h5",
"tokenizer.json",
"tokenizer_config.json",
"vocab.txt"
] | 8f23c25b06e129b6c986331a13d8d025a92cf0ea | [
"transformers",
"pytorch",
"tf",
"jax",
"safetensors",
"bert",
"fill-mask",
"zh",
"arxiv:1810.04805",
"license:apache-2.0",
"endpoints_compatible",
"deploy:azure",
"region:us"
] | null | {"architectures": ["BertForMaskedLM"], "model_type": "bert", "tokenizer_config": {}} | {
"auto_model": "AutoModelForMaskedLM",
"custom_class": null,
"pipeline_tag": "fill-mask",
"processor": "AutoTokenizer"
} | {"base_model": null, "datasets": null, "eval_results": null, "language": "zh", "library_name": null, "license": "apache-2.0", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": null, "tags": null} | null | null | [
"apache-2.0"
] | null | [
"zh"
] | null | null | null | [
"AutoModelForMaskedLM",
"bert",
"BertForMaskedLM"
] | [
"fill-mask"
] | [
"text"
] | [
"text"
] | [
"logits"
] |
621ffdc036468d709f174358 | google-t5/t5-small | google-t5 | null | 1,813,634 | 174,084,827 | False | 2022-03-02T23:29:04Z | 2023-06-30T02:31:26Z | transformers | 536 | 3 | null | translation | {"parameters": {"F32": 60506880}, "total": 60506880} | [
".gitattributes",
"README.md",
"config.json",
"flax_model.msgpack",
"generation_config.json",
"model.safetensors",
"onnx/decoder_model.onnx",
"onnx/decoder_model_merged.onnx",
"onnx/decoder_model_merged_quantized.onnx",
"onnx/decoder_model_quantized.onnx",
"onnx/decoder_with_past_model.onnx",
... | df1b051c49625cf57a3d0d8d3863ed4d13564fe4 | [
"transformers",
"pytorch",
"tf",
"jax",
"rust",
"onnx",
"safetensors",
"t5",
"text2text-generation",
"summarization",
"translation",
"en",
"fr",
"ro",
"de",
"multilingual",
"dataset:c4",
"arxiv:1805.12471",
"arxiv:1708.00055",
"arxiv:1704.05426",
"arxiv:1606.05250",
"arxiv:... | null | {"architectures": ["T5ForConditionalGeneration"], "model_type": "t5", "tokenizer_config": {"eos_token": "</s>", "pad_token": "<pad>", "unk_token": "<unk>"}} | {
"auto_model": "AutoModelForSeq2SeqLM",
"custom_class": null,
"pipeline_tag": "text2text-generation",
"processor": "AutoTokenizer"
} | {"base_model": null, "datasets": ["c4"], "eval_results": null, "language": ["en", "fr", "ro", "de", "multilingual"], "library_name": null, "license": "apache-2.0", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": null, "tags": ["summarization", "translation"]} | # Model Card for T5 Small

# Table of Contents
1. [Model Details](#model-details)
2. [Uses](#uses)
3. ... | null | [
"apache-2.0"
] | [
"c4"
] | [
"en",
"fr",
"ro",
"de",
"multilingual"
] | 60,506,880 | null | null | [
"t5",
"T5ForConditionalGeneration",
"AutoModelForSeq2SeqLM"
] | [
"text2text-generation",
"translation",
"summarization"
] | [
"text"
] | [
"text"
] | [
"text"
] |
621ffdc136468d709f17a149 | cointegrated/rubert-tiny2 | cointegrated | null | 1,135,308 | 17,129,341 | False | 2022-03-02T23:29:05Z | 2025-12-30T08:33:24Z | sentence-transformers | 166 | 3 | null | sentence-similarity | {"parameters": {"I64": 2048, "F32": 29376502}, "total": 29378550} | [
".gitattributes",
".gitignore",
"1_Pooling/config.json",
"README.md",
"config.json",
"model.safetensors",
"modules.json",
"pytorch_model.bin",
"sentence_bert_config.json",
"special_tokens_map.json",
"tinybert-ru-labse-adapter-v2.pt",
"tokenizer.json",
"tokenizer_config.json",
"vocab.txt"
] | e8ed3b0c8bbf4fb6984c3de043bf7d2f4e5969ae | [
"sentence-transformers",
"pytorch",
"safetensors",
"bert",
"pretraining",
"russian",
"fill-mask",
"embeddings",
"masked-lm",
"tiny",
"feature-extraction",
"sentence-similarity",
"transformers",
"ru",
"license:mit",
"text-embeddings-inference",
"endpoints_compatible",
"region:us"
] | null | {"architectures": ["BertForPreTraining"], "model_type": "bert", "tokenizer_config": {"unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]"}} | {
"auto_model": "AutoModelForPreTraining",
"custom_class": null,
"pipeline_tag": "pretraining",
"processor": "AutoTokenizer"
} | {"base_model": null, "datasets": null, "eval_results": null, "language": ["ru"], "library_name": null, "license": "mit", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "sentence-similarity", "tags": ["russian", "fill-mask", "pretraining", "embeddings", "masked-lm", "tin... | This is an updated version of [cointegrated/rubert-tiny](https://huggingface.co/cointegrated/rubert-tiny): a small Russian BERT-based encoder with high-quality sentence embeddings. This [post in Russian](https://habr.com/ru/post/669674/) gives more details.
The differences from the previous version include:
- a larger... | null | [
"mit"
] | null | [
"ru"
] | 29,378,550 | null | null | [
"bert",
"AutoModelForPreTraining",
"BertForPreTraining"
] | [
"fill-mask",
"sentence-similarity",
"feature-extraction",
"pretraining"
] | [
"text",
"multimodal"
] | [
"text"
] | [
"logits",
"embeddings"
] |
621ffdc136468d709f17adb6 | facebook/bart-large-cnn | facebook | null | 2,055,150 | 138,785,394 | False | 2022-03-02T23:29:05Z | 2024-02-13T18:02:05Z | transformers | 1,552 | 3 | [{"name": "facebook/bart-large-cnn", "results": [{"task": {"type": "summarization", "name": "Summarization"}, "dataset": {"name": "cnn_dailymail", "type": "cnn_dailymail", "config": "3.0.0", "split": "train"}, "metrics": [{"name": "ROUGE-1", "type": "rouge", "value": 42.9486, "verified": false}, {"name": "ROUGE-2", "ty... | summarization | {"parameters": {"F32": 406290432}, "total": 406290432} | [
".gitattributes",
"README.md",
"config.json",
"flax_model.msgpack",
"generation_config.json",
"generation_config_for_summarization.json",
"merges.txt",
"model.safetensors",
"pytorch_model.bin",
"rust_model.ot",
"tf_model.h5",
"tokenizer.json",
"vocab.json"
] | 37f520fa929c961707657b28798b30c003dd100b | [
"transformers",
"pytorch",
"tf",
"jax",
"rust",
"safetensors",
"bart",
"text2text-generation",
"summarization",
"en",
"dataset:cnn_dailymail",
"arxiv:1910.13461",
"license:mit",
"model-index",
"endpoints_compatible",
"deploy:azure",
"region:us"
] | null | {"architectures": ["BartForConditionalGeneration"], "model_type": "bart"} | {
"auto_model": "AutoModelForSeq2SeqLM",
"custom_class": null,
"pipeline_tag": "text2text-generation",
"processor": "AutoTokenizer"
} | {"datasets": ["cnn_dailymail"], "language": ["en"], "license": "mit", "pipeline_tag": "summarization", "thumbnail": "https://huggingface.co/front/thumbnails/facebook.png", "model-index": [{"name": "facebook/bart-large-cnn", "results": [{"task": {"type": "summarization", "name": "Summarization"}, "dataset": {"name": "cn... | # BART (large-sized model), fine-tuned on CNN Daily Mail
BART model pre-trained on English language, and fine-tuned on [CNN Daily Mail](https://huggingface.co/datasets/cnn_dailymail). It was introduced in the paper [BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Co... | null | [
"mit"
] | [
"cnn_dailymail"
] | [
"en"
] | 406,290,432 | null | null | [
"bart",
"BartForConditionalGeneration",
"AutoModelForSeq2SeqLM"
] | [
"text2text-generation",
"summarization"
] | [
"text"
] | [
"text"
] | [
"text"
] |
62c8018851eb740327229910 | facebook/nllb-200-3.3B | facebook | null | 58,098 | 2,270,668 | False | 2022-07-08T10:06:00Z | 2023-02-11T20:19:13Z | transformers | 434 | 3 | null | translation | null | [
".gitattributes",
"README.md",
"config.json",
"generation_config.json",
"pytorch_model-00001-of-00003.bin",
"pytorch_model-00002-of-00003.bin",
"pytorch_model-00003-of-00003.bin",
"pytorch_model.bin.index.json",
"sentencepiece.bpe.model",
"special_tokens_map.json",
"tokenizer.json",
"tokenizer... | 1a07f7d195896b2114afcb79b7b57ab512e7b43e | [
"transformers",
"pytorch",
"m2m_100",
"text2text-generation",
"nllb",
"translation",
"ace",
"acm",
"acq",
"aeb",
"af",
"ajp",
"ak",
"als",
"am",
"apc",
"ar",
"ars",
"ary",
"arz",
"as",
"ast",
"awa",
"ayr",
"azb",
"azj",
"ba",
"bm",
"ban",
"be",
"bem",
"b... | null | {"architectures": ["M2M100ForConditionalGeneration"], "model_type": "m2m_100", "tokenizer_config": {"bos_token": "<s>", "cls_token": "<s>", "eos_token": "</s>", "mask_token": {"__type": "AddedToken", "content": "<mask>", "lstrip": true, "normalized": true, "rstrip": false, "single_word": false}, "pad_token": "<pad>", "... | {
"auto_model": "AutoModelForSeq2SeqLM",
"custom_class": null,
"pipeline_tag": "text2text-generation",
"processor": "AutoTokenizer"
} | {"base_model": null, "datasets": ["flores-200"], "eval_results": null, "language": ["ace", "acm", "acq", "aeb", "af", "ajp", "ak", "als", "am", "apc", "ar", "ars", "ary", "arz", "as", "ast", "awa", "ayr", "azb", "azj", "ba", "bm", "ban", "be", "bem", "bn", "bho", "bjn", "bo", "bs", "bug", "bg", "ca", "ceb", "cs", "cjk"... | # NLLB-200
This is the model card of NLLB-200's 3.3B variant.
Here are the [metrics](https://tinyurl.com/nllb200dense3bmetrics) for that particular checkpoint.
- Information about training algorithms, parameters, fairness constraints or other applied approaches, and features. The exact training algorithm, data and t... | null | [
"cc-by-nc-4.0"
] | [
"flores-200"
] | [
"ace",
"acm",
"acq",
"aeb",
"af",
"ajp",
"ak",
"als",
"am",
"apc",
"ar",
"ars",
"ary",
"arz",
"as",
"ast",
"awa",
"ayr",
"azb",
"azj",
"ba",
"bm",
"ban",
"be",
"bem",
"bn",
"bho",
"bjn",
"bo",
"bs",
"bug",
"bg",
"ca",
"ceb",
"cs",
"cjk",
"ckb",... | null | null | [
"bleu",
"spbleu",
"chrf++"
] | [
"AutoModelForSeq2SeqLM",
"m2m_100",
"M2M100ForConditionalGeneration"
] | [
"text2text-generation",
"translation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
63314bb6acb6472115aa55a9 | openai/whisper-tiny | openai | null | 776,748 | 14,064,428 | False | 2022-09-26T06:50:30Z | 2024-02-29T10:57:33Z | transformers | 420 | 3 | [{"name": "whisper-tiny", "results": [{"task": {"name": "Automatic Speech Recognition", "type": "automatic-speech-recognition"}, "dataset": {"name": "LibriSpeech (clean)", "type": "librispeech_asr", "config": "clean", "split": "test", "args": {"language": "en"}}, "metrics": [{"name": "Test WER", "type": "wer", "value":... | automatic-speech-recognition | null | [
".gitattributes",
"README.md",
"added_tokens.json",
"config.json",
"flax_model.msgpack",
"generation_config.json",
"merges.txt",
"model.safetensors",
"normalizer.json",
"preprocessor_config.json",
"pytorch_model.bin",
"special_tokens_map.json",
"tf_model.h5",
"tokenizer.json",
"tokenizer... | 169d4a4341b33bc18d8881c4b69c2e104e1cc0af | [
"transformers",
"pytorch",
"tf",
"jax",
"safetensors",
"whisper",
"automatic-speech-recognition",
"audio",
"hf-asr-leaderboard",
"en",
"zh",
"de",
"es",
"ru",
"ko",
"fr",
"ja",
"pt",
"tr",
"pl",
"ca",
"nl",
"ar",
"sv",
"it",
"id",
"hi",
"fi",
"vi",
"he",
"... | null | {"architectures": ["WhisperForConditionalGeneration"], "model_type": "whisper", "tokenizer_config": {"bos_token": "<|endoftext|>", "eos_token": "<|endoftext|>", "pad_token": "<|endoftext|>", "unk_token": "<|endoftext|>"}} | {
"auto_model": "AutoModelForSpeechSeq2Seq",
"custom_class": null,
"pipeline_tag": "automatic-speech-recognition",
"processor": "AutoProcessor"
} | {"language": ["en", "zh", "de", "es", "ru", "ko", "fr", "ja", "pt", "tr", "pl", "ca", "nl", "ar", "sv", "it", "id", "hi", "fi", "vi", "he", "uk", "el", "ms", "cs", "ro", "da", "hu", "ta", "no", "th", "ur", "hr", "bg", "lt", "la", "mi", "ml", "cy", "sk", "te", "fa", "lv", "bn", "sr", "az", "sl", "kn", "et", "mk", "br", ... | # Whisper
Whisper is a pre-trained model for automatic speech recognition (ASR) and speech translation. Trained on 680k hours
of labelled data, Whisper models demonstrate a strong ability to generalise to many datasets and domains **without** the need
for fine-tuning.
Whisper was proposed in the paper [Robust Speec... | null | [
"apache-2.0"
] | null | [
"en",
"zh",
"de",
"es",
"ru",
"ko",
"fr",
"ja",
"pt",
"tr",
"pl",
"ca",
"nl",
"ar",
"sv",
"it",
"id",
"hi",
"fi",
"vi",
"he",
"uk",
"el",
"ms",
"cs",
"ro",
"da",
"hu",
"ta",
"no",
"th",
"ur",
"hr",
"bg",
"lt",
"la",
"mi",
"ml",
"cy",
"sk"... | null | null | null | [
"WhisperForConditionalGeneration",
"whisper",
"AutoModelForSpeechSeq2Seq"
] | [
"automatic-speech-recognition"
] | [
"multimodal"
] | [
"audio"
] | [
"text"
] |
639861ac02d4b462c93d29f5 | Salesforce/blip-image-captioning-large | Salesforce | null | 1,572,616 | 45,731,127 | False | 2022-12-13T11:27:40Z | 2025-02-03T06:42:42Z | transformers | 1,460 | 3 | null | image-to-text | {"parameters": {"I64": 512, "F32": 469732924}, "total": 469733436} | [
".gitattributes",
"README.md",
"config.json",
"model.safetensors",
"preprocessor_config.json",
"pytorch_model.bin",
"special_tokens_map.json",
"tf_model.h5",
"tokenizer.json",
"tokenizer_config.json",
"vocab.txt"
] | 353689b859fcf0523410b1806dace5fb46ecdf41 | [
"transformers",
"pytorch",
"tf",
"safetensors",
"blip",
"image-text-to-text",
"image-captioning",
"image-to-text",
"arxiv:2201.12086",
"license:bsd-3-clause",
"endpoints_compatible",
"region:us"
] | null | {"architectures": ["BlipForConditionalGeneration"], "model_type": "blip", "tokenizer_config": {"cls_token": "[CLS]", "mask_token": "[MASK]", "pad_token": "[PAD]", "sep_token": "[SEP]", "unk_token": "[UNK]"}} | {
"auto_model": "AutoModelForImageTextToText",
"custom_class": null,
"pipeline_tag": "image-text-to-text",
"processor": "AutoProcessor"
} | {"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": null, "license": "bsd-3-clause", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "image-to-text", "tags": ["image-captioning"], "languages": ["en"]} | # BLIP: Bootstrapping Language-Image Pre-training for Unified Vision-Language Understanding and Generation
Model card for image captioning pretrained on COCO dataset - base architecture (with ViT large backbone).
| "}, "metrics": [{"name": "pass@1", "type": "pass@1", "value": 0.408, "verified": false}]}, {"task": {"type": "text-generation"}, "dataset": {"type": "openai_humaneval", "name": "... | text-generation | {"parameters": {"F32": 15819446272}, "total": 15819446272} | [
".gitattributes",
"README.md",
"config.json",
"generation_config.json",
"merges.txt",
"model-00001-of-00007.safetensors",
"model-00002-of-00007.safetensors",
"model-00003-of-00007.safetensors",
"model-00004-of-00007.safetensors",
"model-00005-of-00007.safetensors",
"model-00006-of-00007.safetens... | 827cd7721231872c152e3faef5cae1b22964ed58 | [
"transformers",
"pytorch",
"safetensors",
"gpt_bigcode",
"text-generation",
"code",
"dataset:bigcode/the-stack-dedup",
"arxiv:1911.02150",
"arxiv:2205.14135",
"arxiv:2207.14255",
"arxiv:2305.06161",
"license:bigcode-openrail-m",
"model-index",
"text-generation-inference",
"endpoints_comp... | null | {"architectures": ["GPTBigCodeForCausalLM"], "model_type": "gpt_bigcode", "tokenizer_config": {"bos_token": "<|endoftext|>", "eos_token": "<|endoftext|>", "unk_token": "<|endoftext|>"}} | {
"auto_model": "AutoModelForCausalLM",
"custom_class": null,
"pipeline_tag": "text-generation",
"processor": "AutoTokenizer"
} | {"datasets": ["bigcode/the-stack-dedup"], "library_name": "transformers", "license": "bigcode-openrail-m", "metrics": ["code_eval"], "pipeline_tag": "text-generation", "tags": ["code"], "inference": true, "widget": [{"text": "def print_hello_world():", "example_title": "Hello world", "group": "Python"}], "extra_gated_p... | null | null | [
"bigcode-openrail-m"
] | [
"bigcode/the-stack-dedup"
] | null | 15,819,446,272 | null | [
"code_eval"
] | [
"gpt_bigcode",
"AutoModelForCausalLM",
"GPTBigCodeForCausalLM"
] | [
"text-generation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
650aedb6238a644cb93a52c3 | mistralai/Mistral-7B-v0.1 | mistralai | null | 454,618 | 23,222,227 | False | 2023-09-20T13:03:50Z | 2025-07-24T16:44:02Z | transformers | 4,057 | 3 | null | text-generation | {"parameters": {"BF16": 7241732096}, "total": 7241732096} | [
".gitattributes",
"README.md",
"config.json",
"generation_config.json",
"model-00001-of-00002.safetensors",
"model-00002-of-00002.safetensors",
"model.safetensors.index.json",
"pytorch_model-00001-of-00002.bin",
"pytorch_model-00002-of-00002.bin",
"pytorch_model.bin.index.json",
"special_tokens_... | 27d67f1b5f57dc0953326b2601d68371d40ea8da | [
"transformers",
"pytorch",
"safetensors",
"mistral",
"text-generation",
"pretrained",
"mistral-common",
"en",
"arxiv:2310.06825",
"license:apache-2.0",
"text-generation-inference",
"region:us"
] | null | {"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "eos_token": "</s>", "pad_token": null, "unk_token": "<unk>", "use_default_system_prompt": false}} | {
"auto_model": "AutoModelForCausalLM",
"custom_class": null,
"pipeline_tag": "text-generation",
"processor": "AutoTokenizer"
} | {"base_model": null, "datasets": null, "eval_results": null, "language": ["en"], "library_name": "transformers", "license": "apache-2.0", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": null, "tags": ["pretrained", "mistral-common"], "inference": false, "extra_gated_desc... | null | null | [
"apache-2.0"
] | null | [
"en"
] | 7,241,732,096 | null | null | [
"AutoModelForCausalLM",
"mistral",
"MistralForCausalLM"
] | [
"text-generation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
65d4e83bc2e85b8b98af701f | bigcode/starcoder2-15b | bigcode | null | 10,053 | 517,419 | False | 2024-02-20T17:58:19Z | 2024-06-05T19:52:45Z | transformers | 664 | 3 | [{"name": "starcoder2-15b", "results": [{"task": {"type": "text-generation"}, "dataset": {"name": "CruxEval-I", "type": "cruxeval-i"}, "metrics": [{"type": "pass@1", "value": 48.1, "verified": false}]}, {"task": {"type": "text-generation"}, "dataset": {"name": "DS-1000", "type": "ds-1000"}, "metrics": [{"type": "pass@1... | text-generation | null | [
".gitattributes",
"README.md",
"config.json",
"generation_config.json",
"merges.txt",
"model-00001-of-00014.safetensors",
"model-00002-of-00014.safetensors",
"model-00003-of-00014.safetensors",
"model-00004-of-00014.safetensors",
"model-00005-of-00014.safetensors",
"model-00006-of-00014.safetens... | 46d44742909c03ac8cee08eb03fdebce02e193ec | [
"transformers",
"safetensors",
"starcoder2",
"text-generation",
"code",
"dataset:bigcode/the-stack-v2-train",
"arxiv:2305.13245",
"arxiv:2205.14135",
"arxiv:2004.05150",
"arxiv:2207.14255",
"arxiv:2402.19173",
"license:bigcode-openrail-m",
"model-index",
"text-generation-inference",
"end... | null | {"architectures": ["Starcoder2ForCausalLM"], "model_type": "starcoder2", "tokenizer_config": {"bos_token": "<|endoftext|>", "eos_token": "<|endoftext|>", "unk_token": "<|endoftext|>"}} | {
"auto_model": "AutoModelForCausalLM",
"custom_class": null,
"pipeline_tag": "text-generation",
"processor": "AutoTokenizer"
} | {"datasets": ["bigcode/the-stack-v2-train"], "library_name": "transformers", "license": "bigcode-openrail-m", "pipeline_tag": "text-generation", "tags": ["code"], "inference": {"parameters": {"temperature": 0.2, "top_p": 0.95}}, "widget": [{"text": "def print_hello_world():", "example_title": "Hello world", "group": "P... | # StarCoder2
<center>
<img src="https://huggingface.co/datasets/bigcode/admin_private/resolve/main/starcoder2_banner.png" alt="SC2" width="900" height="600">
</center>
## Table of Contents
1. [Model Summary](#model-summary)
2. [Use](#use)
3. [Limitations](#limitations)
4. [Training](#training)
5. [License](#lic... | null | [
"bigcode-openrail-m"
] | [
"bigcode/the-stack-v2-train"
] | null | null | null | null | [
"AutoModelForCausalLM",
"starcoder2",
"Starcoder2ForCausalLM"
] | [
"text-generation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
65fc66d12c813664be7731ae | google/codegemma-7b | google | null | 1,183 | 401,887 | manual | 2024-03-21T16:56:49Z | 2024-08-07T18:27:09Z | transformers | 215 | 3 | null | text-generation | null | [
".gitattributes",
"README.md",
"codegemma_nl_benchmarks.png",
"config.json",
"generation_config.json",
"model-00001-of-00004.safetensors",
"model-00002-of-00004.safetensors",
"model-00003-of-00004.safetensors",
"model-00004-of-00004.safetensors",
"model.safetensors.index.json",
"special_tokens_m... | 2ec970089fe8f083d10d1ce6bac30538ac4badaa | [
"transformers",
"safetensors",
"gemma",
"text-generation",
"license:gemma",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | null | {"architectures": ["GemmaForCausalLM"], "model_type": "gemma", "tokenizer_config": {"bos_token": "<bos>", "eos_token": "<eos>", "pad_token": "<pad>", "unk_token": "<unk>", "use_default_system_prompt": false}} | {
"auto_model": "AutoModelForCausalLM",
"custom_class": null,
"pipeline_tag": "text-generation",
"processor": "AutoTokenizer"
} | {"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "gemma", "license_name": null, "license_link": "https://ai.google.dev/gemma/terms", "metrics": null, "model_name": null, "pipeline_tag": null, "tags": null, "extra_gated_heading": "Access CodeGemma ... | null | null | [
"gemma",
"https://ai.google.dev/gemma/terms"
] | null | null | null | null | null | [
"AutoModelForCausalLM",
"gemma",
"GemmaForCausalLM"
] | [
"text-generation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
6623f2c68fee5ba324a5fecc | protectai/deberta-v3-base-prompt-injection-v2 | protectai | {
"models": [
{
"_id": "621ffdc136468d709f17dee4",
"id": "microsoft/deberta-v3-base"
}
],
"relation": "quantized"
} | 436,329 | 4,305,262 | False | 2024-04-20T16:52:22Z | 2024-05-28T07:07:49Z | transformers | 104 | 3 | [{"name": "deberta-v3-base-prompt-injection-v2", "results": []}] | text-classification | {"parameters": {"F32": 184423682}, "total": 184423682} | [
".gitattributes",
"LICENSE",
"README.md",
"added_tokens.json",
"config.json",
"deberta-v3-base-prompt-injection-v2_emissions.csv",
"model.safetensors",
"onnx/added_tokens.json",
"onnx/config.json",
"onnx/model.onnx",
"onnx/special_tokens_map.json",
"onnx/spm.model",
"onnx/tokenizer.json",
... | e6535ca4ce3ba852083e75ec585d7c8aeb4be4c5 | [
"transformers",
"onnx",
"safetensors",
"deberta-v2",
"text-classification",
"prompt-injection",
"injection",
"security",
"llm-security",
"generated_from_trainer",
"en",
"dataset:natolambert/xstest-v2-copy",
"dataset:VMware/open-instruct",
"dataset:alespalla/chatbot_instruction_prompts",
... | null | {"architectures": ["DebertaV2ForSequenceClassification"], "model_type": "deberta-v2", "tokenizer_config": {"bos_token": "[CLS]", "cls_token": "[CLS]", "eos_token": "[SEP]", "mask_token": "[MASK]", "pad_token": "[PAD]", "sep_token": "[SEP]", "unk_token": "[UNK]"}} | {
"auto_model": "AutoModelForSequenceClassification",
"custom_class": null,
"pipeline_tag": "text-classification",
"processor": "AutoTokenizer"
} | {"base_model": "microsoft/deberta-v3-base", "datasets": ["natolambert/xstest-v2-copy", "VMware/open-instruct", "alespalla/chatbot_instruction_prompts", "HuggingFaceH4/grok-conversation-harmless", "Harelix/Prompt-Injection-Mixed-Techniques-2024", "OpenSafetyLab/Salad-Data", "jackhhao/jailbreak-classification"], "eval_re... | # Model Card for deberta-v3-base-prompt-injection-v2
This model is a fine-tuned version of [microsoft/deberta-v3-base](https://huggingface.co/microsoft/deberta-v3-base) specifically developed to detect and classify prompt injection attacks which can manipulate language models into producing unintended outputs.
## Int... | null | [
"apache-2.0"
] | [
"natolambert/xstest-v2-copy",
"VMware/open-instruct",
"alespalla/chatbot_instruction_prompts",
"HuggingFaceH4/grok-conversation-harmless",
"Harelix/Prompt-Injection-Mixed-Techniques-2024",
"OpenSafetyLab/Salad-Data",
"jackhhao/jailbreak-classification"
] | [
"en"
] | 184,423,682 | null | [
"accuracy",
"recall",
"precision",
"f1"
] | [
"AutoModelForSequenceClassification",
"deberta-v2",
"DebertaV2ForSequenceClassification"
] | [
"text-classification"
] | [
"text"
] | [
"text"
] | [
"logits"
] |
66268dc9cd54f5fffb4e2e24 | microsoft/Phi-3-mini-4k-instruct | microsoft | null | 747,020 | 27,675,955 | False | 2024-04-22T16:18:17Z | 2025-12-10T20:22:18Z | transformers | 1,403 | 3 | null | text-generation | {"parameters": {"BF16": 3821079552}, "total": 3821079552} | [
".gitattributes",
"CODE_OF_CONDUCT.md",
"LICENSE",
"NOTICE.md",
"README.md",
"SECURITY.md",
"added_tokens.json",
"config.json",
"configuration_phi3.py",
"data_summary_card.md",
"generation_config.json",
"model-00001-of-00002.safetensors",
"model-00002-of-00002.safetensors",
"model.safetens... | f39ac1d28e925b323eae81227eaba4464caced4e | [
"transformers",
"safetensors",
"phi3",
"text-generation",
"nlp",
"code",
"conversational",
"custom_code",
"en",
"fr",
"license:mit",
"eval-results",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | null | {"architectures": ["Phi3ForCausalLM"], "auto_map": {"AutoConfig": "configuration_phi3.Phi3Config", "AutoModelForCausalLM": "modeling_phi3.Phi3ForCausalLM"}, "model_type": "phi3", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{% for message in messages %}{% if message['role'] == 'system' %}{{'<|system|>\n' ... | {
"auto_model": "AutoModelForCausalLM",
"custom_class": null,
"pipeline_tag": "text-generation",
"processor": "AutoTokenizer"
} | {"base_model": null, "datasets": null, "eval_results": null, "language": ["en", "fr"], "library_name": null, "license": "mit", "license_name": null, "license_link": "https://huggingface.co/microsoft/Phi-3-mini-4k-instruct/resolve/main/LICENSE", "metrics": null, "model_name": null, "pipeline_tag": "text-generation", "ta... | 🎉 **Phi-3.5**: [[mini-instruct]](https://huggingface.co/microsoft/Phi-3.5-mini-instruct); [[MoE-instruct]](https://huggingface.co/microsoft/Phi-3.5-MoE-instruct) ; [[vision-instruct]](https://huggingface.co/microsoft/Phi-3.5-vision-instruct)
## Model Summary
The Phi-3-Mini-4K-Instruct is a 3.8B parameters, lightweig... | null | [
"mit",
"https://huggingface.co/microsoft/Phi-3-mini-4k-instruct/resolve/main/LICENSE"
] | null | [
"en",
"fr"
] | 3,821,079,552 | null | null | [
"AutoModelForCausalLM",
"phi3",
"Phi3ForCausalLM"
] | [
"text-generation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
66286c9842709fe826603f58 | aaditya/Llama3-OpenBioLLM-70B | aaditya | {
"models": [
{
"_id": "661f97be9cae93c02393e8ef",
"id": "meta-llama/Meta-Llama-3-70B-Instruct"
}
],
"relation": "finetune"
} | 3,282 | 336,928 | False | 2024-04-24T02:21:12Z | 2025-01-18T07:39:15Z | transformers | 503 | 3 | [{"name": "OpenBioLLM-70B", "results": []}] | text-generation | null | [
".gitattributes",
"LICENSE",
"README.md",
"config.json",
"generation_config.json",
"pytorch_model-00001-of-00030.bin",
"pytorch_model-00002-of-00030.bin",
"pytorch_model-00003-of-00030.bin",
"pytorch_model-00004-of-00030.bin",
"pytorch_model-00005-of-00030.bin",
"pytorch_model-00006-of-00030.bin... | 7ad17ef0d2185811f731f89d20885b2f99b1e994 | [
"transformers",
"pytorch",
"llama",
"text-generation",
"llama-3",
"Mixtral",
"instruct",
"finetune",
"chatml",
"DPO",
"RLHF",
"gpt4",
"distillation",
"heathcare",
"medical",
"clinical",
"med",
"lifescience",
"Pharmaceutical",
"Pharma",
"conversational",
"en",
"arxiv:2305.... | null | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|e... | {
"auto_model": "AutoModelForCausalLM",
"custom_class": null,
"pipeline_tag": "text-generation",
"processor": "AutoTokenizer"
} | {"base_model": "meta-llama/Meta-Llama-3-70B-Instruct", "datasets": null, "eval_results": [], "language": ["en"], "library_name": null, "license": "llama3", "license_name": null, "license_link": null, "metrics": null, "model_name": "OpenBioLLM-70B", "pipeline_tag": null, "tags": ["llama-3", "llama", "Mixtral", "instruct... | <div align="center">
<img width="260px" src="https://cdn-uploads.huggingface.co/production/uploads/5f3fe13d79c1ba4c353d0c19/BrQCb95lmEIFz79QAmoNA.png"></div>

<div align="center">
<h1>Advancin... | null | [
"llama3"
] | null | [
"en"
] | null | null | null | [
"AutoModelForCausalLM",
"llama",
"LlamaForCausalLM"
] | [
"text-generation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
666ce6f4067382b3e98317c6 | microsoft/Florence-2-base | microsoft | null | 811,173 | 8,822,257 | False | 2024-06-15T00:57:24Z | 2025-08-04T17:36:57Z | transformers | 357 | 3 | null | image-text-to-text | {"parameters": {"F16": 231567705}, "total": 231567705} | [
".gitattributes",
"CODE_OF_CONDUCT.md",
"LICENSE",
"README.md",
"SECURITY.md",
"SUPPORT.md",
"config.json",
"configuration_florence2.py",
"model.safetensors",
"modeling_florence2.py",
"preprocessor_config.json",
"processing_florence2.py",
"pytorch_model.bin",
"tokenizer.json",
"tokenizer... | 5ca5edf5bd017b9919c05d08aebef5e4c7ac3bac | [
"transformers",
"pytorch",
"safetensors",
"florence2",
"image-text-to-text",
"vision",
"custom_code",
"arxiv:2311.06242",
"license:mit",
"endpoints_compatible",
"region:us"
] | null | {"architectures": ["Florence2ForConditionalGeneration"], "auto_map": {"AutoConfig": "configuration_florence2.Florence2Config", "AutoModelForCausalLM": "modeling_florence2.Florence2ForConditionalGeneration"}, "model_type": "florence2", "tokenizer_config": {}} | {
"auto_model": "AutoModelForImageTextToText",
"custom_class": null,
"pipeline_tag": "image-text-to-text",
"processor": "AutoProcessor"
} | {"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": null, "license": "mit", "license_name": null, "license_link": "https://huggingface.co/microsoft/Florence-2-base/resolve/main/LICENSE", "metrics": null, "model_name": null, "pipeline_tag": "image-text-to-text", "tags": ["visio... | null | null | [
"mit",
"https://huggingface.co/microsoft/Florence-2-base/resolve/main/LICENSE"
] | null | null | 231,567,705 | null | null | [
"florence2",
"AutoModelForImageTextToText",
"Florence2ForConditionalGeneration"
] | [
"image-text-to-text"
] | [
"multimodal"
] | [
"text",
"image"
] | [
"text"
] |
669b774854729d168aeb6207 | Alibaba-NLP/gte-multilingual-base | Alibaba-NLP | null | 960,722 | 22,385,282 | False | 2024-07-20T08:37:28Z | 2025-07-05T01:38:16Z | sentence-transformers | 350 | 3 | [{"name": "gte-multilingual-base (dense)", "results": [{"task": {"type": "Clustering"}, "dataset": {"type": "PL-MTEB/8tags-clustering", "name": "MTEB 8TagsClustering", "config": "default", "split": "test", "revision": "None"}, "metrics": [{"type": "v_measure", "value": 33.66681726329994, "verified": false}]}, {"task": ... | sentence-similarity | {"parameters": {"F16": 305369089}, "total": 305369089} | [
".gitattributes",
"1_Pooling/config.json",
"README.md",
"config.json",
"images/mgte-loco.png",
"images/mgte-mldr.png",
"images/mgte-mteb.png",
"images/mgte-reranker.png",
"images/mgte-retrieval.png",
"model.safetensors",
"modules.json",
"scripts/gte_embedding.py",
"sentence_bert_config.json"... | 9bbca17d9273fd0d03d5725c7a4b0f6b45142062 | [
"sentence-transformers",
"safetensors",
"new",
"feature-extraction",
"mteb",
"transformers",
"multilingual",
"sentence-similarity",
"text-embeddings-inference",
"custom_code",
"af",
"ar",
"az",
"be",
"bg",
"bn",
"ca",
"ceb",
"cs",
"cy",
"da",
"de",
"el",
"en",
"es",
... | null | {"architectures": ["NewModel", "NewForTokenClassification"], "auto_map": {"AutoConfig": "Alibaba-NLP/new-impl--configuration.NewConfig", "AutoModelForMaskedLM": "Alibaba-NLP/new-impl--modeling.NewForMaskedLM", "AutoModel": "Alibaba-NLP/new-impl--modeling.NewModel", "AutoModelForMultipleChoice": "Alibaba-NLP/new-impl--m... | {
"auto_model": "AutoModel",
"custom_class": "Alibaba-NLP/new-impl--modeling.NewModel",
"pipeline_tag": "feature-extraction",
"processor": null
} | {"language": ["af", "ar", "az", "be", "bg", "bn", "ca", "ceb", "cs", "cy", "da", "de", "el", "en", "es", "et", "eu", "fa", "fi", "fr", "gl", "gu", "he", "hi", "hr", "ht", "hu", "hy", "id", "is", "it", "ja", "jv", "ka", "kk", "km", "kn", "ko", "ky", "lo", "lt", "lv", "mk", "ml", "mn", "mr", "ms", "my", "ne", "nl", "no",... | null | null | [
"apache-2.0"
] | null | [
"af",
"ar",
"az",
"be",
"bg",
"bn",
"ca",
"ceb",
"cs",
"cy",
"da",
"de",
"el",
"en",
"es",
"et",
"eu",
"fa",
"fi",
"fr",
"gl",
"gu",
"he",
"hi",
"hr",
"ht",
"hu",
"hy",
"id",
"is",
"it",
"ja",
"jv",
"ka",
"kk",
"km",
"kn",
"ko",
"ky",
"lo... | 305,369,089 | null | null | [
"NewForTokenClassification",
"Alibaba-NLP/new-impl--modeling.NewModel",
"AutoModel",
"new",
"NewModel"
] | [
"sentence-similarity",
"feature-extraction"
] | [
"text",
"multimodal"
] | [
"text"
] | [
"logits",
"embeddings"
] |
66e6cfeb6998c3d86c14e163 | Qwen/Qwen2.5-0.5B | Qwen | null | 1,824,056 | 17,297,091 | False | 2024-09-15T12:15:39Z | 2024-09-25T12:32:36Z | transformers | 388 | 3 | null | text-generation | {"parameters": {"BF16": 494032768}, "total": 494032768} | [
".gitattributes",
"LICENSE",
"README.md",
"config.json",
"generation_config.json",
"merges.txt",
"model.safetensors",
"tokenizer.json",
"tokenizer_config.json",
"vocab.json"
] | 060db6499f32faf8b98477b0a26969ef7d8b9987 | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"conversational",
"en",
"arxiv:2407.10671",
"license:apache-2.0",
"text-generation-inference",
"endpoints_compatible",
"deploy:azure",
"region:us"
] | null | {"architectures": ["Qwen2ForCausalLM"], "model_type": "qwen2", "tokenizer_config": {"bos_token": null, "chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0]['role'] == 'system' %}\n {{- messages[0]['content'] }}\n {%- else %}\n {{- 'You are a helpful assistant.' ... | {
"auto_model": "AutoModelForCausalLM",
"custom_class": null,
"pipeline_tag": "text-generation",
"processor": "AutoTokenizer"
} | {"base_model": null, "datasets": null, "eval_results": null, "language": ["en"], "library_name": "transformers", "license": "apache-2.0", "license_name": null, "license_link": "https://huggingface.co/Qwen/Qwen2.5-0.5B/blob/main/LICENSE", "metrics": null, "model_name": null, "pipeline_tag": "text-generation", "tags": nu... | # Qwen2.5-0.5B
## Introduction
Qwen2.5 is the latest series of Qwen large language models. For Qwen2.5, we release a number of base language models and instruction-tuned language models ranging from 0.5 to 72 billion parameters. Qwen2.5 brings the following improvements upon Qwen2:
- Significantly **more knowledge**... | null | [
"apache-2.0",
"https://huggingface.co/Qwen/Qwen2.5-0.5B/blob/main/LICENSE"
] | null | [
"en"
] | 494,032,768 | null | null | [
"AutoModelForCausalLM",
"Qwen2ForCausalLM",
"qwen2"
] | [
"text-generation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
66e6d032663c452ad2400439 | Qwen/Qwen2.5-1.5B | Qwen | null | 704,321 | 8,715,182 | False | 2024-09-15T12:16:50Z | 2024-10-08T10:56:50Z | transformers | 171 | 3 | null | text-generation | {"parameters": {"BF16": 1543714304}, "total": 1543714304} | [
".gitattributes",
"LICENSE",
"README.md",
"config.json",
"generation_config.json",
"merges.txt",
"model.safetensors",
"tokenizer.json",
"tokenizer_config.json",
"vocab.json"
] | 8faed761d45a263340a0528343f099c05c9a4323 | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"conversational",
"en",
"arxiv:2407.10671",
"license:apache-2.0",
"text-generation-inference",
"endpoints_compatible",
"deploy:azure",
"region:us"
] | null | {"architectures": ["Qwen2ForCausalLM"], "model_type": "qwen2", "tokenizer_config": {"bos_token": null, "chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0]['role'] == 'system' %}\n {{- messages[0]['content'] }}\n {%- else %}\n {{- 'You are a helpful assistant.' ... | {
"auto_model": "AutoModelForCausalLM",
"custom_class": null,
"pipeline_tag": "text-generation",
"processor": "AutoTokenizer"
} | {"base_model": null, "datasets": null, "eval_results": null, "language": ["en"], "library_name": "transformers", "license": "apache-2.0", "license_name": null, "license_link": "https://huggingface.co/Qwen/Qwen2.5-1.5B/blob/main/LICENSE", "metrics": null, "model_name": null, "pipeline_tag": "text-generation", "tags": nu... | # Qwen2.5-1.5B
## Introduction
Qwen2.5 is the latest series of Qwen large language models. For Qwen2.5, we release a number of base language models and instruction-tuned language models ranging from 0.5 to 72 billion parameters. Qwen2.5 brings the following improvements upon Qwen2:
- Significantly **more knowledge**... | null | [
"apache-2.0",
"https://huggingface.co/Qwen/Qwen2.5-1.5B/blob/main/LICENSE"
] | null | [
"en"
] | 1,543,714,304 | null | null | [
"AutoModelForCausalLM",
"Qwen2ForCausalLM",
"qwen2"
] | [
"text-generation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
66e902f34eeaced24a2787f8 | Qwen/Qwen2.5-32B-Instruct | Qwen | {
"models": [
{
"_id": "66e6d09952865ca1e41c6566",
"id": "Qwen/Qwen2.5-32B"
}
],
"relation": "finetune"
} | 4,098,183 | 19,629,113 | False | 2024-09-17T04:17:55Z | 2024-09-25T12:33:09Z | transformers | 341 | 3 | null | text-generation | null | [
".gitattributes",
"LICENSE",
"README.md",
"config.json",
"generation_config.json",
"merges.txt",
"model-00001-of-00017.safetensors",
"model-00002-of-00017.safetensors",
"model-00003-of-00017.safetensors",
"model-00004-of-00017.safetensors",
"model-00005-of-00017.safetensors",
"model-00006-of-0... | 5ede1c97bbab6ce5cda5812749b4c0bdf79b18dd | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"chat",
"conversational",
"en",
"arxiv:2309.00071",
"arxiv:2407.10671",
"base_model:Qwen/Qwen2.5-32B",
"base_model:finetune:Qwen/Qwen2.5-32B",
"license:apache-2.0",
"text-generation-inference",
"endpoints_compatible",
"deploy:azu... | null | {"architectures": ["Qwen2ForCausalLM"], "model_type": "qwen2", "tokenizer_config": {"bos_token": null, "chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0]['role'] == 'system' %}\n {{- messages[0]['content'] }}\n {%- else %}\n {{- 'You are Qwen, created by Aliba... | {
"auto_model": "AutoModelForCausalLM",
"custom_class": null,
"pipeline_tag": "text-generation",
"processor": "AutoTokenizer"
} | {"base_model": "Qwen/Qwen2.5-32B", "datasets": null, "eval_results": null, "language": ["en"], "library_name": "transformers", "license": "apache-2.0", "license_name": null, "license_link": "https://huggingface.co/Qwen/Qwen2.5-32B-Instruct/blob/main/LICENSE", "metrics": null, "model_name": null, "pipeline_tag": "text-g... | # Qwen2.5-32B-Instruct
## Introduction
Qwen2.5 is the latest series of Qwen large language models. For Qwen2.5, we release a number of base language models and instruction-tuned language models ranging from 0.5 to 72 billion parameters. Qwen2.5 brings the following improvements upon Qwen2:
- Significantly **more kno... | null | [
"apache-2.0",
"https://huggingface.co/Qwen/Qwen2.5-32B-Instruct/blob/main/LICENSE"
] | null | [
"en"
] | null | null | null | [
"AutoModelForCausalLM",
"Qwen2ForCausalLM",
"qwen2"
] | [
"text-generation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.