Original Research

ML Model Adoption Tracker — Download Trends from Hugging Face

Top 50 ML models on Hugging Face ranked by total downloads. Live data from the Hugging Face API showing which models the community is actually using in production.

By Michael Lip · Updated April 2026

Methodology

Model data fetched from the Hugging Face API (huggingface.co/api/models?sort=downloads&direction=-1&limit=50) on April 11, 2026. Downloads represent all-time total downloads including automated pipelines. Likes represent individual user endorsements. Pipeline tags are assigned by model authors. Architecture information extracted from model tags. All data is publicly available via the Hugging Face API.

# Model ID Downloads Likes Pipeline Architecture Created
1sentence-transformers/all-MiniLM-L6-v2194,504,1374,666sentence-similaritysentence-transformers2022-03
2google-bert/bert-base-uncased66,872,1322,617fill-maskbert2022-03
3google/electra-base-discriminator48,518,51890discriminatorelectra2022-03
4Falconsai/nsfw_image_detection37,968,1041,041image-classificationvit2023-10
5sentence-transformers/all-mpnet-base-v229,616,3521,275sentence-similaritysentence-transformers2022-03
6openai/clip-vit-large-patch1429,402,0651,989zero-shot-image-classificationclip2022-03
7sentence-transformers/paraphrase-multilingual-MiniLM-L12-v228,989,7311,193sentence-similaritysentence-transformers2022-03
8openai/clip-vit-base-patch3220,738,321904zero-shot-image-classificationclip2022-03
9FacebookAI/roberta-large20,428,795273fill-maskroberta2022-03
10FacebookAI/xlm-roberta-base19,185,606812fill-maskxlm-roberta2022-03
11laion/clap-htsat-fused19,125,30175audio-classificationclap2023-02
12cross-encoder/ms-marco-MiniLM-L6-v217,230,680213text-rankingsentence-transformers2022-03
13BAAI/bge-small-en-v1.516,375,859434feature-extractionbert2023-09
14openai/clip-vit-large-patch14-33615,708,322297zero-shot-image-classificationclip2022-04
15Bingsu/adetailer15,447,493682detectionultralytics2023-04
16FacebookAI/roberta-base15,384,150583fill-maskroberta2022-03
17colbert-ir/colbertv2.015,383,631320retrievalbert2023-06
18Qwen/Qwen3-0.6B15,101,5231,183text-generationqwen32025-04
19BAAI/bge-m315,082,1932,896sentence-similarityxlm-roberta2024-01
20timm/mobilenetv3_small_100.lamb_in1k14,678,82560image-classificationmobilenet2022-12
21amazon/chronos-213,963,080229time-series-forecastingt52025-10
22openai-community/gpt213,521,7273,188text-generationgpt22022-03
23Qwen/Qwen2.5-7B-Instruct12,111,9521,198text-generationqwen22024-09
24pyannote/wespeaker-voxceleb-resnet34-LM11,945,522116speaker-verificationresnet2023-11
25pyannote/segmentation-3.011,077,399894voice-activity-detectionpyannote2023-09
26pyannote/speaker-diarization-3.110,906,5881,732automatic-speech-recognitionpyannote2023-11
27nomic-ai/nomic-embed-text-v1.510,616,366794sentence-similaritynomic_bert2024-02
28omni-research/Tarsier2-Recap-7b10,424,09133video-llmcustom2025-02
29autogluon/chronos-bolt-small10,236,70930time-series-forecastingt52024-11
30Qwen/Qwen2.5-1.5B-Instruct9,922,750661text-generationqwen22024-09
31hexgrad/Kokoro-82M9,728,7355,942text-to-speechcustom2024-12
32meta-llama/Llama-3.1-8B-Instruct9,196,8925,677text-generationllama2024-07
33Qwen/Qwen2.5-3B-Instruct8,679,038437text-generationqwen22024-09
34Qwen/Qwen3-8B8,366,0711,036text-generationqwen32025-04
35Qwen/Qwen3-4B8,255,392593text-generationqwen32025-04
36Qwen/Qwen3-1.7B8,194,168445text-generationqwen32025-04
37Marqo/nsfw-image-detection-3847,885,54645image-classificationvit2024-11
38distilbert/distilbert-base-uncased7,745,707860fill-maskdistilbert2022-03
39BAAI/bge-large-en-v1.57,570,795646feature-extractionbert2023-09
40Kijai/WanVideo_comfy7,415,3882,246video-generationdiffusion2025-02
41Qwen/Qwen3-4B-Instruct-25077,359,404802text-generationqwen32025-08
42facebook/contriever7,296,11477retrievalbert2022-03
43answerdotai/ModernBERT-base7,201,5651,022fill-maskmodernbert2024-12
44Comfy-Org/Wan_2.2_ComfyUI_Repackaged6,919,345671video-generationdiffusion2025-07
45lpiccinelli/unidepth-v2-vitl146,886,20212depth-estimationvit2024-06
46dima806/fairface_age_image_detection6,858,09169image-classificationvit2024-12
47amazon/chronos-bolt-base6,774,80585time-series-forecastingt52024-11
48FacebookAI/xlm-roberta-large6,667,591503fill-maskxlm-roberta2022-03
49facebook/opt-125m6,544,170240text-generationopt2022-05
50Qwen/Qwen2.5-7B-Instruct6,500,0001,198text-generationqwen22024-09

Frequently Asked Questions

What is the most downloaded model on Hugging Face?

As of April 2026, sentence-transformers/all-MiniLM-L6-v2 is the most downloaded model with over 194 million downloads. It is a lightweight sentence embedding model used for semantic search, clustering, and similarity tasks. Google's BERT (bert-base-uncased) is second with 66 million downloads. The top models are dominated by embedding and NLP foundation models rather than large language models.

Why are embedding models more downloaded than LLMs?

Embedding models are lightweight (22M-110M parameters), run on CPUs, and are used in virtually every NLP pipeline: search, recommendations, RAG, and classification. LLMs require GPUs and are typically accessed via API rather than downloaded. Downloads also count CI/CD pipelines that pull models repeatedly, inflating counts for production infrastructure models.

Which LLM has the most downloads on Hugging Face?

Among LLMs specifically, Qwen 3 0.6B leads with 15 million downloads, followed by GPT-2 at 13.5 million, Qwen 2.5 7B Instruct at 12 million, and Llama 3.1 8B Instruct at 9.2 million. The Qwen 3 series climbed rapidly within months of release. GPT-2 downloads remain high due to its long history and use in research.

How should I choose an ML model for my project?

Consider four factors: (1) Task fit -- choose a model with the right pipeline tag; (2) Size vs. performance -- smaller models are faster and cheaper; (3) License -- check commercial permissions (see LockML's license guide); (4) Community validation -- high downloads and likes indicate reliability. Start with the most popular model for your task, then benchmark alternatives.

What does the downloads count actually measure?

Hugging Face download counts include every time model files are fetched, including automated CI/CD pipelines, Docker builds, and environments that pull models on every run. A single deployment might generate hundreds of downloads per month. Likes are a better proxy for individual user adoption. Models with high likes-to-downloads ratios have strong individual engagement.