Original Research

PyTorch Model Zoo Stats

Architecture Popularity and Download Trends

By Michael Lip · Published April 11, 2026 · Data source: HuggingFace API · Last updated:
50
Models Ranked
194M
Top Downloads
5,942
Most Liked
12
Task Categories

Which PyTorch models are developers actually using? We queried the HuggingFace API for the top 50 most-downloaded PyTorch models and found that sentence embedding models dominate, with all-MiniLM-L6-v2 leading at 194M+ downloads. Text generation models from Qwen and Meta occupy 12 of the top 50 spots, reflecting the ongoing LLM boom.

This data helps developers choose battle-tested models and understand which architectures the community trusts for production workloads.

Top 50 PyTorch Models by Downloads

# Model ID Downloads Likes Task License
1sentence-transformers/all-MiniLM-L6-v2194,504,1374,666sentence-similarityApache-2.0
2google-bert/bert-base-uncased66,872,1322,617fill-maskApache-2.0
3google/electra-base-discriminator48,518,51890discriminativeApache-2.0
4Falconsai/nsfw_image_detection37,968,1041,041image-classificationApache-2.0
5sentence-transformers/all-mpnet-base-v229,616,3521,275sentence-similarityApache-2.0
6openai/clip-vit-large-patch1429,402,0651,989zero-shot-imageCustom
7sentence-transformers/paraphrase-multilingual-MiniLM-L12-v228,989,7311,193sentence-similarityApache-2.0
8openai/clip-vit-base-patch3220,738,321904zero-shot-imageCustom
9FacebookAI/roberta-large20,428,795273fill-maskMIT
10FacebookAI/xlm-roberta-base19,185,606812fill-maskMIT
11laion/clap-htsat-fused19,125,30175audio-classificationApache-2.0
12cross-encoder/ms-marco-MiniLM-L6-v217,230,680213text-rankingApache-2.0
13BAAI/bge-small-en-v1.516,375,859434feature-extractionMIT
14openai/clip-vit-large-patch14-33615,708,322297zero-shot-imageCustom
15Bingsu/adetailer15,447,493682detectionApache-2.0
16FacebookAI/roberta-base15,384,150583fill-maskMIT
17colbert-ir/colbertv2.015,383,631320retrievalMIT
18Qwen/Qwen3-0.6B15,101,5231,183text-generationApache-2.0
19BAAI/bge-m315,082,1932,896sentence-similarityMIT
20timm/mobilenetv3_small_100.lamb_in1k14,678,82560image-classificationApache-2.0
21amazon/chronos-213,963,080229time-seriesApache-2.0
22openai-community/gpt213,521,7273,188text-generationMIT
23Qwen/Qwen2.5-7B-Instruct12,111,9521,198text-generationApache-2.0
24pyannote/wespeaker-voxceleb-resnet34-LM11,945,522116speaker-embedCC-BY-4.0
25pyannote/segmentation-3.011,077,399894voice-activityMIT
26pyannote/speaker-diarization-3.110,906,5881,732speech-recognitionMIT
27nomic-ai/nomic-embed-text-v1.510,616,366794sentence-similarityApache-2.0
28omni-research/Tarsier2-Recap-7b10,424,09133multimodalApache-2.0
29autogluon/chronos-bolt-small10,236,70930time-seriesApache-2.0
30Qwen/Qwen2.5-1.5B-Instruct9,922,750661text-generationApache-2.0
31hexgrad/Kokoro-82M9,728,7355,942text-to-speechApache-2.0
32meta-llama/Llama-3.1-8B-Instruct9,196,8925,677text-generationLlama 3.1
33autogluon/chronos-29,162,04712time-seriesApache-2.0
34Qwen/Qwen2.5-3B-Instruct8,679,038437text-generationOther
35Qwen/Qwen3-8B8,366,0711,036text-generationApache-2.0
36Qwen/Qwen3-4B8,255,392593text-generationApache-2.0
37Qwen/Qwen3-1.7B8,194,168445text-generationApache-2.0
38Marqo/nsfw-image-detection-3847,885,54645image-classificationApache-2.0
39distilbert/distilbert-base-uncased7,745,707860fill-maskApache-2.0
40BAAI/bge-large-en-v1.57,570,795646feature-extractionMIT
41Kijai/WanVideo_comfy7,415,3882,246video-generationCustom
42Qwen/Qwen3-4B-Instruct-25077,359,404802text-generationApache-2.0
43facebook/contriever7,296,11477retrievalCustom
44answerdotai/ModernBERT-base7,201,5651,022fill-maskApache-2.0
45Comfy-Org/Wan_2.2_ComfyUI_Repackaged6,919,345671image-generationCustom
46lpiccinelli/unidepth-v2-vitl146,886,20212depth-estimationCustom
47dima806/fairface_age_image_detection6,858,09169image-classificationApache-2.0
48amazon/chronos-bolt-base6,774,80585time-seriesApache-2.0
49FacebookAI/xlm-roberta-large6,667,591503fill-maskMIT
50facebook/opt-125m6,544,170240text-generationOther

Methodology

Model data in this index is sourced directly from the HuggingFace API:

Download counts include programmatic API access and may not reflect unique users. Models used as dependencies by other packages will have inflated download numbers.

Frequently Asked Questions

What is the most downloaded PyTorch model on HuggingFace?

As of April 2026, the most downloaded PyTorch model is sentence-transformers/all-MiniLM-L6-v2 with over 194 million downloads. This sentence embedding model is popular for semantic search, clustering, and retrieval-augmented generation (RAG) pipelines. Google's BERT-base-uncased ranks second with 66M+ downloads.

Which PyTorch model architecture is most popular in 2026?

Sentence embedding models dominate the top downloads, with sentence-transformers appearing 4 times in the top 20. Text generation models (Qwen3, Llama 3.1, GPT-2) are the second most popular category. CLIP models for zero-shot image classification hold strong in the top 10. The trend shows a shift toward specialized embedding and multimodal models.

How do I choose between PyTorch models for my project?

Consider: 1) Task match — use the pipeline_tag to find models for your task, 2) Model size vs. hardware — MiniLM (22M params) runs on CPU while Llama-3.1-8B needs a GPU, 3) License — Apache 2.0 and MIT are commercially permissive, 4) Community validation — higher likes and downloads indicate battle-tested models, 5) Recency — newer architectures like ModernBERT often outperform older ones.

What license do most popular PyTorch models use?

Apache 2.0 is the most common license among the top 50, used by over 60% including Qwen, sentence-transformers, and BERT. MIT is second, used by RoBERTa, BGE, and pyannote models. Some models like CLIP and Llama use custom licenses with varying commercial restrictions.

Why do some models have high downloads but low likes?

High downloads with low likes typically indicates the model is used programmatically as a dependency rather than being directly chosen. For example, google/electra-base-discriminator has 48M downloads but only 90 likes because it is pulled automatically by downstream packages. Models with high likes relative to downloads (like Kokoro-82M with 5,942 likes) tend to have strong community engagement.

Related Tools

Parameter Counter Memory Calculator ML Framework Comparison Layer Performance

Free to use under CC BY 4.0 license. Cite this page when sharing.