Hugging Face · Pricing Plans

Hugging Face Plans Pricing

Hugging Face uses a hybrid pricing model. Account subscriptions (Free, PRO, Team, Enterprise) layer on top of usage-based pricing for Inference Providers (pay-as-you-go), Inference Endpoints (per-instance-hour), and Spaces hardware (per-hour). Storage is metered in TB/month with volume discounts. Subscriptions include monthly inference credits that pass through provider rates with no markup.

8 Plans API Commons Plans
View Source
AIMachine LearningInferenceModel HubDatasets

Plans

Free freemium

Free Hub access for collaboration and learning. Includes $0.10/month in Inference Providers credits.

Inference Providers monthly credits (USD · month) 0.00 USD
PRO subscription

Individual professional tier with 10x private storage, 2x public storage, 20x inference credits, 8x ZeroGPU quota, and Spaces Dev Mode.

PRO subscription (month · month) $9 USD
Inference Providers monthly credits (USD · month) included USD
Team subscription

Per-seat plan for organizations with SSO, audit logs, storage regions, resource groups, and analytics.

Team seat (seat · month) $20 USD
Inference Providers credits per seat (USD · month) included USD
Enterprise Hub enterprise

Enterprise per-seat plan with the highest limits, SCIM provisioning, dedicated support, and contractual terms.

Enterprise seat (seat · month) $50+ USD
Inference Providers credits per seat (USD · month) included USD
Inference Providers (Pay-As-You-Go) usage-based

Routed access to 200+ models across partner inference providers (Cerebras, Cohere, Fireworks, Groq, Together, Replicate, fal, Sambanova, etc). Pass-through provider rates with no Hugging Face markup.

Inference request (routed) (token · usage) see provider rate card USD
Inference Endpoints usage-based

Dedicated managed inference deployments billed per replica-hour by hardware tier.

CPU - AWS Sapphire Rapids (1 vCPU) (instance-hour · usage) $0.03 USD
CPU - AWS Sapphire Rapids (16 vCPU) (instance-hour · usage) $0.54 USD
GPU - NVIDIA T4 (1x, 14GB) (instance-hour · usage) $0.50 USD
GPU - NVIDIA L40S (1x, 48GB) (instance-hour · usage) $1.80 USD
GPU - NVIDIA A100 (1x, 80GB) (instance-hour · usage) $2.50 USD
GPU - NVIDIA H100 (1x, 80GB) (instance-hour · usage) $4.50 USD
GPU - NVIDIA H200 (1x, 141GB) (instance-hour · usage) $5.00 USD
GPU - NVIDIA B200 (1x, 179GB) (instance-hour · usage) $9.25 USD
GPU - NVIDIA A100 (8x, 640GB) (instance-hour · usage) $20.00 USD
Accelerator - AWS Inf2 Neuron (instance-hour · usage) $0.75 USD
Accelerator - GCP TPU v5e (1x1) (instance-hour · usage) $1.20 USD
Spaces Hardware usage-based

Hosted Gradio/Streamlit/Docker apps billed per hour by hardware. ZeroGPU is free with dynamic allocation on H200.

CPU Basic (instance-hour · usage) 0.00 USD
CPU Upgrade (8 vCPU, 32GB) (instance-hour · usage) $0.03 USD
NVIDIA T4 small (instance-hour · usage) $0.40 USD
NVIDIA L4 (1x) (instance-hour · usage) $0.80 USD
NVIDIA A10G small (instance-hour · usage) $1.00 USD
NVIDIA A100 large (1x, 80GB) (instance-hour · usage) $2.50 USD
8x NVIDIA A100 (instance-hour · usage) $20.00 USD
ZeroGPU (H200, dynamic) (instance-hour · usage) 0.00 USD
Hub Storage usage-based

Tiered TB/month storage pricing for repositories. Public repos cheaper than private; volume discounts at 50TB, 200TB, 500TB.

Public storage (0-50TB) (gb · month) $12/TB USD
Public storage (50TB+) (gb · month) $10/TB USD
Public storage (200TB+) (gb · month) $9/TB USD
Public storage (500TB+) (gb · month) $8/TB USD
Private storage (0-50TB) (gb · month) $18/TB USD
Private storage (50TB+) (gb · month) $16/TB USD
Private storage (200TB+) (gb · month) $14/TB USD
Private storage (500TB+) (gb · month) $12/TB USD

Sources