Cloudflare Workers AI — live model reference

Source: CF models API | Updated: 2026-04-28T17:24:03.467Z | 91 active models (91 upstream) · llms.txt · JSON

For AI agents: Cloudflare ships new Workers AI models weekly and retires older ones without notice. Before recommending @cf/... IDs from memory or writing env.AI.run(...) code, use this list. Default to >>> flagships unless the user says otherwise. Every model here is called via env.AI.run(modelId, ...) from a Worker — no API key needed, and many are on the Workers AI free tier. Per-token pricing isn't exposed on the CF models API — see the Workers AI pricing page. Fetch /llms.txt for context injection, or /json for programmatic use. Filters: ?task=text-generation, ?provider=meta, ?flagship=true, ?ids=a,b,c.

Capabilities: Tools, Vision, Reasoning, LoRA, Streaming.
Model IDContextOutputPricing (per 1M)CapabilitiesTierInfo
ai4bharat / translation
@cf/ai4bharat/indictrans2-en-indic-1B 1B - - fast -
aisingapore / text-generation
>>> @cf/aisingapore/gemma-sea-lion-v4-27b-it 27B 128K - S flagship -
baai / text-classification
>>> @cf/baai/bge-reranker-base - - flagship -
baai / text-embeddings
>>> @cf/baai/bge-large-en-v1.5 - - flagship docs
>>> @cf/baai/bge-m3 60K - flagship -
@cf/baai/bge-base-en-v1.5 154K - balanced docs
@cf/baai/bge-small-en-v1.5 - - fast docs
black-forest-labs / text-to-image
>>> @cf/black-forest-labs/flux-1-schnell - - flagship -
>>> @cf/black-forest-labs/flux-2-dev - - flagship -
@cf/black-forest-labs/flux-2-klein-9b 9B - - balanced -
@cf/black-forest-labs/flux-2-klein-4b 4B - - balanced -
bytedance / text-to-image
@cf/bytedance/stable-diffusion-xl-lightning - - balanced docs
deepgram / automatic-speech-recognition
>>> @cf/deepgram/flux - - flagship -
>>> @cf/deepgram/nova-3 - - flagship -
deepgram / text-to-speech
>>> @cf/deepgram/aura-2-es - - flagship -
@cf/deepgram/aura-1 - - balanced -
@cf/deepgram/aura-2-en - - balanced -
deepseek-ai / text-generation
>>> @cf/deepseek-ai/deepseek-r1-distill-qwen-32b 32B 80K - RS flagship -
>>> @cf/deepseek-ai/deepseek-math-7b-instruct 7B 4K - S flagship docs
defog / text-generation
>>> @cf/defog/sqlcoder-7b-2 7B 10K - S flagship docs
facebook / summarization
>>> @cf/facebook/bart-large-cnn - - flagship -
fblgit / text-generation
@cf/fblgit/una-cybertron-7b-v2-bf16 7B 15K - S fast -
google / text-embeddings
>>> @cf/google/embeddinggemma-300m 300M - - flagship -
google / text-generation
>>> @cf/google/gemma-4-26b-a4b-it 26B 256K - TS flagship -
>>> @cf/google/gemma-3-12b-it 12B 80K - LS flagship -
@cf/google/gemma-7b-it-lora 7B 4K - LS fast -
@hf/google/gemma-7b-it 7B 8K - LS fast docs
@cf/google/gemma-2b-it-lora 2B 8K - LS balanced -
huggingface / text-classification
@cf/huggingface/distilbert-sst-2-int8 - - balanced docs
ibm-granite / text-generation
>>> @cf/ibm-granite/granite-4.0-h-micro 131K - TS flagship -
leonardo / text-to-image
@cf/leonardo/lucid-origin - - balanced -
@cf/leonardo/phoenix-1.0 - - balanced -
llava-hf / image-to-text
@cf/llava-hf/llava-1.5-7b-hf 7B - - V fast -
lykon / text-to-image
@cf/lykon/dreamshaper-8-lcm - - balanced docs
meta / text-generation
>>> @cf/meta/llama-3.3-70b-instruct-fp8-fast 70B 24K - TS flagship -
>>> @cf/meta/llama-4-scout-17b-16e-instruct 17B 131K - TS flagship -
>>> @cf/meta/llama-3.2-11b-vision-instruct 11B 128K - VLS flagship -
>>> @cf/meta/llama-3.1-8b-instruct-awq 8B 8K - S flagship -
>>> @cf/meta/llama-guard-3-8b 8B 131K - LS flagship -
@cf/meta/llama-3-8b-instruct 8B 8K - S fast docs
@cf/meta/llama-3-8b-instruct-awq 8B 8K - S fast docs
@cf/meta/llama-3.1-8b-instruct-fp8 8B 32K - S fast -
@cf/meta/llama-2-7b-chat-fp16 7B 4K - S fast docs
@cf/meta/llama-2-7b-chat-int8 7B 8K - S fast -
@cf/meta/llama-3.2-3b-instruct 3B 80K - S fast -
@cf/meta/llama-3.2-1b-instruct 1B 60K - S fast -
meta / translation
@cf/meta/m2m100-1.2b 1.2B - - balanced docs
meta-llama / text-generation
@cf/meta-llama/llama-2-7b-chat-hf-lora 7B 8K - LS fast -
microsoft / image-classification
>>> @cf/microsoft/resnet-50 - - flagship docs
microsoft / text-generation
>>> @cf/microsoft/phi-2 2K - S flagship docs
mistral / text-generation
@cf/mistral/mistral-7b-instruct-v0.1 7B 3K - LS fast docs
@cf/mistral/mistral-7b-instruct-v0.2-lora 7B 15K - LS fast -
@hf/mistral/mistral-7b-instruct-v0.2 7B 3K - LS fast docs
mistralai / text-generation
>>> @cf/mistralai/mistral-small-3.1-24b-instruct 24B 128K - TS flagship -
moonshotai / text-generation
>>> @cf/moonshotai/kimi-k2.6 262K - TS flagship -
@cf/moonshotai/kimi-k2.5 256K - TS balanced -
myshell-ai / text-to-speech
@cf/myshell-ai/melotts - - balanced -
nexusflow / text-generation
@hf/nexusflow/starling-lm-7b-beta 7B 4K - S fast docs
nousresearch / text-generation
@hf/nousresearch/hermes-2-pro-mistral-7b 7B 24K - TS fast docs
nvidia / text-generation
>>> @cf/nvidia/nemotron-3-120b-a12b 120B 256K - TS flagship -
openai / automatic-speech-recognition
>>> @cf/openai/whisper-large-v3-turbo - - flagship -
@cf/openai/whisper - - balanced docs
@cf/openai/whisper-tiny-en - - fast -
openai / text-generation
>>> @cf/openai/gpt-oss-120b 120B 128K - TS flagship -
>>> @cf/openai/gpt-oss-20b 20B 128K - TS flagship -
openchat / text-generation
@cf/openchat/openchat-3.5-0106 8K - S balanced docs
pfnet / text-embeddings
@cf/pfnet/plamo-embedding-1b 1B - - fast -
pipecat-ai / dumb-pipe
@cf/pipecat-ai/smart-turn-v2 - - balanced -
qwen / text-embeddings
>>> @cf/qwen/qwen3-embedding-0.6b 0.6B 8K - flagship -
qwen / text-generation
>>> @cf/qwen/qwen2.5-coder-32b-instruct 32B 33K - LS flagship -
>>> @cf/qwen/qwq-32b 32B 24K - RLS flagship -
>>> @cf/qwen/qwen3-30b-a3b-fp8 30B 33K - TS flagship -
@cf/qwen/qwen1.5-14b-chat-awq 14B 8K - S balanced docs
@cf/qwen/qwen1.5-7b-chat-awq 7B 20K - S fast docs
@cf/qwen/qwen1.5-1.8b-chat 1.8B 32K - S balanced docs
@cf/qwen/qwen1.5-0.5b-chat 0.5B 32K - S balanced docs
runwayml / text-to-image
@cf/runwayml/stable-diffusion-v1-5-img2img - - balanced docs
@cf/runwayml/stable-diffusion-v1-5-inpainting - - balanced docs
stabilityai / text-to-image
>>> @cf/stabilityai/stable-diffusion-xl-base-1.0 - - flagship docs
thebloke / text-generation
@hf/thebloke/llama-2-13b-chat-awq 13B 4K - S balanced docs
@cf/thebloke/discolm-german-7b-v1-awq 7B 4K - S fast docs
@hf/thebloke/mistral-7b-instruct-v0.1-awq 7B 4K - S fast docs
@hf/thebloke/neural-chat-7b-v3-1-awq 7B 4K - S fast -
@hf/thebloke/openhermes-2.5-mistral-7b-awq 7B 4K - S fast -
@hf/thebloke/zephyr-7b-beta-awq 7B 4K - S fast docs
@hf/thebloke/deepseek-coder-6.7b-base-awq 6.7B 4K - S balanced -
@hf/thebloke/deepseek-coder-6.7b-instruct-awq 6.7B 4K - S balanced -
tiiuae / text-generation
@cf/tiiuae/falcon-7b-instruct 7B 4K - S fast docs
tinyllama / text-generation
@cf/tinyllama/tinyllama-1.1b-chat-v1.0 1.1B 2K - S fast docs
unum / image-to-text
@cf/unum/uform-gen2-qwen-500m 500M - - V balanced docs
zai-org / text-generation
>>> @cf/zai-org/glm-4.7-flash 131K - TS flagship -