| ai4bharat / translation |
| @cf/ai4bharat/indictrans2-en-indic-1B 1B |
- |
- |
— |
|
fast |
- |
| aisingapore / text-generation |
| >>> @cf/aisingapore/gemma-sea-lion-v4-27b-it 27B |
128K |
- |
— |
S |
flagship |
- |
| baai / text-classification |
| >>> @cf/baai/bge-reranker-base |
- |
- |
— |
|
flagship |
- |
| baai / text-embeddings |
| >>> @cf/baai/bge-large-en-v1.5 |
- |
- |
— |
|
flagship |
docs |
| >>> @cf/baai/bge-m3 |
60K |
- |
— |
|
flagship |
- |
| @cf/baai/bge-base-en-v1.5 |
154K |
- |
— |
|
balanced |
docs |
| @cf/baai/bge-small-en-v1.5 |
- |
- |
— |
|
fast |
docs |
| black-forest-labs / text-to-image |
| >>> @cf/black-forest-labs/flux-1-schnell |
- |
- |
— |
|
flagship |
- |
| >>> @cf/black-forest-labs/flux-2-dev |
- |
- |
— |
|
flagship |
- |
| @cf/black-forest-labs/flux-2-klein-9b 9B |
- |
- |
— |
|
balanced |
- |
| @cf/black-forest-labs/flux-2-klein-4b 4B |
- |
- |
— |
|
balanced |
- |
| bytedance / text-to-image |
| @cf/bytedance/stable-diffusion-xl-lightning |
- |
- |
— |
|
balanced |
docs |
| deepgram / automatic-speech-recognition |
| >>> @cf/deepgram/flux |
- |
- |
— |
|
flagship |
- |
| >>> @cf/deepgram/nova-3 |
- |
- |
— |
|
flagship |
- |
| deepgram / text-to-speech |
| >>> @cf/deepgram/aura-2-es |
- |
- |
— |
|
flagship |
- |
| @cf/deepgram/aura-1 |
- |
- |
— |
|
balanced |
- |
| @cf/deepgram/aura-2-en |
- |
- |
— |
|
balanced |
- |
| deepseek-ai / text-generation |
| >>> @cf/deepseek-ai/deepseek-r1-distill-qwen-32b 32B |
80K |
- |
— |
RS |
flagship |
- |
| >>> @cf/deepseek-ai/deepseek-math-7b-instruct 7B |
4K |
- |
— |
S |
flagship |
docs |
| defog / text-generation |
| >>> @cf/defog/sqlcoder-7b-2 7B |
10K |
- |
— |
S |
flagship |
docs |
| facebook / summarization |
| >>> @cf/facebook/bart-large-cnn |
- |
- |
— |
|
flagship |
- |
| fblgit / text-generation |
| @cf/fblgit/una-cybertron-7b-v2-bf16 7B |
15K |
- |
— |
S |
fast |
- |
| google / text-embeddings |
| >>> @cf/google/embeddinggemma-300m 300M |
- |
- |
— |
|
flagship |
- |
| google / text-generation |
| >>> @cf/google/gemma-4-26b-a4b-it 26B |
256K |
- |
— |
TS |
flagship |
- |
| >>> @cf/google/gemma-3-12b-it 12B |
80K |
- |
— |
LS |
flagship |
- |
| @cf/google/gemma-7b-it-lora 7B |
4K |
- |
— |
LS |
fast |
- |
| @hf/google/gemma-7b-it 7B |
8K |
- |
— |
LS |
fast |
docs |
| @cf/google/gemma-2b-it-lora 2B |
8K |
- |
— |
LS |
balanced |
- |
| huggingface / text-classification |
| @cf/huggingface/distilbert-sst-2-int8 |
- |
- |
— |
|
balanced |
docs |
| ibm-granite / text-generation |
| >>> @cf/ibm-granite/granite-4.0-h-micro |
131K |
- |
— |
TS |
flagship |
- |
| leonardo / text-to-image |
| @cf/leonardo/lucid-origin |
- |
- |
— |
|
balanced |
- |
| @cf/leonardo/phoenix-1.0 |
- |
- |
— |
|
balanced |
- |
| llava-hf / image-to-text |
| @cf/llava-hf/llava-1.5-7b-hf 7B |
- |
- |
— |
V |
fast |
- |
| lykon / text-to-image |
| @cf/lykon/dreamshaper-8-lcm |
- |
- |
— |
|
balanced |
docs |
| meta / text-generation |
| >>> @cf/meta/llama-3.3-70b-instruct-fp8-fast 70B |
24K |
- |
— |
TS |
flagship |
- |
| >>> @cf/meta/llama-4-scout-17b-16e-instruct 17B |
131K |
- |
— |
TS |
flagship |
- |
| >>> @cf/meta/llama-3.2-11b-vision-instruct 11B |
128K |
- |
— |
VLS |
flagship |
- |
| >>> @cf/meta/llama-3.1-8b-instruct-awq 8B |
8K |
- |
— |
S |
flagship |
- |
| >>> @cf/meta/llama-guard-3-8b 8B |
131K |
- |
— |
LS |
flagship |
- |
| @cf/meta/llama-3-8b-instruct 8B |
8K |
- |
— |
S |
fast |
docs |
| @cf/meta/llama-3-8b-instruct-awq 8B |
8K |
- |
— |
S |
fast |
docs |
| @cf/meta/llama-3.1-8b-instruct-fp8 8B |
32K |
- |
— |
S |
fast |
- |
| @cf/meta/llama-2-7b-chat-fp16 7B |
4K |
- |
— |
S |
fast |
docs |
| @cf/meta/llama-2-7b-chat-int8 7B |
8K |
- |
— |
S |
fast |
- |
| @cf/meta/llama-3.2-3b-instruct 3B |
80K |
- |
— |
S |
fast |
- |
| @cf/meta/llama-3.2-1b-instruct 1B |
60K |
- |
— |
S |
fast |
- |
| meta / translation |
| @cf/meta/m2m100-1.2b 1.2B |
- |
- |
— |
|
balanced |
docs |
| meta-llama / text-generation |
| @cf/meta-llama/llama-2-7b-chat-hf-lora 7B |
8K |
- |
— |
LS |
fast |
- |
| microsoft / image-classification |
| >>> @cf/microsoft/resnet-50 |
- |
- |
— |
|
flagship |
docs |
| microsoft / text-generation |
| >>> @cf/microsoft/phi-2 |
2K |
- |
— |
S |
flagship |
docs |
| mistral / text-generation |
| @cf/mistral/mistral-7b-instruct-v0.1 7B |
3K |
- |
— |
LS |
fast |
docs |
| @cf/mistral/mistral-7b-instruct-v0.2-lora 7B |
15K |
- |
— |
LS |
fast |
- |
| @hf/mistral/mistral-7b-instruct-v0.2 7B |
3K |
- |
— |
LS |
fast |
docs |
| mistralai / text-generation |
| >>> @cf/mistralai/mistral-small-3.1-24b-instruct 24B |
128K |
- |
— |
TS |
flagship |
- |
| moonshotai / text-generation |
| >>> @cf/moonshotai/kimi-k2.6 |
262K |
- |
— |
TS |
flagship |
- |
| @cf/moonshotai/kimi-k2.5 |
256K |
- |
— |
TS |
balanced |
- |
| myshell-ai / text-to-speech |
| @cf/myshell-ai/melotts |
- |
- |
— |
|
balanced |
- |
| nexusflow / text-generation |
| @hf/nexusflow/starling-lm-7b-beta 7B |
4K |
- |
— |
S |
fast |
docs |
| nousresearch / text-generation |
| @hf/nousresearch/hermes-2-pro-mistral-7b 7B |
24K |
- |
— |
TS |
fast |
docs |
| nvidia / text-generation |
| >>> @cf/nvidia/nemotron-3-120b-a12b 120B |
256K |
- |
— |
TS |
flagship |
- |
| openai / automatic-speech-recognition |
| >>> @cf/openai/whisper-large-v3-turbo |
- |
- |
— |
|
flagship |
- |
| @cf/openai/whisper |
- |
- |
— |
|
balanced |
docs |
| @cf/openai/whisper-tiny-en |
- |
- |
— |
|
fast |
- |
| openai / text-generation |
| >>> @cf/openai/gpt-oss-120b 120B |
128K |
- |
— |
TS |
flagship |
- |
| >>> @cf/openai/gpt-oss-20b 20B |
128K |
- |
— |
TS |
flagship |
- |
| openchat / text-generation |
| @cf/openchat/openchat-3.5-0106 |
8K |
- |
— |
S |
balanced |
docs |
| pfnet / text-embeddings |
| @cf/pfnet/plamo-embedding-1b 1B |
- |
- |
— |
|
fast |
- |
| pipecat-ai / dumb-pipe |
| @cf/pipecat-ai/smart-turn-v2 |
- |
- |
— |
|
balanced |
- |
| qwen / text-embeddings |
| >>> @cf/qwen/qwen3-embedding-0.6b 0.6B |
8K |
- |
— |
|
flagship |
- |
| qwen / text-generation |
| >>> @cf/qwen/qwen2.5-coder-32b-instruct 32B |
33K |
- |
— |
LS |
flagship |
- |
| >>> @cf/qwen/qwq-32b 32B |
24K |
- |
— |
RLS |
flagship |
- |
| >>> @cf/qwen/qwen3-30b-a3b-fp8 30B |
33K |
- |
— |
TS |
flagship |
- |
| @cf/qwen/qwen1.5-14b-chat-awq 14B |
8K |
- |
— |
S |
balanced |
docs |
| @cf/qwen/qwen1.5-7b-chat-awq 7B |
20K |
- |
— |
S |
fast |
docs |
| @cf/qwen/qwen1.5-1.8b-chat 1.8B |
32K |
- |
— |
S |
balanced |
docs |
| @cf/qwen/qwen1.5-0.5b-chat 0.5B |
32K |
- |
— |
S |
balanced |
docs |
| runwayml / text-to-image |
| @cf/runwayml/stable-diffusion-v1-5-img2img |
- |
- |
— |
|
balanced |
docs |
| @cf/runwayml/stable-diffusion-v1-5-inpainting |
- |
- |
— |
|
balanced |
docs |
| stabilityai / text-to-image |
| >>> @cf/stabilityai/stable-diffusion-xl-base-1.0 |
- |
- |
— |
|
flagship |
docs |
| thebloke / text-generation |
| @hf/thebloke/llama-2-13b-chat-awq 13B |
4K |
- |
— |
S |
balanced |
docs |
| @cf/thebloke/discolm-german-7b-v1-awq 7B |
4K |
- |
— |
S |
fast |
docs |
| @hf/thebloke/mistral-7b-instruct-v0.1-awq 7B |
4K |
- |
— |
S |
fast |
docs |
| @hf/thebloke/neural-chat-7b-v3-1-awq 7B |
4K |
- |
— |
S |
fast |
- |
| @hf/thebloke/openhermes-2.5-mistral-7b-awq 7B |
4K |
- |
— |
S |
fast |
- |
| @hf/thebloke/zephyr-7b-beta-awq 7B |
4K |
- |
— |
S |
fast |
docs |
| @hf/thebloke/deepseek-coder-6.7b-base-awq 6.7B |
4K |
- |
— |
S |
balanced |
- |
| @hf/thebloke/deepseek-coder-6.7b-instruct-awq 6.7B |
4K |
- |
— |
S |
balanced |
- |
| tiiuae / text-generation |
| @cf/tiiuae/falcon-7b-instruct 7B |
4K |
- |
— |
S |
fast |
docs |
| tinyllama / text-generation |
| @cf/tinyllama/tinyllama-1.1b-chat-v1.0 1.1B |
2K |
- |
— |
S |
fast |
docs |
| unum / image-to-text |
| @cf/unum/uform-gen2-qwen-500m 500M |
- |
- |
— |
V |
balanced |
docs |
| zai-org / text-generation |
| >>> @cf/zai-org/glm-4.7-flash |
131K |
- |
— |
TS |
flagship |
- |