Limits
During the open beta, the following limits will be in place:
Inference requests per minute (per model)
- @cf/meta/llama-2-7b-chat-int8 - 5 reqs/min
- @cf/openai/whisper - 120 reqs/min
- @cf/meta/m2m100-1.2b - 120 reqs/min
- @cf/huggingface/distilbert-sst-2-int8 - 180 reqs/min
- @cf/microsoft/resnet-50 - 180 reqs/min
- @cf/baai/bge-base-en-v1.5 -1 80 reqs/min
Other Limits
- @cf/meta/llama-2-7b-chat-int8 (max tokens) - 768 input / 256 output
- @cf/meta/m2m100-1.2b (max tokens) - 256 output