Configs2026年4月7日·1 分钟阅读

Cloudflare AI Workers — Deploy AI Apps at the Edge

Run AI models on Cloudflare's global edge network. Workers AI provides serverless inference for LLMs, embeddings, image generation, and speech-to-text at low latency.

What is Workers AI?

Cloudflare Workers AI runs AI model inference on the global edge network — no GPU management, auto-scaling, pay-per-request.

In one sentence: Cloudflare Workers AI provides serverless AI inference in 300+ cities worldwide, supporting Llama, Stable Diffusion, Whisper, and more.

For: Developers needing low-latency serverless AI deployment.

Core Features

1. Preloaded Model Catalog

Text generation, embeddings, image generation, speech-to-text, and more.

2. Built-In Vector Database

Vectorize provides embedding storage and query.

3. AI Gateway

Routing, caching, and monitoring for AI API calls.

4. Edge Deployment

GPU clusters in 300+ cities worldwide with p50 latency < 50ms.

FAQ

Q: Is there a free tier? A: Yes — 10,000 neurons per day are free.

Q: How does it compare to AWS Bedrock? A: Workers AI is edge-native with lower latency, simpler, and cheaper for small-to-medium workloads.

🙏

来源与感谢

讨论

登录后参与讨论。
还没有评论,来写第一条吧。

相关资产