fbpx
Techitup Middle East
AIB2B Technology

Cloudflare: Global Deployment for AI Applications with Hugging Face 

Workers AI and Hugging Face integration generally available; deploying serverless AI is now easier and more affordable than ever 

Cloudflare, Inc. announced that developers can now carry deployment for AI applications on Cloudflare’s global network in one simple click directly from Hugging Face, the leading open and collaborative platform for AI builders. With Workers AI now generally available, Cloudflare is the first serverless inference partner integrated on the Hugging Face Hub for deploying models, enabling developers to quickly, easily, and affordably deploy AI globally, without managing infrastructure or paying for unused compute capacity. 

Despite significant strides in AI innovation, there is still a disconnect between its potential and the value it brings businesses. Organizations and their developers need to be able to experiment and iterate quickly and affordably, without having to set up, manage, or maintain GPUs or infrastructure. Businesses are in need of a straightforward platform that unlocks speed, security, performance, observability, and compliance to bring innovative, production-ready applications to their customers faster. 

Workers AI is generally available with GPUs now deployed in more than 150 cities globally 

Today, Workers AI is generally available, providing the end-to-end infrastructure needed to scale and deploy AI models efficiently and affordably for the next era of AI applications. Cloudflare now has GPUs deployed across more than 150 cities globally, most recently launching in Cape Town, Durban, Johannesburg, and Lagos for the first locations in Africa, as well as Amman, Buenos Aires, Mexico City, Mumbai, New Delhi, and Seoul, to provide low-latency inference around the world. Workers AI is also expanding to support fine-tuned model weights, enabling organizations to build and deploy more specialized, domain-specific applications. 

In addition to Workers AI, Cloudflare’s AI Gateway offers a control plane for your AI applications, allowing developers to dynamically evaluate and route requests to different models and providers, eventually enabling developers to use data to create fine tunes and run the fine-tuned jobs directly on the Workers AI platform. 

Cloudflare powers one-click deployment with Hugging Face 

With Workers AI generally available, developers can now deploy AI models in one click directly from Hugging Face, for the fastest way to access a variety of models and run inference requests on Cloudflare’s global network of GPUs. Developers can choose one of the popular open source models and then simply click “Deploy to Cloudflare Workers AI” to deploy a model instantly. There are 14 curated Hugging Face models now optimized for Cloudflare’s global serverless inference platform, supporting three different task categories including text generation, embeddings, and sentence similarity. 

AI-first companies are building with Workers AI 

Companies around the world trust Workers AI and Cloudflare’s global network to power their AI applications, including: 

Related posts

Check Point Unveils Quantum Force Gateway Series

Editor

SolarWinds Bolsters Database Observability for Cloud-Native Platform  

Editor

Veeam Leads Data Replication and Protection Software in 2H23 

Editor

Leave a Comment