为什么Replicate加入Cloudflare
Why Replicate is joining Cloudflare

原始链接: https://blog.cloudflare.com/why-replicate-joining-cloudflare/

Replicate,一个生成式AI平台,已被Cloudflare收购。Replicate成立于2019年,旨在通过简化部署,使开发者更容易获取强大的AI模型——通过Cog工具和Replicate平台本身,抽象机器学习和GPU管理的复杂性,从而实现AI的民主化。 该公司随着Stable Diffusion的发布而受到关注,并成为创新AI应用中心。然而,Replicate意识到现代AI应用需要的不仅仅是模型推理,还包括完整的服务栈。 加入Cloudflare将使Replicate能够实现其构建完整AI基础设施层的愿景,利用Cloudflare的网络、无服务器函数(Workers)和存储解决方案,提供更快、更高效的AI体验——包括边缘计算和简化的模型流水线。Replicate为率先提供生成式AI服务并建立关键行业标准而感到自豪。

Hacker News 新闻 | 过去 | 评论 | 提问 | 展示 | 招聘 | 提交 登录 为什么 Replicate 要加入 Cloudflare (cloudflare.com) 9 分,作者 chmaynard 2 小时前 | 隐藏 | 过去 | 收藏 | 3 条评论 fodkodrasz 15 分钟前 | 下一个 [–] 我搜索了,但在营销文章中找不到“收购”、“钱”、“美元”或“股票”这些词,所以它显然没有回答标题中的问题。交易价值是多少?回复 dmoy 8 分钟前 | 父评论 | 下一个 [–] > 交易价值是多少? 我认为这被有意隐瞒,所以知道的人都不会说。 ChrisArchitect 1 小时前 | 上一个 [–] 相关: Replicate 要加入 Cloudflare https://news.ycombinator.com/item?id=45953702 指南 | 常见问题 | 列表 | API | 安全 | 法律 | 申请 YC | 联系 搜索:
相关文章

原文

We're happy to announce that as of today Replicate is officially part of Cloudflare.

When we started Replicate in 2019, OpenAI had just open sourced GPT-2, and few people outside of the machine learning community paid much attention to AI. But for those of us in the field, it felt like something big was about to happen. Remarkable models were being created in academic labs, but you needed a metaphorical lab coat to be able to run them.

We made it our mission to get research models out of the lab into the hands of developers. We wanted programmers to creatively bend and twist these models into products that the researchers would never have thought of.

We approached this as a tooling problem. Just like tools like Heroku made it possible to run websites without managing web servers, we wanted to build tools for running models without having to understand backpropagation or deal with CUDA errors.

The first tool we built was Cog: a standard packaging format for machine learning models. Then we built Replicate as the platform to run Cog models as API endpoints in the cloud. We abstracted away both the low-level machine learning, and the complicated GPU cluster management you need to run inference at scale.

It turns out the timing was just right. When Stable Diffusion was released in 2022 we had mature infrastructure that could handle the massive developer interest in running these models. A ton of fantastic apps and products were built on Replicate, apps that often ran a single model packaged in a slick UI to solve a particular use case.

Since then, AI Engineering has matured into a serious craft. AI apps are no longer just about running models. The modern AI stack has model inference, but also microservices, content delivery, object storage, caching, databases, telemetry, etc. We see many of our customers building complex heterogenous stacks where the Replicate models are one part of a higher-order system across several platforms.

This is why we’re joining Cloudflare. Replicate has the tools and primitives for running models. Cloudflare has the best network, Workers, R2, Durable Objects, and all the other primitives you need to build a full AI stack.

The AI stack lives entirely on the network. Models run on data center GPUs and are glued together by small cloud functions that call out to vector databases, fetch objects from blob storage, call MCP servers, etc. “The network is the computer” has never been more true.

At Cloudflare, we’ll now be able to build the AI infrastructure layer we have dreamed of since we started. We’ll be able to do things like run fast models on the edge, run model pipelines on instantly-booting Workers, stream model inputs and outputs with WebRTC, etc.

We’re proud of what we’ve built at Replicate. We were the first generative AI serving platform, and we defined the abstractions and design patterns that most of our peers have adopted. We’ve grown a wonderful community of builders and researchers around our product.

联系我们 contact @ memedata.com