CompactifAI 推理 API
嗨,HN,
我们一直在进行模型压缩,并已在我们的AWS集群上部署了压缩后的模型以及原始版本,用户可以通过推理API访问它们。我们非常希望开发者能对集成体验和整体表现提供反馈。如果您想尝试,请发送邮件至 hyperscalers@multiversecomputing.com,我们将为您发送API密钥和简要文档。我们将向有限数量的用户开放访问权限,为期3个月,且没有使用限制,因此您可以在此期间运行实际工作负载。
模型基准测试可以在这里查看: https://artificialanalysis.ai/providers/compactifai
查看原文
Hi HN,<p>We’ve been working on model compression and have deployed our compressed models—as well as the original versions—on our AWS cluster, accessible via an inference API.
We’d love feedback from developers on the integration experience and overall behavior. If you’d like to try it, email hyperscalers@multiversecomputing.com and we’ll send an API key and brief docs. We’re granting access to a limited number of users for a 3-month period with no usage caps, so you can run real workloads during that window.<p>Model benchmarking is available here: https://artificialanalysis.ai/providers/compactifai