Inference for text-embeddings in Python
Playground for devs to finetune & deploy LLMs
Simple API for deploying any RAG or LLM that you want adding plugins.
An open-source GPU cluster manager for running LLMs
Fine-tune, serve, deploy, and monitor any open-source LLMs in production. Used in production at BentoML for LLMs-based applications.
Formerly langchain-ChatGLM, local knowledge based LLM (like ChatGLM) QA app with langchain.
Confidently evaluate, test, and ship LLM applications with a suite of observability tools to calibrate language model outputs across your dev and production lifecycle.
Your email address will not be published. Required fields are marked *
Comment *
Name *
Email *
Website
Captcha: 10 + 14 = ?*
Save my name, email, and website in this browser for the next time I comment.
Playground for devs to finetune & deploy LLMs