Lightweight alternative to LangChain for composing LLMs
Blazingly fast LLM inference.
a chat interface crafted with llama.cpp for running Alpaca models. No API keys, entirely self-hosted!
A distributed multi-model LLM serving system with web UI and OpenAI-compatible RESTful APIs.
A method designed to enhance the efficiency of Transformer models
MII makes low-latency and high-throughput inference, similar to vLLM powered by DeepSpeed.
Open Source LLM Engineering Platform 🪢 Tracing, Evaluations, Prompt Management, Evaluations and Playground.
Your email address will not be published. Required fields are marked *
Comment *
Name *
Email *
Website
Captcha: 19 - 13 = ?*
Save my name, email, and website in this browser for the next time I comment.
Blazingly fast LLM inference.