
Inference Engines
OpenLLM
Fine-tune, serve, deploy, and monitor any open-source LLMs in production. Used in production at BentoML for LLMs-based applications.
Fine-tune, serve, deploy, and monitor any open-source LLMs in production. Used in production at BentoML for LLMs-based applications.
A Device-Inference framework, including LLM Inference on device(Mobile Phone/PC/IOT)