An open-source GPU cluster manager for running LLMs
Seamlessly integrate LLMs as Python functions
Confidently evaluate, test, and ship LLM applications with a suite of observability tools to calibrate language model outputs across your dev and production lifecycle.
Inference for text-embeddings in Python
WebAssembly binding for llama.cpp - Enabling in-browser LLM inference
Inference for text-embeddings in Rust, HFOIL Licence.
Nvidia Framework for LLM Inference
Your email address will not be published. Required fields are marked *
Comment *
Name *
Email *
Website
Captcha: 15 + 14 = ?*
Save my name, email, and website in this browser for the next time I comment.
Seamlessly integrate LLMs as Python functions