Playground for devs to finetune & deploy LLMs
Data integration platform for LLMs.
An open-source GPU cluster manager for running LLMs
a toolkit for deploying and serving Large Language Models (LLMs).
A more memory-efficient rewrite of the HF transformers implementation of Llama for use with quantized weights.
Lightweight alternative to LangChain for composing LLMs
NVIDIA Framework for LLM Inference(Transitioned to TensorRT-LLM)
Your email address will not be published. Required fields are marked *
Comment *
Name *
Email *
Website
Captcha: 11 + 15 = ?*
Save my name, email, and website in this browser for the next time I comment.
Data integration platform for LLMs.