Inference for text-embeddings in Python
Lightweight alternative to LangChain for composing LLMs
NVIDIA Framework for LLM Inference(Transitioned to TensorRT-LLM)
LLM inference in C/C++.
a chat interface crafted with llama.cpp for running Alpaca models. No API keys, entirely self-hosted!
An interactive chat project that leverages Ollama/OpenAI/MistralAI LLMs for rapid understanding and navigation of GitHub code repository or compressed file resources.
Test your prompts. Evaluate and compare LLM outputs, catch regressions, and improve prompt quality.
Your email address will not be published. Required fields are marked *
Comment *
Name *
Email *
Website
Captcha: 16 - 18 = ?*
Save my name, email, and website in this browser for the next time I comment.
Lightweight alternative to LangChain for composing LLMs