Inference for text-embeddings in Rust, HFOIL Licence.
Blazingly fast LLM inference.
An interactive chat project that leverages Ollama/OpenAI/MistralAI LLMs for rapid understanding and navigation of GitHub code repository or compressed file resources.
AI gateway and marketplace for developers, enables streamlined integration of AI features into products
Nvidia Framework for LLM Inference
a chat interface crafted with llama.cpp for running Alpaca models. No API keys, entirely self-hosted!
Get up and running with Llama 3, Mistral, Gemma, and other large language models.
Your email address will not be published. Required fields are marked *
Comment *
Name *
Email *
Website
Captcha: 14 + 20 = ?*
Save my name, email, and website in this browser for the next time I comment.
Blazingly fast LLM inference.