a toolkit for deploying and serving Large Language Models (LLMs).
An open-source GPU cluster manager for running LLMs
LLM inference in C/C++.
A distributed multi-model LLM serving system with web UI and OpenAI-compatible RESTful APIs.
An interactive chat project that leverages Ollama/OpenAI/MistralAI LLMs for rapid understanding and navigation of GitHub code repository or compressed file resources.
Interact with LLM using Ollama models(or openAI, mistralAI)via pure shell scripts on your Linux(or MacOS) system, enhancing intelligent system management without any dependencies.
Data integration platform for LLMs.
Your email address will not be published. Required fields are marked *
Comment *
Name *
Email *
Website
Captcha: 16 - 18 = ?*
Save my name, email, and website in this browser for the next time I comment.
An open-source GPU cluster manager for running LLMs