Data integration platform for LLMs.
An interactive chat project that leverages Ollama/OpenAI/MistralAI LLMs for rapid understanding and navigation of GitHub code repository or compressed file resources.
Lightweight alternative to LangChain for composing LLMs
FlexLLMGen is a high-throughput generation engine for running large language models with limited GPU memory. FlexLLMGen allows high-throughput generation by IO-efficient offloading, compression, and large effective batch sizes.
Comprehensive set of tools for working with local LLMs for various tasks.
a toolkit for deploying and serving Large Language Models (LLMs).
An open-source GPU cluster manager for running LLMs
Your email address will not be published. Required fields are marked *
Comment *
Name *
Email *
Website
Captcha: 18 - 19 = ?*
Save my name, email, and website in this browser for the next time I comment.
An interactive chat project that leverages Ollama/OpenAI/MistralAI LLMs for rapid understanding and navigation of GitHub code repository or compressed file resources.