Use ChatGPT On Wechat via wechaty
A high-throughput and memory-efficient inference and serving engine for LLMs.
WebAssembly binding for llama.cpp - Enabling in-browser LLM inference
A Device-Inference framework, including LLM Inference on device(Mobile Phone/PC/IOT)
AI gateway and marketplace for developers, enables streamlined integration of AI features into products
Create, deploy and operate Actions using Python anywhere to enhance your AI agents and assistants. Batteries included with an extensive set of libraries, helpers and logging.
An interactive chat project that leverages Ollama/OpenAI/MistralAI LLMs for rapid understanding and navigation of GitHub code repository or compressed file resources.
Your email address will not be published. Required fields are marked *
Comment *
Name *
Email *
Website
Captcha: 19 + 15 = ?*
Save my name, email, and website in this browser for the next time I comment.
A high-throughput and memory-efficient inference and serving engine for LLMs.