Effortlessly run LLM backends, APIs, frontends, and services with one command.
-
Updated
Sep 30, 2024 - TypeScript
Effortlessly run LLM backends, APIs, frontends, and services with one command.
A fast, lightweight, parallel inference server for Llama LLMs.
LLM translation toolkit for subtitle files.
Various image processing scripts.
Add a description, image, and links to the exllamav2 topic page so that developers can more easily learn about it.
To associate your repository with the exllamav2 topic, visit your repo's landing page and select "manage topics."