Stars
SGLang is a fast serving framework for large language models and vision language models.
A DreamGen Native App + Some added features and fixes
Scripts I use to customize my experience on DreamGen AI
a lightweight no-dependency fork from transformers.js (only tokenizers)
Convert Compute And Books Into Instruct-Tuning Datasets! Makes: QA, RP, Classifiers.
An OpenAI API compatible LLM inference server based on ExLlamaV2.
MII makes low-latency and high-throughput inference possible, powered by DeepSpeed.
Simple and efficient pytorch-native transformer text generation in <1000 LOC of python.
LightLLM is a Python-based LLM (Large Language Model) inference and serving framework, notable for its lightweight design, easy scalability, and high-speed performance.
Finetune Llama 3.3, DeepSeek-R1, Gemma 3 & Reasoning LLMs 2x faster with 70% less memory! 🦥