- Beijing
-
11:50
(UTC +08:00)
Pinned Loading
-
vllm-project/vllm
vllm-project/vllm PublicA high-throughput and memory-efficient inference and serving engine for LLMs
-
bytedance/lightseq
bytedance/lightseq PublicLightSeq: A High Performance Library for Sequence Processing and Generation
-
microsoft/Megatron-DeepSpeed
microsoft/Megatron-DeepSpeed PublicForked from NVIDIA/Megatron-LM
Ongoing research training transformer language models at scale, including: BERT & GPT-2
-
AniZpZ/AutoSmoothQuant
AniZpZ/AutoSmoothQuant PublicAn easy-to-use package for implementing SmoothQuant for LLMs
-
IST-DASLab/marlin
IST-DASLab/marlin PublicFP16xINT4 LLM inference kernel that can achieve near-ideal ~4x speedups up to medium batchsizes of 16-32 tokens.
Something went wrong, please refresh the page to try again.
If the problem persists, check the GitHub status page or contact support.
If the problem persists, check the GitHub status page or contact support.