-
π Iβm currently working on Meituan, focusing on algorithms and engineering related to LLM inference acceleration, familar with model quantization.
-
π» I'm lucky to contribute to some open source projects: SGLang, vLLM, TorchAO, Megatron-DeepSpeed and LightSeq.
-
π I'm proud to build some projects from scratch:
- AutoSmoothQuant: An easy-to-use package for implementing SmoothQuant for LLMs.
- QQQ: QQQ is an innovative and hardware-optimized W4A8 quantization solution for LLMs.
-
π« Contact: [email protected]
-
π Google Scholar: https://scholar.google.com/citations?hl=zh-CN&user=MBR97ZIAAAAJ
- Beijing
-
11:08
(UTC +08:00) - https://scholar.google.com/citations?hl=zh-CN&user=MBR97ZIAAAAJ
Pinned Loading
-
vllm-project/vllm
vllm-project/vllm PublicA high-throughput and memory-efficient inference and serving engine for LLMs
-
bytedance/lightseq
bytedance/lightseq PublicLightSeq: A High Performance Library for Sequence Processing and Generation
-
deepspeedai/Megatron-DeepSpeed
deepspeedai/Megatron-DeepSpeed PublicForked from NVIDIA/Megatron-LM
Ongoing research training transformer language models at scale, including: BERT & GPT-2
-
AniZpZ/AutoSmoothQuant
AniZpZ/AutoSmoothQuant PublicAn easy-to-use package for implementing SmoothQuant for LLMs
-
sglang
sglang PublicForked from sgl-project/sglang
SGLang is a fast serving framework for large language models and vision language models.
Python
Something went wrong, please refresh the page to try again.
If the problem persists, check the GitHub status page or contact support.
If the problem persists, check the GitHub status page or contact support.