Skip to content

LightLLM is a Python-based LLM (Large Language Model) inference and serving framework, notable for its lightweight design, easy scalability, and high-speed performance.

License

Notifications You must be signed in to change notification settings

ModelTC/lightllm

Folders and files

NameName
Last commit message
Last commit date

Latest commit

SangChengCsangchengmeng
and
sangchengmeng
Apr 7, 2025
11ec65a · Apr 7, 2025
Oct 9, 2024
Jul 22, 2023
Sep 4, 2024
Mar 28, 2025
Mar 28, 2025
Apr 7, 2025
Apr 3, 2025
Aug 12, 2023
Apr 2, 2025
Mar 28, 2025
Mar 18, 2024
Jan 12, 2024
Apr 7, 2025
Jul 22, 2023
Mar 28, 2025
Jul 22, 2023
Sep 18, 2023
Oct 16, 2023
Apr 7, 2025
Apr 2, 2025

Repository files navigation

LightLLM

docs Docker stars visitors Discord Banner license

LightLLM is a Python-based LLM (Large Language Model) inference and serving framework, notable for its lightweight design, easy scalability, and high-speed performance. LightLLM harnesses the strengths of numerous well-regarded open-source implementations, including but not limited to FasterTransformer, TGI, vLLM, and FlashAttention.

English Docs | 中文文档 | Blogs

News

  • [2025/02] 🔥 LightLLM v1.0.0 release, achieving the fastest DeepSeek-R1 serving performance on single H200 machine.

Get started

Performance

Learn more in the release blogs: v1.0.0 blog.

FAQ

Please refer to the FAQ for more information.

Projects using lightllm

We welcome any coopoeration and contribution. If there is a project requires lightllm's support, please contact us via email or create a pull request.

  1. LazyLLM: Easyest and lazyest way for building multi-agent LLMs applications.

    Once you have installed lightllm and lazyllm, and then you can use the following code to build your own chatbot:

    from lazyllm import TrainableModule, deploy, WebModule
    # Model will be download automatically if you have an internet connection
    m = TrainableModule('internlm2-chat-7b').deploy_method(deploy.lightllm)
    WebModule(m).start().wait()

    Documents: https://lazyllm.readthedocs.io/

Community

For further information and discussion, join our discord server. Welcome to be a member and look forward to your contribution!

License

This repository is released under the Apache-2.0 license.

Acknowledgement

We learned a lot from the following projects when developing LightLLM.

About

LightLLM is a Python-based LLM (Large Language Model) inference and serving framework, notable for its lightweight design, easy scalability, and high-speed performance.

Topics

Resources

License

Stars

Watchers

Forks

Languages