Skip to content

Bump the flash-attn-4 version ; Add the cu13 extra#12

Merged
MasterJH5574 merged 1 commit intomlc-ai:mainfrom
haok1402:0404-bump-flashattn4
Apr 4, 2026
Merged

Bump the flash-attn-4 version ; Add the cu13 extra#12
MasterJH5574 merged 1 commit intomlc-ai:mainfrom
haok1402:0404-bump-flashattn4

Conversation

@haok1402
Copy link
Copy Markdown
Collaborator

@haok1402 haok1402 commented Apr 4, 2026

Our framework requires cu13.0. The official flash-attn-4 pypi page mentions having cu13 extra for the best performance. Additionally, we ping the flash-attn-4 to its latest beta release 4.0.0b7.

Copy link
Copy Markdown

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request updates the flash-attn-4 dependency in pyproject.toml to version 4.0.0b7 with the cu13 extra. A review comment suggests pinning this dependency to the exact beta version to prevent potential instability from future releases.

@MasterJH5574 MasterJH5574 merged commit bfb2826 into mlc-ai:main Apr 4, 2026
1 check passed
@haok1402 haok1402 deleted the 0404-bump-flashattn4 branch April 4, 2026 21:34
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants