Show use torch.compile on TorchAO charts #6368
Open
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Some data is missing on the chart otherwise. For example https://hud.pytorch.org/benchmark/llms?startTime=Thu%2C%2027%20Feb%202025%2001%3A12%3A08%20GMT&stopTime=Thu%2C%2006%20Mar%202025%2001%3A12%3A08%20GMT&granularity=day&lBranch=main&lCommit=6946094f6fe9ff77ca02736567ddf2fc031d5fd300ce7c60d832cf9c301c4c48&rBranch=main&rCommit=6946094f6fe9ff77ca02736567ddf2fc031d5fd300ce7c60d832cf9c301c4c48&repoName=pytorch%2Fao&benchmarkName=&modelName=AlbertForMaskedLM&backendName=&modeName=inference&dtypeName=autoquant&deviceName=cuda%20(NVIDIA%20H100)&archName=All%20Platforms shows only 0.99 on
compile v.s eager speedup
chart, while it should include the other value (6.85)Preview
https://torchci-git-fork-huydhn-fix-ao-data-fbopensource.vercel.app/benchmark/llms?startTime=Thu%2C%2027%20Feb%202025%2001%3A12%3A08%20GMT&stopTime=Thu%2C%2006%20Mar%202025%2001%3A12%3A08%20GMT&granularity=day&lBranch=main&lCommit=6946094f6fe9ff77ca02736567ddf2fc031d5fd300ce7c60d832cf9c301c4c48&rBranch=main&rCommit=6946094f6fe9ff77ca02736567ddf2fc031d5fd300ce7c60d832cf9c301c4c48&repoName=pytorch%2Fao&benchmarkName=&modelName=AlbertForMaskedLM&backendName=&modeName=inference&dtypeName=autoquant&deviceName=cuda%20(NVIDIA%20H100)&archName=All%20Platforms shows up the missing value
@jerryzh168 I notice that there could be multiple records of the same model in the same commit, mostly from
noquant
eager. For example,6946094f6fe9ff77ca02736567ddf2fc031d5fd300ce7c60d832cf9c301c4c48
andAlbertForMaskedLM
has 1time_ms(avg)
record forautoquant
, 1time_ms(avg)
record fornoquant
using torch.compile, but 5 of them fornoquant
eager. Any thoughts?