Bugfix: Change get() -> GetDLTensorPtr() in cutlass FusedMoE validations#1995
Bugfix: Change get() -> GetDLTensorPtr() in cutlass FusedMoE validations#1995yzh119 merged 2 commits intoflashinfer-ai:mainfrom
Conversation
Signed-off-by: Amir Klein <203507526+amirkl94@users.noreply.github.com>
Summary of ChangesHello @amirkl94, I'm Gemini Code Assist1! I'm currently reviewing this pull request and will post my feedback shortly. In the meantime, here's a summary to help you and other reviewers quickly get up to speed! This pull request addresses a compatibility issue arising from an Highlights
Using Gemini Code AssistThe full guide for Gemini Code Assist can be found on our documentation page, here are some quick tips. Invoking Gemini You can request assistance from Gemini at any point by creating a comment using either
Customization To customize Gemini Code Assist for GitHub experience, repository maintainers can create a configuration file and/or provide a custom code review style guide (such as PEP-8 for Python) by creating and adding files to a Limitations & Feedback Gemini Code Assist may make mistakes. Please leave feedback on any instances where its feedback is incorrect or counter productive. You can react with 👍 and 👎 on @gemini-code-assist comments. If you're interested in giving your feedback about your experience with Gemini Code Assist for Github and other Google products, sign up here. You can also get AI-powered code generation, chat, as well as code reviews directly in the IDE at no cost with the Gemini Code Assist IDE Extension. Footnotes
|
|
/bot run |
|
Note Other AI code review bot(s) detectedCodeRabbit has detected other AI code review bot(s) in this pull request and will avoid duplicating their findings in the review comments. This may lead to a less comprehensive review. WalkthroughUpdated null-pointer checks in the Cutlass FusedMoE FP8 quantization validation to use Changes
Estimated code review effort🎯 1 (Trivial) | ⏱️ ~5 minutes
Possibly related PRs
Suggested reviewers
Poem
Pre-merge checks and finishing touches❌ Failed checks (2 warnings)
✅ Passed checks (1 passed)
✨ Finishing touches
🧪 Generate unit tests (beta)
📜 Recent review detailsConfiguration used: CodeRabbit UI Review profile: CHILL Plan: Pro 📒 Files selected for processing (1)
🚧 Files skipped from review as they are similar to previous changes (1)
Comment |
|
@amirkl94 is not authorized to trigger this CI job. cc: @yzh119, @sricketts, @yongwww |
There was a problem hiding this comment.
Code Review
This pull request correctly updates the API call from .get() to .GetDLTensorPtr() for tvm::ffi::Tensor objects, which is necessary due to a version bump in the apache-tvm-ffi library. The change is straightforward and correct. I've identified a minor typo in an error message on one of the modified lines and have provided a suggestion to fix it.
Signed-off-by: Amir Klein <203507526+amirkl94@users.noreply.github.com>
There was a problem hiding this comment.
Actionable comments posted: 0
🧹 Nitpick comments (1)
csrc/fused_moe/cutlass_backend/flashinfer_cutlass_fused_moe_sm100_binding.cu (1)
829-829: Fix typo in error message.The error message contains a typo: "fc2_dequant_dequant" should be "fc2_dequant". This is a pre-existing issue, not introduced by this PR.
Apply this diff to fix the typo:
- << "Expecting fc2_dequant_dequant to be non null"; + << "Expecting fc2_dequant to be non null";
📜 Review details
Configuration used: CodeRabbit UI
Review profile: CHILL
Plan: Pro
📒 Files selected for processing (1)
csrc/fused_moe/cutlass_backend/flashinfer_cutlass_fused_moe_sm100_binding.cu(1 hunks)
⏰ Context from checks skipped due to timeout of 90000ms. You can increase the timeout in your CodeRabbit configuration to a maximum of 15 minutes (900000ms). (1)
- GitHub Check: Deploy Docs
🔇 Additional comments (1)
csrc/fused_moe/cutlass_backend/flashinfer_cutlass_fused_moe_sm100_binding.cu (1)
825-831: LGTM! API migration correctly applied.The change from
get()toGetDLTensorPtr()is consistent across all four null-pointer checks for the FP8 quantization parameters. This correctly addresses the API change after bumping the apache-tvm-ffi version.
|
/bot run |
|
I can repro the previously failing unit tests now passing with this PR on B200. Waiting for results from CI bot's pipeline |
yzh119
left a comment
There was a problem hiding this comment.
Thanks for the timely fix!
…ons (flashinfer-ai#1995) ## 📌 Description Using different API after `apach-tvm-ffi` version bump. <!-- This is an auto-generated comment: release notes by coderabbit.ai --> ## Summary by CodeRabbit * **Bug Fixes** * Improved null-pointer validation for FP8 quantization tensors used during inference, increasing robustness and reducing risk of runtime errors. <!-- end of auto-generated comment: release notes by coderabbit.ai --> --------- Signed-off-by: Amir Klein <203507526+amirkl94@users.noreply.github.com>
📌 Description
Using different API after
apach-tvm-ffiversion bump.Summary by CodeRabbit