Teak mem fraction#20
Merged
merrymercy merged 2 commits intomainfrom Jan 17, 2024
Merged
Conversation
5 tasks
timethink
pushed a commit
to timethink/sglang
that referenced
this pull request
Mar 9, 2025
chunyuan-w
pushed a commit
to chunyuan-w/sglang
that referenced
this pull request
Mar 25, 2025
* Add fused biased_grouped_topk * add record function
NorthmanPKU
pushed a commit
to NorthmanPKU/sglang
that referenced
this pull request
May 16, 2025
* feat: automated pip install, but manual environmental variable * fix:small bug * feat: added requirements.txt, but encountered some other issues * fix: run_time lib configed correctly, but _cython does not exist * fix: worked? * fix: no .vscode * fix: "re-Deleted .vscode directory"
5 tasks
chunyuan-w
pushed a commit
to chunyuan-w/sglang
that referenced
this pull request
May 28, 2025
* Add fused biased_grouped_topk * add record function
yanbing-j
added a commit
to yanbing-j/sglang
that referenced
this pull request
Jun 3, 2025
* Add fused biased_grouped_topk * add record function
yanbing-j
added a commit
to yanbing-j/sglang
that referenced
this pull request
Jun 4, 2025
* Add fused biased_grouped_topk * add record function
yanbing-j
added a commit
to yanbing-j/sglang
that referenced
this pull request
Jun 10, 2025
* Add fused biased_grouped_topk * add record function
yanbing-j
added a commit
to yanbing-j/sglang
that referenced
this pull request
Jun 18, 2025
* Add fused biased_grouped_topk * add record function
5 tasks
blzheng
referenced
this pull request
in blzheng/sglang
Jun 23, 2025
yichiche
pushed a commit
to yichiche/sglang
that referenced
this pull request
Jul 30, 2025
Signed-off-by: Ivan Butygin <ivan.butygin@gmail.com>
yichiche
pushed a commit
to yichiche/sglang
that referenced
this pull request
Aug 7, 2025
Signed-off-by: Ivan Butygin <ivan.butygin@gmail.com>
yichiche
pushed a commit
to yichiche/sglang
that referenced
this pull request
Aug 11, 2025
Signed-off-by: Ivan Butygin <ivan.butygin@gmail.com>
5 tasks
5 tasks
zhuyijie88
added a commit
to zhuyijie88/sglang
that referenced
this pull request
Sep 1, 2025
zhuyijie88
pushed a commit
to zhuyijie88/sglang
that referenced
this pull request
Sep 4, 2025
* sync models and datasets every 6 hours Signed-off-by: mywaaagh_admin <pkwarcraft@gmail.com>
key4ng
pushed a commit
to key4ng/sglang
that referenced
this pull request
Nov 9, 2025
key4ng
pushed a commit
to key4ng/sglang
that referenced
this pull request
Nov 9, 2025
5 tasks
amd-youchen
pushed a commit
to amd-youchen/sglang
that referenced
this pull request
Nov 24, 2025
…sed-new [feat] enable fused qknorm and rope
CloudRipple
added a commit
to CloudRipple/sglang
that referenced
this pull request
Jan 25, 2026
…and revert necessary changes (sgl-project#20) * refactor: enhance KL divergence method in DiagonalGaussianDistribution for flexible dimension handling and clean up DAC class by removing unused code * refactor: update DacVAE architecture, configuration and its customized loader. * Revert "fix: update adjust_frames parameter to False for improved multi-GPU compatibility" * revert changes in base pipeline configs * revert changes in configs/sample/__init__.py * [Feature] Remove weight norm in DAC * [Fix] Use legacy weight norm, which can be removed * [Fix] remove weight norm at the right place * [Chore] update test script * Revert "[Fix] remove weight norm at the right place" This reverts commit 3a0accbae41650e926c5828025323a12454827a4. * Revert "[Fix] Use legacy weight norm, which can be removed" This reverts commit eb93f20f134888adba4a5124fa1d167b93d180e7. * Revert "[Feature] Remove weight norm in DAC" This reverts commit aaa64abbc25112a706bf3d3604ffeac390a1d8a8. * [Feature] Remove all weight norm from DAC modeling --------- Co-authored-by: CloudRipple <yiyangzhang25@m.fudan.edu.cn>
sywangyi
pushed a commit
to sywangyi/sglang
that referenced
this pull request
Feb 26, 2026
sywangyi
added a commit
to sywangyi/sglang
that referenced
this pull request
Feb 27, 2026
* port layernorm 3d * apply layernorm * support for bias * fix * intf fix * add support for CPU * fix tp=3/6 padding issue in encoder vision * fix tp=3/6 padding issue in qwen3-omni * refactor code * add mrope * change attention_mask shape to use flash attn * add kernel apply_rotary_pos_emb_cpu * replace nn.Linear with ReplicatedLinear * enable torch.compile * construct mask using query.dtype instead of bool on CPU * add fast path for sparse attention * fix double free segfault by wrong setting of BLOCK_M * improve extend kernel performance for long context length * update test_extend.py * update comment * fix topk softmax performance issue * port optimization for image preprocessor in Qwen2VLImageProcessorFast * apply optimization for image preprocessor * update docker file * optimize conv3d used in patch embedding * resolve conflict * apply optimized conv3d * apply optimization for flash_attn_varlen_func (sgl-project#19) * port optimization for flash_attn_varlen_func * apply flash_attn_varlen_func * remove contiguous before rope (sgl-project#20) * Revert "resolve conflict" This reverts commit 7622f6d. * fix after rebase * Update pyproject_cpu.toml * Update xeon.Dockerfile * minor fix after rebase * rope: add support for bf16 sincos (sgl-project#102) * format * Update xeon.Dockerfile * odd tp for cpu * Apply linear_gelu_linear and fix numa memory bind (sgl-project#22) * [CPU] Optimize small oc GEMM for Qwen3-next on CPU (sgl-project#12446) Co-authored-by: Zheng, Beilei <beilei.zheng@intel.com> * port linear_gelu_linear kernel * apply linear_gelu_linear for TP=1 * fix numa memory bind * apply parallel partition patch --------- Co-authored-by: jianan-gu <jianan.gu@intel.com> * Revert "Fix: test_vlm_offline_throughput output throughput (sgl-project#13279)" (sgl-project#101) This reverts commit 7ee3e36. * fix input dtype mismatch issue * apply optimized layernorm --------- Co-authored-by: Zheng, Beilei <beilei.zheng@intel.com> Co-authored-by: ZailiWang <zaili.wang@intel.com> Co-authored-by: mingfeima <mingfei.ma@intel.com> Co-authored-by: jianan-gu <jianan.gu@intel.com>
5 tasks
SCDESPERTATE
pushed a commit
to SCDESPERTATE/sglang
that referenced
this pull request
Mar 12, 2026
support Qwen3.5
alphabetc1
pushed a commit
to alphabetc1/sglang
that referenced
this pull request
Mar 14, 2026
Add Acknowledgment To the Team mates
mmangkad
pushed a commit
to mmangkad-dev/sglang
that referenced
this pull request
Apr 3, 2026
Gemma 4 audio/vision tower changes their weights from `qkv_proj.weight` to `qkv_proj.linear.weight` Gemma 4 vision pipeline have some misc changes
wisclmy0611
pushed a commit
that referenced
this pull request
Apr 7, 2026
* converted supported-models/reward-models.mdx * fixed reward models * converting rerank models
lujangus
added a commit
to tails-mpt/sglang
that referenced
this pull request
Apr 30, 2026
…e Squeeze)
Adds two server-args flags and routes them into the Eagle3 verifier path:
--speculative-verify-mode rejection_sampling | typical_acceptance
--speculative-typical-acceptance-alpha float (default 0.8)
Defaults to "rejection_sampling" so existing deployments are unchanged.
When set to "typical_acceptance", the Eagle3 verifier in eagle_info.py
(single-layer) and eagle_info_v2.py (multi-layer) routes the alpha value
into both threshold_single and threshold_acc that the existing CUDA kernel
already consumes:
// sgl-kernel/csrc/speculative/speculative_sampling.cuh:80
if (coin <= prob_acc / threshold_acc || target_prob_single >= threshold_single)
accept;
When alpha=1.0 this reduces to strict rejection sampling. When alpha<1.0
this is exactly Medusa's typical-acceptance algorithm. No CUDA changes.
Pre-flight context (Crucible umbrella):
experiments/MiniMax-M2.5/squeeze/relaxed/B1-typical-acceptance/preflight.md
Plan reference (Crucible umbrella):
docs/plans/minimax-squeeze.md §184-191 (Track B B1)
Squeeze pipeline coordination (Crucible umbrella):
CLAUDE.md rules sgl-project#20/sgl-project#21/sgl-project#22
Branch policy: this commit lives on `squeeze-relaxed` only. The umbrella
crucible repo's sglang submodule pointer continues to track origin/main of
sglang. The flag will be merged to main after the alpha-sweep bench
validates per the squeeze plan's quality floor (≤3% per-dataset drift
from Exp F lossless baseline).
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.This suggestion is invalid because no changes were made to the code.Suggestions cannot be applied while the pull request is closed.Suggestions cannot be applied while viewing a subset of changes.Only one suggestion per line can be applied in a batch.Add this suggestion to a batch that can be applied as a single commit.Applying suggestions on deleted lines is not supported.You must change the existing code in this line in order to create a valid suggestion.Outdated suggestions cannot be applied.This suggestion has been applied or marked resolved.Suggestions cannot be applied from pending reviews.Suggestions cannot be applied on multi-line comments.Suggestions cannot be applied while the pull request is queued to merge.Suggestion cannot be applied right now. Please check back later.
No description provided.