Skip to content

Fix for T4 GPUs#16

Merged
Ying1123 merged 4 commits intomainfrom
fix
Jan 16, 2024
Merged

Fix for T4 GPUs#16
Ying1123 merged 4 commits intomainfrom
fix

Conversation

@Ying1123
Copy link
Copy Markdown
Contributor

No description provided.

@Ying1123 Ying1123 merged commit ffe4aae into main Jan 16, 2024
@Ying1123 Ying1123 deleted the fix branch January 16, 2024 23:49
@Ying1123 Ying1123 mentioned this pull request Jan 17, 2024
timethink pushed a commit to timethink/sglang that referenced this pull request Mar 9, 2025
Co-authored-by: Lianmin Zheng <lianminzheng@gmail.com>
yanbing-j pushed a commit to yanbing-j/sglang that referenced this pull request Mar 19, 2025
pi314ever pushed a commit to pi314ever/sglang that referenced this pull request Apr 23, 2025
chunyuan-w added a commit to chunyuan-w/sglang that referenced this pull request May 27, 2025
chunyuan-w added a commit to chunyuan-w/sglang that referenced this pull request May 28, 2025
chunyuan-w added a commit to chunyuan-w/sglang that referenced this pull request May 28, 2025
chunyuan-w added a commit to chunyuan-w/sglang that referenced this pull request Jun 3, 2025
chunyuan-w added a commit to chunyuan-w/sglang that referenced this pull request Jun 6, 2025
pengxin99 pushed a commit to pengxin99/sglang that referenced this pull request Jun 19, 2025
yichiche pushed a commit to yichiche/sglang that referenced this pull request Jul 30, 2025
yichiche pushed a commit to yichiche/sglang that referenced this pull request Aug 7, 2025
yichiche pushed a commit to yichiche/sglang that referenced this pull request Aug 11, 2025
amd-youchen pushed a commit to amd-youchen/sglang that referenced this pull request Nov 18, 2025
[Feature] Accelerate VisionAttention by precompute H2D part in every …
apinge pushed a commit to apinge/sglang that referenced this pull request Nov 18, 2025
[Feature] Accelerate VisionAttention by precompute H2D part in every …
nithinsubbiah pushed a commit to nithinsubbiah/sglang that referenced this pull request Nov 21, 2025
Signed-off-by: Stanley Winata <stanley.winata@amd.com>

[Wave] Add wave extend attention kernel

Signed-off-by: Harsh Menon <harsh@nod-labs.com>

[Wave] Adding logit_cap and layer scaling to API

Also add support for the wave backend to the model
runner. And use Triton decode kernels for now.

[Wave] Run chunked prefill for perf comparison on Wave test

Need to rename the non chunked/regular prefill version because otherwise
rpd will treat it as the same kernel

Signed-off-by: Stanley Winata <stanley.winata@amd.com>

[Wave] Cache the function that loads the wave kernel

Also maintain a global kernel hash to avoid
recomputing the hash on every call.

[Wave] Don't specify block size and enable buffer ops

[Wave] Enable wave runtime and update scheduling API

[Wave] Update API to use wave_compile & WaveCompileOptions

[Wave] Update wave backend and extend attention to latest

[Wave] Add speculative decode kernel

Signed-off-by: nithinsubbiah <nithinsubbiah@gmail.com>

cache kernels using lru_cache

Update WaveBackend to use Wave Decode  (sgl-project#6)

Signed-off-by: Ivan Butygin <ivan.butygin@gmail.com>

Revert "Update WaveBackend to use Wave Decode  (sgl-project#6)" (sgl-project#7)

This reverts commit eac4599.

Wave Backend decode (sgl-project#8)

* align shapes

Signed-off-by: Ivan Butygin <ivan.butygin@gmail.com>

* fix

Signed-off-by: Ivan Butygin <ivan.butygin@gmail.com>

---------

Signed-off-by: Ivan Butygin <ivan.butygin@gmail.com>

Wave backend fixes (sgl-project#10)

Signed-off-by: Ivan Butygin <ivan.butygin@gmail.com>

More fixes to Wave decode (sgl-project#12)

Signed-off-by: Ivan Butygin <ivan.butygin@gmail.com>

is_causal

Signed-off-by: Ivan Butygin <ivan.butygin@gmail.com>

Enable the grok in3 model (sgl-project#14)

Set unique cache dir for each worker (sgl-project#16)

update kernel (sgl-project#18)

Signed-off-by: Ivan Butygin <ivan.butygin@gmail.com>

updated spec decode test as per wave

Signed-off-by: xintin <gaurav.verma@amd.com>

fix extend (sgl-project#23)

Signed-off-by: Ivan Butygin <ivan.butygin@gmail.com>

Refactor paged decode intermediate arrays shapes (sgl-project#24)

Signed-off-by: Ivan Butygin <ivan.butygin@gmail.com>

remove dyn symbols (sgl-project#26)

Signed-off-by: Ivan Butygin <ivan.butygin@gmail.com>

cleanup shapes (sgl-project#27)

Some fields were removed from `paged_decode_attention_shape`.

Signed-off-by: Ivan Butygin <ivan.butygin@gmail.com>

Remove `mha` param from Wave decode attention kernel (sgl-project#28)

Depends on iree-org/iree-turbine#1039

Signed-off-by: Paul Zhang <paul.zhang@amd.com>

nfc: fix problems reported by linting

update references from iree.turbine to wave_lang
Garrybest pushed a commit to Garrybest/sglang that referenced this pull request Jan 9, 2026
…ate_sgl-jax_discussions

bugfix: use sgl-jax discussions address
wzrf pushed a commit to wzrf/sglang-fusionrag that referenced this pull request Feb 8, 2026
alphabetc1 pushed a commit to alphabetc1/sglang that referenced this pull request Mar 14, 2026
docs: polish sglang launch & add python request examples
KHAEntertainment referenced this pull request in Clarit-AI/Engram Mar 30, 2026
5 adverse-condition tests (4/5 PASS): client disconnect, SIGKILL mid-inference
(with startup preload verification), SIGKILL during write, SIGTERM, abort+save.
New bug #16: SIGTERM graceful shutdown hangs >60s.

Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
wisclmy0611 pushed a commit that referenced this pull request Apr 7, 2026
* feat: added classification-model.mdx

* fix: modified the description
prakashkagitha added a commit to prakashkagitha/sglang that referenced this pull request May 9, 2026
Replace MTP/EAGLE speculative decoding benchmarks with standard baseline
results using plain sglang serve (no speculative flags), per sgl-cookbook
issue sgl-project#16. All 4 variants (35B-A3B FP8/BF16 and 27B FP8/BF16) benchmarked
on 1× H100 NVL across Chat (1K/1K), Reasoning (1K/8K), and Summarization
(8K/1K) scenarios at concurrency 1, 16, and 64/100.

Co-Authored-By: Claude Sonnet 4.6 <noreply@anthropic.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants