Skip to content

Improve docs & Rename Gemini -> VertexAI#19

Merged
merrymercy merged 6 commits intomainfrom
docs
Jan 17, 2024
Merged

Improve docs & Rename Gemini -> VertexAI#19
merrymercy merged 6 commits intomainfrom
docs

Conversation

@merrymercy
Copy link
Contributor

No description provided.

@merrymercy merrymercy merged commit bf51ddc into main Jan 17, 2024
@merrymercy merrymercy deleted the docs branch January 17, 2024 10:54
timethink pushed a commit to timethink/sglang that referenced this pull request Mar 9, 2025
yanbing-j pushed a commit to yanbing-j/sglang that referenced this pull request Mar 25, 2025
* support W8A8 for DeepSeek-R1

* add assertion on cpu_has_amx_support()

* refactor device check in MoE
pi314ever pushed a commit to pi314ever/sglang that referenced this pull request May 16, 2025
chunyuan-w added a commit to chunyuan-w/sglang that referenced this pull request May 28, 2025
* support W8A8 for DeepSeek-R1

* add assertion on cpu_has_amx_support()

* refactor device check in MoE
chunyuan-w added a commit to chunyuan-w/sglang that referenced this pull request May 28, 2025
* support W8A8 for DeepSeek-R1

* add assertion on cpu_has_amx_support()

* refactor device check in MoE
chunyuan-w added a commit to chunyuan-w/sglang that referenced this pull request Jun 6, 2025
* support W8A8 for DeepSeek-R1

* add assertion on cpu_has_amx_support()

* refactor device check in MoE
pengxin99 pushed a commit to pengxin99/sglang that referenced this pull request Jun 19, 2025
Support V2 V3 R1 style MLA decode
chunyuan-w added a commit to chunyuan-w/sglang that referenced this pull request Jun 25, 2025
* support W8A8 for DeepSeek-R1

* add assertion on cpu_has_amx_support()

* refactor device check in MoE
key4ng added a commit to key4ng/sglang that referenced this pull request Nov 9, 2025
Co-authored-by: Robert Ru <key4ng@Roberts-Mac-Studio.local>
wzrf pushed a commit to wzrf/sglang-fusionrag that referenced this pull request Feb 8, 2026
sywangyi pushed a commit to sywangyi/sglang that referenced this pull request Feb 26, 2026
* port optimization for flash_attn_varlen_func

* apply flash_attn_varlen_func
sywangyi added a commit to sywangyi/sglang that referenced this pull request Feb 27, 2026
* port layernorm 3d

* apply layernorm

* support for bias

* fix

* intf fix

* add support for CPU

* fix tp=3/6 padding issue in encoder vision

* fix tp=3/6 padding issue in qwen3-omni

* refactor code

* add mrope

* change attention_mask shape to use flash attn

* add kernel apply_rotary_pos_emb_cpu

* replace nn.Linear with ReplicatedLinear

* enable torch.compile

* construct mask using query.dtype instead of bool on CPU

* add fast path for sparse attention

* fix double free segfault by wrong setting of BLOCK_M

* improve extend kernel performance for long context length

* update test_extend.py

* update comment

* fix topk softmax performance issue

* port optimization for image preprocessor in Qwen2VLImageProcessorFast

* apply optimization for image preprocessor

* update docker file

* optimize conv3d used in patch embedding

* resolve conflict

* apply optimized conv3d

* apply optimization for flash_attn_varlen_func (sgl-project#19)

* port optimization for flash_attn_varlen_func

* apply flash_attn_varlen_func

* remove contiguous before rope (sgl-project#20)

* Revert "resolve conflict"

This reverts commit 7622f6d.

* fix after rebase

* Update pyproject_cpu.toml

* Update xeon.Dockerfile

* minor fix after rebase

* rope: add support for bf16 sincos (sgl-project#102)

* format

* Update xeon.Dockerfile

* odd tp for cpu

* Apply linear_gelu_linear and fix numa memory bind (sgl-project#22)

* [CPU]  Optimize small oc GEMM for Qwen3-next on CPU (sgl-project#12446)

Co-authored-by: Zheng, Beilei <beilei.zheng@intel.com>

* port linear_gelu_linear kernel

* apply linear_gelu_linear for TP=1

* fix numa memory bind

* apply parallel partition patch

---------

Co-authored-by: jianan-gu <jianan.gu@intel.com>

* Revert "Fix: test_vlm_offline_throughput output throughput (sgl-project#13279)" (sgl-project#101)

This reverts commit 7ee3e36.

* fix input dtype mismatch issue

* apply optimized layernorm

---------

Co-authored-by: Zheng, Beilei <beilei.zheng@intel.com>
Co-authored-by: ZailiWang <zaili.wang@intel.com>
Co-authored-by: mingfeima <mingfei.ma@intel.com>
Co-authored-by: jianan-gu <jianan.gu@intel.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant