Skip to content

Create LICENSE#11

Merged
ApostaC merged 1 commit intomainfrom
license
Jul 1, 2024
Merged

Create LICENSE#11
ApostaC merged 1 commit intomainfrom
license

Conversation

@ApostaC
Copy link
Copy Markdown
Contributor

@ApostaC ApostaC commented Jul 1, 2024

No description provided.

@ApostaC ApostaC merged commit 67ca2fa into main Jul 1, 2024
@ApostaC ApostaC deleted the license branch July 1, 2024 16:06
orozery added a commit to orozery/LMCache that referenced this pull request Apr 29, 2025
This commit re-enables p2p sharing for both vllm v0 and v1.

Signed-off-by: Or Ozeri <oro@il.ibm.com>
guymguym pushed a commit to guymguym/LMCache that referenced this pull request Jun 11, 2025
Add async sending based on async V1 connector API updates
KevinCheung2259 pushed a commit to KevinCheung2259/LMCache that referenced this pull request Nov 5, 2025
sheperdh added a commit to sheperdh/LMCache2 that referenced this pull request Jan 6, 2026
* [XWKV-48] Implement gismo backend (LMCache#7)

* [XWKV-48] Cache fd to optimize performance (LMCache#11)

Cache fd to optimize performance.

Add thread pool to read/write file in parallel.

* [XWKV-48] Support remotely read kv (LMCache#12)

As Gismo supports locally read files
written remotely, we need to change our code
to make sure we can read files not cached.

* [XWKV-68] Use vram API to boost performance (LMCache#13)

As vram read/write API to boost performance.

Use new get/put method to read/write meta file.

* [XWKV-68] Use batched contains API

Use mvfs batched contains API to implement
batch contains interface to
boost performance.

Add retry when get metadata in case lmcache
reads faster than backend.

---------

Co-authored-by: Jinwen <287310886@qq.com>
DongDongJu pushed a commit to DongDongJu/LMCache that referenced this pull request Feb 22, 2026
hyunyul-XCENA added a commit to hyunyul-XCENA/LMCache that referenced this pull request Mar 18, 2026
- Add assert memory_obj.tensor is not None in batched_submit_put_task
  for parity with single-key submit_put_task
- Add summary debug logs to batched_get_blocking and
  batched_get_non_blocking for production observability
- Deduplicate batched_async_contains by delegating to batched_contains
yoo-kumaneko pushed a commit to yoo-kumaneko/LMCache that referenced this pull request Apr 8, 2026
* [MP][optimize] optimize save when mla enabled

Signed-off-by: idellzheng <idellzheng@tencent.com>

* use ParallelStrategy

Signed-off-by: idellzheng <idellzheng@tencent.com>

* rename world_sise and worker_id

Signed-off-by: idellzheng <idellzheng@tencent.com>

* optimize mla save

Signed-off-by: idellzheng <idellzheng@tencent.com>

---------

Signed-off-by: idellzheng <idellzheng@tencent.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant