Skip to content

[MP] Update the MP docs and pass telemetry config into http_server#2730

Merged
ApostaC merged 2 commits intoLMCache:devfrom
ApostaC:local-dev/mp-update-doc-start-command
Mar 10, 2026
Merged

[MP] Update the MP docs and pass telemetry config into http_server#2730
ApostaC merged 2 commits intoLMCache:devfrom
ApostaC:local-dev/mp-update-doc-start-command

Conversation

@ApostaC
Copy link
Copy Markdown
Contributor

@ApostaC ApostaC commented Mar 10, 2026

What this PR does / why we need it:

Documents the --engine-type CLI argument introduced in the previous PR (#2722).

Adds the telemetry config parsing (introduced in #2696 ) to http server

Special notes for your reviewers:

Docs-only change.

If applicable:

  • this PR contains user facing changes - docs added
  • this PR contains unit tests

Signed-off-by: ApostaC <yihua98@uchicago.edu>
@gemini-code-assist
Copy link
Copy Markdown
Contributor

Summary of Changes

Hello, I'm Gemini Code Assist1! I'm currently reviewing this pull request and will post my feedback shortly. In the meantime, here's a summary to help you and other reviewers quickly get up to speed!

This pull request focuses on updating the documentation for the multiprocessing server to incorporate the recently added --engine-type command-line argument. The changes ensure that users have clear guidance on configuring different cache engine backends, particularly how to enable BlendEngineV2 through the http_server entry point, and correctly identifies the blend_server as a legacy component.

Highlights

  • Documentation Update: The --engine-type CLI argument, previously introduced, has been fully documented across relevant sections.
  • Configuration Table: The configuration.rst file now includes --engine-type in the MP Server config table, detailing its choices and usage.
  • Server Variants: The index.rst file has been updated to clarify the use of --engine-type blend with http_server and marks the standalone blend_server as a legacy option.
  • Full Example: The full example in configuration.rst was updated to use http_server and explicitly include --engine-type default.

🧠 New Feature in Public Preview: You can now enable Memory to help Gemini Code Assist learn from your team's feedback. This makes future code reviews more consistent and personalized to your project's style. Click here to enable Memory in your admin console.

Changelog
  • docs/source/mp/configuration.rst
    • Added the --engine-type argument with its description and choices to the MP Server configuration table.
    • Updated the 'Full Example' code block to use http_server instead of server and included the --engine-type default argument.
  • docs/source/mp/index.rst
    • Modified the description for python3 -m lmcache.v1.multiprocess.server to specify MPCacheEngine.
    • Added a note to python3 -m lmcache.v1.multiprocess.http_server explaining how to use --engine-type blend for BlendEngineV2.
    • Updated the description for python3 -m lmcache.v1.multiprocess.blend_server to mark it as a legacy server and recommend http_server --engine-type blend.
Activity
  • No human activity has been recorded for this pull request yet.
Using Gemini Code Assist

The full guide for Gemini Code Assist can be found on our documentation page, here are some quick tips.

Invoking Gemini

You can request assistance from Gemini at any point by creating a comment using either /gemini <command> or @gemini-code-assist <command>. Below is a summary of the supported commands on the current page.

Feature Command Description
Code Review /gemini review Performs a code review for the current pull request in its current state.
Pull Request Summary /gemini summary Provides a summary of the current pull request in its current state.
Comment @gemini-code-assist Responds in comments when explicitly tagged, both in pull request comments and review comments.
Help /gemini help Displays a list of available commands.

Customization

To customize Gemini Code Assist for GitHub experience, repository maintainers can create a configuration file and/or provide a custom code review style guide (such as PEP-8 for Python) by creating and adding files to a .gemini/ folder in the base of the repository. Detailed instructions can be found here.

Limitations & Feedback

Gemini Code Assist may make mistakes. Please leave feedback on any instances where its feedback is incorrect or counter productive. You can react with 👍 and 👎 on @gemini-code-assist comments. If you're interested in giving your feedback about your experience with Gemini Code Assist for GitHub and other Google products, sign up here.

You can also get AI-powered code generation, chat, as well as code reviews directly in the IDE at no cost with the Gemini Code Assist IDE Extension.

Footnotes

  1. Review the Privacy Notices, Generative AI Prohibited Use Policy, Terms of Service, and learn how to configure Gemini Code Assist in GitHub here. Gemini can make mistakes, so double check it and use code with caution.

Copy link
Copy Markdown
Contributor

@gemini-code-assist gemini-code-assist Bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

The pull request updates the lmcache.v1.multiprocess documentation to introduce the --engine-type configuration option, allowing users to select between MPCacheEngine and BlendEngineV2 for the cache backend. This change clarifies that the http_server can now support both engine types, rendering the standalone blend_server legacy. The documentation examples are updated to use http_server with the new --engine-type flag. A review comment suggests an improvement to the configuration.rst example, recommending the inclusion of the --http-port argument for the http_server to make the example more complete.

python3 -m lmcache.v1.multiprocess.server \
python3 -m lmcache.v1.multiprocess.http_server \
--host 0.0.0.0 \
--port 6555 \
Copy link
Copy Markdown
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

medium

Since the example now uses the http_server entry point, it would be more complete and informative to include its specific configuration arguments, like --http-port. This would clarify how to configure the HTTP frontend that this server variant provides.

Suggested change
--port 6555 \
--port 6555 \
--http-port 8000 \

@ApostaC ApostaC changed the title [Chore][MP] Update the docs for newly introduced --engine-type arg [MP] Update the MP docs and pass telemetry config into http_server Mar 10, 2026
Signed-off-by: ApostaC <yihua98@uchicago.edu>
@ApostaC ApostaC added the full Run comprehensive tests on this PR label Mar 10, 2026
Copy link
Copy Markdown
Contributor

@sammshen sammshen left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM

Copy link
Copy Markdown
Contributor

@KuntaiDu KuntaiDu left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM

@KuntaiDu KuntaiDu enabled auto-merge (squash) March 10, 2026 22:21
@ApostaC ApostaC disabled auto-merge March 10, 2026 22:51
@github-actions github-actions Bot removed the full Run comprehensive tests on this PR label Mar 10, 2026
@ApostaC ApostaC merged commit a13ad66 into LMCache:dev Mar 10, 2026
29 of 30 checks passed
shaoxiawjc pushed a commit to shaoxiawjc/LMCache that referenced this pull request Mar 11, 2026
…MCache#2730)

* Update docs for engine type

* [add] telemetry config pass to http server and blend server

Signed-off-by: ApostaC <yihua98@uchicago.edu>
Signed-off-by: shaoxiawjc <wjc2800@163.com>
realAaronWu pushed a commit to realAaronWu/LMCache that referenced this pull request Mar 20, 2026
…MCache#2730)

* Update docs for engine type

* [add] telemetry config pass to http server and blend server

Signed-off-by: ApostaC <yihua98@uchicago.edu>
Signed-off-by: Aaron Wu <aaron.wu@dell.com>
jooho-XCENA pushed a commit to xcena-dev/LMCache that referenced this pull request Apr 2, 2026
…MCache#2730)

* Update docs for engine type

* [add] telemetry config pass to http server and blend server

Signed-off-by: ApostaC <yihua98@uchicago.edu>
jooho-XCENA pushed a commit to xcena-dev/LMCache that referenced this pull request Apr 2, 2026
…MCache#2730)

* Update docs for engine type

* [add] telemetry config pass to http server and blend server

Signed-off-by: ApostaC <yihua98@uchicago.edu>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants