Skip to content

Conversation

@SanjanShiv
Copy link
Contributor

@SanjanShiv SanjanShiv commented Apr 25, 2025

…res section

What this PR does / why we need it

This PR updates the README file to improve clarity on multiple host support and installation guidelines by specifying Kubernetes version requirements, helping to prevent setup issues.

Which issue(s) this PR fixes

Fixes #
#379

Special notes for your reviewer

Does this PR introduce a user-facing change?

Kubernetes version >= 1.26

@InftyAI-Agent InftyAI-Agent added needs-triage Indicates an issue or PR lacks a label and requires one. needs-priority Indicates a PR lacks a label and requires one. do-not-merge/needs-kind Indicates a PR lacks a label and requires one. labels Apr 25, 2025
@InftyAI-Agent InftyAI-Agent requested a review from kerthcet April 25, 2025 11:10
- **Accelerator Fungibility**: llmaz supports serving the same LLM with various accelerators to optimize cost and performance.
- **Various Model Providers**: llmaz supports a wide range of model providers, such as [HuggingFace](https://huggingface.co/), [ModelScope](https://www.modelscope.cn), ObjectStores. llmaz will automatically handle the model loading, requiring no effort from users.
- **Multi-Host Support**: llmaz supports both single-host and multi-host scenarios with [LWS](https://github.com/kubernetes-sigs/lws) from day 0.
- **Multi-Host Support**: llmaz supports both single-host and multi-host scenarios with [LWS](https://github.com/kubernetes-sigs/lws) from day 0. **Important**: LWS requires Kubernetes version **v1.26 or higher**. If you are using a lower Kubernetes version and most of your workloads rely on single-node inference, we may consider replacing LWS with a deployment-based approach. This fallback plan would involve using Kubernetes Deployments to manage single-node inference workloads efficiently. See [#32](https://github.com/InftyAI/llmaz/issues/32) for more details and updates.
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Sure. I will complete it and make a commit.

Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Please revert this change, the note within installation.md is enough. Thanks

@kerthcet
Copy link
Member

/kind documentation

@InftyAI-Agent InftyAI-Agent added documentation Categorizes issue or PR as related to documentation. and removed do-not-merge/needs-kind Indicates a PR lacks a label and requires one. labels Apr 25, 2025
@SanjanShiv
Copy link
Contributor Author

I've updated the Installation file. Can you please review it?

- **Accelerator Fungibility**: llmaz supports serving the same LLM with various accelerators to optimize cost and performance.
- **Various Model Providers**: llmaz supports a wide range of model providers, such as [HuggingFace](https://huggingface.co/), [ModelScope](https://www.modelscope.cn), ObjectStores. llmaz will automatically handle the model loading, requiring no effort from users.
- **Multi-Host Support**: llmaz supports both single-host and multi-host scenarios with [LWS](https://github.com/kubernetes-sigs/lws) from day 0.
- **Multi-Host Support**: llmaz supports both single-host and multi-host scenarios with [LWS](https://github.com/kubernetes-sigs/lws) from day 0. **Important**: LWS requires Kubernetes version **v1.26 or higher**. If you are using a lower Kubernetes version and most of your workloads rely on single-node inference, we may consider replacing LWS with a deployment-based approach. This fallback plan would involve using Kubernetes Deployments to manage single-node inference workloads efficiently. See [#32](https://github.com/InftyAI/llmaz/issues/32) for more details and updates.
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Please revert this change, the note within installation.md is enough. Thanks

@kerthcet
Copy link
Member

/lgtm
/approve

Thanks for your patience. Welcome onboard!

@InftyAI-Agent InftyAI-Agent added lgtm Looks good to me, indicates that a PR is ready to be merged. approved Indicates a PR has been approved by an approver from all required OWNERS files. labels Apr 26, 2025
@InftyAI-Agent InftyAI-Agent removed the lgtm Looks good to me, indicates that a PR is ready to be merged. label Apr 26, 2025
@SanjanShiv
Copy link
Contributor Author

Thank you! I appreciate the review and approval. Excited to contribute further!

@kerthcet
Copy link
Member

/lgtm

@InftyAI-Agent InftyAI-Agent added the lgtm Looks good to me, indicates that a PR is ready to be merged. label Apr 26, 2025
@kerthcet
Copy link
Member

seems we need to rebase.

@InftyAI-Agent InftyAI-Agent removed the lgtm Looks good to me, indicates that a PR is ready to be merged. label Apr 26, 2025
@SanjanShiv
Copy link
Contributor Author

Are there any final changes required? Please clarify, and I will make the necessary updates.

@kerthcet
Copy link
Member

/lgtm

@InftyAI-Agent InftyAI-Agent added the lgtm Looks good to me, indicates that a PR is ready to be merged. label Apr 26, 2025
@InftyAI-Agent InftyAI-Agent merged commit fc5734f into InftyAI:main Apr 26, 2025
19 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

approved Indicates a PR has been approved by an approver from all required OWNERS files. documentation Categorizes issue or PR as related to documentation. lgtm Looks good to me, indicates that a PR is ready to be merged. needs-priority Indicates a PR lacks a label and requires one. needs-triage Indicates an issue or PR lacks a label and requires one.

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants