-
-
Notifications
You must be signed in to change notification settings - Fork 44
Clarify Kubernetes version requirement and fallback plan in Key Features #380
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
| - **Accelerator Fungibility**: llmaz supports serving the same LLM with various accelerators to optimize cost and performance. | ||
| - **Various Model Providers**: llmaz supports a wide range of model providers, such as [HuggingFace](https://huggingface.co/), [ModelScope](https://www.modelscope.cn), ObjectStores. llmaz will automatically handle the model loading, requiring no effort from users. | ||
| - **Multi-Host Support**: llmaz supports both single-host and multi-host scenarios with [LWS](https://github.com/kubernetes-sigs/lws) from day 0. | ||
| - **Multi-Host Support**: llmaz supports both single-host and multi-host scenarios with [LWS](https://github.com/kubernetes-sigs/lws) from day 0. **Important**: LWS requires Kubernetes version **v1.26 or higher**. If you are using a lower Kubernetes version and most of your workloads rely on single-node inference, we may consider replacing LWS with a deployment-based approach. This fallback plan would involve using Kubernetes Deployments to manage single-node inference workloads efficiently. See [#32](https://github.com/InftyAI/llmaz/issues/32) for more details and updates. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Could you update this in https://github.com/InftyAI/llmaz/blob/main/docs/installation.md#prerequisites? Thanks
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Sure. I will complete it and make a commit.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Please revert this change, the note within installation.md is enough. Thanks
|
/kind documentation |
|
I've updated the Installation file. Can you please review it? |
| - **Accelerator Fungibility**: llmaz supports serving the same LLM with various accelerators to optimize cost and performance. | ||
| - **Various Model Providers**: llmaz supports a wide range of model providers, such as [HuggingFace](https://huggingface.co/), [ModelScope](https://www.modelscope.cn), ObjectStores. llmaz will automatically handle the model loading, requiring no effort from users. | ||
| - **Multi-Host Support**: llmaz supports both single-host and multi-host scenarios with [LWS](https://github.com/kubernetes-sigs/lws) from day 0. | ||
| - **Multi-Host Support**: llmaz supports both single-host and multi-host scenarios with [LWS](https://github.com/kubernetes-sigs/lws) from day 0. **Important**: LWS requires Kubernetes version **v1.26 or higher**. If you are using a lower Kubernetes version and most of your workloads rely on single-node inference, we may consider replacing LWS with a deployment-based approach. This fallback plan would involve using Kubernetes Deployments to manage single-node inference workloads efficiently. See [#32](https://github.com/InftyAI/llmaz/issues/32) for more details and updates. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Please revert this change, the note within installation.md is enough. Thanks
|
/lgtm Thanks for your patience. Welcome onboard! |
|
Thank you! I appreciate the review and approval. Excited to contribute further! |
|
/lgtm |
|
seems we need to rebase. |
Co-authored-by: Kante Yin <kerthcet@gmail.com>
971cf5e to
15fee24
Compare
|
Are there any final changes required? Please clarify, and I will make the necessary updates. |
|
/lgtm |
…res section
What this PR does / why we need it
This PR updates the README file to improve clarity on multiple host support and installation guidelines by specifying Kubernetes version requirements, helping to prevent setup issues.
Which issue(s) this PR fixes
Fixes #
#379
Special notes for your reviewer
Does this PR introduce a user-facing change?