Skip to content

Support torch.distributed scatter/gather natively #9069

@bfolie

Description

@bfolie

🚀 Feature

Support torch.distributed.scatter and torch.distributed.gather using the XLA backend. They are currently not implemented.

Motivation

(Copied from a discussion thread) The fastsafetensors library uses torch.distributed to distribute/shuffle weights across GPU devices on the same host when using tensor parallelism. This loader library was introduced into vLLM last month. I'm trying to understand if the same approach can be used to speed up model loading in vLLM, in conjunction with XLA caching

Metadata

Metadata

Assignees

No one assigned

    Labels

    distributedSPMD and other distributed things.enhancementNew feature or request

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions