Skip to content

Conversation

@wukaixingxp
Copy link

Introduces put_slice() and get_slice() APIs to support distributed tensor storage and retrieval with explicit slice metadata. This enables efficient resharding between different parallelism strategies (e.g., FSDP to TP).

Key changes:

  • Add put_slice() API for storing tensor shards with position metadata
  • Add get_slice() API for fetching specific slices from distributed tensors
  • Add LocalClient.put_slice() implementation
  • Add test_slice_api.py with comprehensive slice API tests
  • Export slice APIs in torchstore.init

These APIs work with the existing transport system and will benefit from GPU-direct transports like CUDA IPC when available.

Developer and others added 2 commits February 6, 2026 23:59
Introduces put_slice() and get_slice() APIs to support distributed tensor
storage and retrieval with explicit slice metadata. This enables efficient
resharding between different parallelism strategies (e.g., FSDP to TP).

Key changes:
- Add put_slice() API for storing tensor shards with position metadata
- Add get_slice() API for fetching specific slices from distributed tensors
- Add LocalClient.put_slice() implementation
- Add test_slice_api.py with comprehensive slice API tests
- Export slice APIs in torchstore.__init__

These APIs work with the existing transport system and will benefit from
GPU-direct transports like CUDA IPC when available.

Co-Authored-By: Claude Sonnet 4 <noreply@anthropic.com>
@meta-cla meta-cla bot added the CLA Signed This label is managed by the Meta Open Source bot. label Feb 7, 2026
@wukaixingxp wukaixingxp marked this pull request as ready for review February 10, 2026 19:19
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

CLA Signed This label is managed by the Meta Open Source bot.

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant