Skip to content

[Feature Request] Support for NPU #27

@zeus-cht

Description

@zeus-cht

I am encountering a compatibility issue when running this project on NPU hardware (Huawei Ascend). The project currently depends on torch_memory_saver, which seems to be heavily optimized for CUDA memory management and fragmentation reduction.

Context

  • Dependency: torch_memory_saver (likely introduced via sglang or vllm).
  • Hardware: Huawei Ascend NPU (CANN).
  • Current Behavior: torch_memory_saver specifically hooks into CUDA-related memory APIs (cudaMalloc, etc.). When running on NPU, it either fails to initializ

Question
Does the roadmap include support for non-CUDA backends like NPU?

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions