Skip to content

Conversation

@CyCle1024
Copy link
Collaborator

Motivation

Some feature are not currently compatitable with npu device. This would lead to errors in RL training.

Key Changes

  1. Add ACCELERATOR and ACCELERATOR_PER_NODE parameter in examples/v1/scripts/run_rl.sh and examples/v1/scripts/run_rl_submit.sh to compatiable with npu device.
  2. Make xtuner.v1.model.base.BaseModel._resolve_compile_cfg return empty dict when using npu device, this imples we shouldn't use torch.compile on npu device for now.
@CyCle1024 CyCle1024 requested a review from YanhuiDua December 22, 2025 13:58
Copy link
Collaborator

@YanhuiDua YanhuiDua left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM

@CyCle1024 CyCle1024 merged commit b095dad into InternLM:main Dec 29, 2025
4 of 7 checks passed
@CyCle1024 CyCle1024 deleted the fix_npu_resources branch December 29, 2025 03:40
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

2 participants