Skip to content

Conversation

@HIT-cwh
Copy link
Collaborator

@HIT-cwh HIT-cwh commented Apr 24, 2024

No description provided.

HIT-cwh added 2 commits April 24, 2024 15:00
…uld set attn_implementation to flash_attention_2 or do not set this attribute.
@HIT-cwh HIT-cwh changed the title [Fix] Do not set attn_implementation to flash_attention_2 or sdpa if users already set it Apr 24, 2024
@pppppM pppppM merged commit 60e0cc9 into InternLM:main Apr 25, 2024
llkn-2 pushed a commit to llkn-2/xtuner that referenced this pull request Jul 31, 2024
…users already set it in XTuner configs. (InternLM#609)

* do not set attn_implementation to flash_attention_2 or sdpa if users already set it

* check cfg: If we want to use varlen attn or sequence parallel, we should set attn_implementation to flash_attention_2 or do not set this attribute.
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

2 participants