Skip to content

Conversation

@yuanlehome
Copy link
Collaborator

@yuanlehome yuanlehome commented Jul 3, 2025

  1. Currently, it is only supported for use in PD, by export FD_ATTENTION_BACKEND=FLASH_ATTN
  2. Currently, We need skip profile, by specifying --num-gpu-blocks-override
  3. By the way, delete the use_fast_ffn parameter

@paddle-bot
Copy link

paddle-bot bot commented Jul 3, 2025

Thanks for your contribution!

@yuanlehome yuanlehome force-pushed the add_fa3_backend_v2 branch from 0f83e81 to f8157ef Compare July 3, 2025 12:14
Copy link
Collaborator

@EmmonsCurse EmmonsCurse left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM

@yuanlehome yuanlehome merged commit 240bdac into PaddlePaddle:develop Jul 3, 2025
2 of 3 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants