Skip to content

[FT] better support for model loading args in transformers #752

@NathanHB

Description

@NathanHB

Hi, i need to evaluate a model with SWA so i think i need to change the attention from spda => flash_attention or others since i get

Sliding Window Attention is enabled but not implemented for `sdpa`; unexpected results may be encountered.

Metadata

Metadata

Assignees

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions