Skip to content

Overriding the new default of 'sdpa'#108

Open
ParthaEth wants to merge 20 commits intoMeiGen-AI:mainfrom
ParthaEth:main
Open

Overriding the new default of 'sdpa'#108
ParthaEth wants to merge 20 commits intoMeiGen-AI:mainfrom
ParthaEth:main

Conversation

@ParthaEth
Copy link

The default value of the attn_implementation argument in Wav2Vec2Model.from_pretrained() depends on the underlying version of Transformers (Hugging Face) you’re using.

As of Transformers v4.38+, the attn_implementation argument can be one of the following:
"eager" — standard PyTorch attention (default prior to v4.37)
"flash_attention_2" — uses FlashAttention v2 if available (faster and more memory-efficient)
"sdpa" — uses PyTorch’s scaled dot-product attention if available

I have not tested with - flash_attention_2. It would be great if that can speed things up.

VRWarg and others added 20 commits July 23, 2025 10:24
…ention (not currently used, but may be soon)
- Updated avatar asset selection to find configuration JSON and image files in a predictable manner.
- Improved input payload building by extracting required fields from avatar config and ensuring necessary data is present.
- Introduced a new base TTS template JSON file for input payload construction.
- Implemented avatar image downloading from a presigned S3 URL in the CLI.
- Updated input payload building to utilize the new template and handle avatar images.
- Adjusted video generation time constant for improved performance.
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants