Avoids requiring xformers installation by shimming xformers.ops.memory_efficient_attention with torch.nn.functional.scaled_dot_product_attention when xformers is not available. Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
11 KiB
11 KiB