Skip to content

Conversation

@EzraReiss
Copy link

@EzraReiss EzraReiss commented Sep 16, 2025

Description

These folders contain the work Juhyoung and Ezra worked on over the summer. We designed an optimized MHA kernel utilizing allo scheduling primitives.

Problems

To add the llm work we've done

Proposed Solutions

Created optimized attention kernel for FPGA architectures

Examples

Its just allo examples that can be used in the future or integrated with other transformer architecture kernels

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant