Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add Batch=32 / paged attention / and new RoPE module support to all Llama3 demo and tests #15327

Draft
wants to merge 24 commits into
base: main
Choose a base branch
from

Conversation

mtairum
Copy link
Contributor

@mtairum mtairum commented Nov 21, 2024

Ticket

#11997

Problem description

TODO

What's changed

TODO

Checklist

TODO

mtairum and others added 22 commits November 26, 2024 10:04
…and batch sizes that don't lead to a rectangular core grid. TODO: confirm if tracing works.
…mode. TODO: Debug inconsistent outputs of batch 1 vs batch 16/32
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants