Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Enable flex attention support #2255

Closed
wants to merge 76 commits into from
Closed

Enable flex attention support #2255

wants to merge 76 commits into from

Conversation

bursteratom
Copy link
Collaborator

Enable flex attention support

bursteratom and others added 15 commits January 5, 2025 23:27
* bumping version, removing non-osx compatible deps

* updating pylintrc

* fixing linters

* reverting changes
* add support for merge groups

* also lint merge groups
* add hf cache caching for GHA

* use modal volume to cache hf data

* make sure to update the cache as we add new fixtures in conftest
* bump axolotl contribs for upstream main conflicts:

* bump datasets, tokenizer, trl

* remove log workarounds in trl

* bump lm-eval

* remove unsloth_ import from critical path

* remove llama fa2 from conftest

* unsloth breaks with latest upstream
@bursteratom bursteratom marked this pull request as draft January 13, 2025 14:43
@bursteratom bursteratom marked this pull request as ready for review January 13, 2025 15:27
@bursteratom bursteratom marked this pull request as draft January 16, 2025 21:41
@bursteratom bursteratom marked this pull request as ready for review February 12, 2025 04:29
@bursteratom bursteratom marked this pull request as draft February 12, 2025 04:34
@bursteratom bursteratom marked this pull request as ready for review February 12, 2025 04:35
@bursteratom bursteratom marked this pull request as draft February 12, 2025 04:36
@bursteratom
Copy link
Collaborator Author

Closing in favour of #2363

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

5 participants