Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Can you add support for apple/ml-cross-entropy? #1298

Open
zfflxx opened this issue Nov 16, 2024 · 6 comments
Open

Can you add support for apple/ml-cross-entropy? #1298

zfflxx opened this issue Nov 16, 2024 · 6 comments

Comments

@zfflxx
Copy link

zfflxx commented Nov 16, 2024

This new method saves a lot of more memory, can you port it to unsloth?
Cut Your Losses in Large-Vocabulary Language Models
https://github.com/apple/ml-cross-entropy

@iiLaurens
Copy link

I was coming here to ask the same. This seems like a perfect fit for Unsloth

@danielhanchen
Copy link
Contributor

Yes! Just tested it and it seems to work really well :)

@danielhanchen
Copy link
Contributor

Will add it probably in the next release :)

@dame-cell
Copy link

hey there @danielhanchen I've been trying to implement this in triton if it's ok and can i open a draft by tomorrow just so we can discuss and check if the code is all right

@danielhanchen
Copy link
Contributor

@dame-cell Oh I already managed to add Apple's one in :)

@shimmyshimmer
Copy link
Collaborator

We're still testing though so it's not final or near finished.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

5 participants