Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

reorder weights according to their precision #252

Closed
wants to merge 1 commit into from

Conversation

ngc92
Copy link
Contributor

@ngc92 ngc92 commented Apr 25, 2024

Simplify our logic by keeping weights of the same precision close together.

(If we want to go with this, we also need to update the fp32 network to match; hence, for now this is a Draft PR)

@karpathy
Copy link
Owner

Random thought I was already going to update how we write mixed precision tensors to file, reading inside mixed precision program. We could couple this change with that one, then we don't have to update the fp32 part of the code.

This is not just a code refactor, afaict this would speed up the code because our adam kernels run over more elements at a time?

@karpathy
Copy link
Owner

seems like we should be able to directly load the version 2 files saved from .py file now on master! will check tomorrow.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants