Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Support for on-loading quantization #8

Merged
merged 22 commits into from
Jul 27, 2023

Conversation

smspillaz
Copy link
Owner

No description provided.

This can be used to pre-configure how we would like the model to be
loaded. Right now it supports a configuration for quantization. A design
principle of GGMLModelConfig is that it can be used also when the
config is NULL - this is why the config_get functions return a gboolean
indicating whether there is any change from the default in that part of
the configuration and also handle the case where the config object is NULL
…e ModelDescNode

We do this after loading the hyperparameters and starting to set up
the ModelDesc. The ModelDesc contains all the information about the types
of the tensors and since we can convert them now on the fly during loading,
this is the perfect place to edit once we know the desired quantization
configuration.
@smspillaz smspillaz force-pushed the quantization-conversion-support branch from 6659075 to bba8014 Compare July 26, 2023 21:56
@smspillaz smspillaz merged commit ea09786 into master Jul 27, 2023
1 check passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

1 participant