-
Notifications
You must be signed in to change notification settings - Fork 235
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Config Serilization No Deps #1875
Conversation
🔗 Helpful Links🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/ao/1875
Note: Links to docs will display an error until the docs builds have been completed. ✅ You can merge normally! (1 Unrelated Failure)As of commit 6f44671 with merge base be09c1d ( BROKEN TRUNK - The following job failed but were present on the merge base:👉 Rebase onto the `viable/strict` branch to avoid these failures
This comment was automatically generated by Dr. CI and updates every 15 minutes. |
|
||
|
||
@pytest.mark.parametrize("config", configs, ids=get_config_ids) | ||
def test_reconstructable_dict_file_round_trip(config): |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
looks great!
Can this be relaxed to allow prototype and sparsity configs to be serializable? Thoughts on how handling of static scales would work with this? Overall looks great! |
4d0495b
to
e06cb0a
Compare
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Looks great! Definitely prefer this to the Pydantic version and separating serde from the string API. Just a couple questions:
-
How do we want versioning to work exactly? We definitely need to bump the version when we remove or modify an existing field, but do we necessarily need to bump the version when adding a new field with a default value? In this case, should we still fail on version mismatch or just let it load the default value?
-
Do we need to handle nested configs / data structures? I feel inner configs (e.g.
Float8MMConfig
) don't need to beTorchAOBaseConfig
, so technically we just need to handle inner dataclasses/named tuples?
processed_data = {} | ||
for key, value in obj_data.items(): | ||
if isinstance(value, dict) and "_type" in value and "_data" in value: | ||
# Recursively handle nested configs |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Do we have any existing configs that have nested configs or fields with data structures? I feel our configs so far have been pretty simple?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Vasiliy mentioned there might be one for float8 training but I could find it for config wrapping configs
We do have configs that wrap objects that need to be handled specially, the float inference configs have this where MMconfig is internal to it and needs special care
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
We do have configs that wrap objects that need to be handled specially, the float inference configs have this where MMconfig is internal to it and needs special care
I don't think MMConfig should be in the BC surface / serializable / etc, that seems like a code smell.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I disagree - it is a convenient way for users to modulate mm behavior on a per instance level and unpacking this struct into three separate booleans doen't feel beneficial.
That being said there are alot of examples of these nested objects, Layouts
, granularities
, etc.. for which this behavior is necessary
0e5aab0
to
896cb53
Compare
@@ -56,6 +58,8 @@ | |||
Int8DynamicActivationInt4WeightConfig, | |||
Int8DynamicActivationInt8WeightConfig, | |||
Int8WeightOnlyConfig, | |||
PlainLayout, |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
is this exposed in public API / on HF / etc?
I don't think this abstraction is clean and I don't want to expose it, this is not really a "layout" rather than a "layout with some other arguments mixed in for convenience". We should talk about what to do with this before exposing it.
896cb53
to
6f44671
Compare
Summary
Similar to #1806 but without using Pydantic
This PR introduces functionality for serializing and deserializing
AOBaseConfig
objects to and from dictionariesThe new
config_to_dict
andconfig_from_dict
functions allow these configurations to be:Another main difference from that PR is that AOBaseConfig remains unchanged besides the overridable class var which is set to 1 by default
This makes it so that configs can opt-in, which hopefully is as cheap is adding your config to the test file, and ensuring that all types that are needed for reconstruction can be found in torchao.quantization import module
Implementation Details
Serialization Format
When an
AOBaseConfig
object is serialized, it creates a dictionary with:Special Cases
The
ConfigJSONEncoder
handles several types of objects:MappingType.SYMMETRIC
Version Management
Each config class defines a
VERSION
class variable (defaulting to 1). During deserialization, we enforce:This prevents using outdated configurations with newer code that might expect different parameters.
Restricted Reconstruction
I only allow importing types from torchao.quantization module path or from torch
Example Usage
Serializing a Configuration
This produces
Deserializing a Configuration
Prints
YAML Support
The serialized format uses basic types (strings, numbers, booleans, lists, dictionaries) that are compatible with YAML. Users can:
config_to_dict
config_from_dict
Example:
The yaml string
Future Improvements
As well we should ensure that all quant-api's configs are added to the test files