Skip to content

Commit

Permalink
add version to quant config
Browse files Browse the repository at this point in the history
  • Loading branch information
horheynm committed Oct 30, 2024
1 parent 8b89c15 commit ec849a9
Showing 1 changed file with 4 additions and 2 deletions.
6 changes: 4 additions & 2 deletions src/compressed_tensors/quantization/quant_config.py
Original file line number Diff line number Diff line change
Expand Up @@ -132,9 +132,10 @@ class QuantizationConfig(BaseModel):
`k_proj` and `v_proj` in their names. If this is not the case
and kv_cache_scheme != None, the quantization of kv cache will fail
:global_compression_ratio: optional informational config to report the model
compression ratio acheived by the quantization config
compression ratio acheived by the quantization config
:ignore: optional list of layers to ignore from config_groups. Layers in this list
are not quantized even if they match up with a target in config_groups
are not quantized even if they match up with a target in config_groups
:version param: version of the compressors used to optimize the given model
"""

config_groups: Dict[str, Union[QuantizationScheme, List[str]]]
Expand All @@ -144,6 +145,7 @@ class QuantizationConfig(BaseModel):
quantization_status: QuantizationStatus = QuantizationStatus.INITIALIZED
global_compression_ratio: Optional[float] = None
ignore: Optional[List[str]] = Field(default_factory=list)
version: str = "0.7.1"

def model_post_init(self, __context):
"""
Expand Down

0 comments on commit ec849a9

Please sign in to comment.