You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
SUPPORTS_BITS= [4] # fused eora only validated for 4 bits
57
+
SUPPORTS_BITS= [4, 8] # fused eora only validated for 4 bits
58
58
SUPPORTS_GROUP_SIZE= [-1, 16, 32, 64, 128]
59
59
SUPPORTS_DESC_ACT= [True, False]
60
60
SUPPORTS_SYM= [True] # TODO: validate False
@@ -156,32 +156,48 @@ def post_init(self):
156
156
defforward(self, x):
157
157
x_dtype=x.dtype
158
158
ifx_dtype!=torch.float16:
159
-
logger.warning_once(
159
+
log.warn.once(
160
160
f"Exllama EoRA kernel requires a float16 input activation, while {x.dtype} was passed. Casting to float16.\nMake sure you loaded your model with torch_dtype=torch.float16, that the model definition does not inadvertently cast to float32, or disable AMP Autocast that may produce float32 intermediate activations in the model."
0 commit comments