You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
* Adding support for gptq models
Signed-off-by: Amit Raj <quic_amitraj@quicinc.com>
* Code cleaning and formating
Signed-off-by: Amit Raj <quic_amitraj@quicinc.com>
* ruff format and fixed some bug
Signed-off-by: Amit Raj <quic_amitraj@quicinc.com>
* Added tests for gptq
Signed-off-by: Amit Raj <quic_amitraj@quicinc.com>
* Bug-fix-1
Signed-off-by: Amit Raj <quic_amitraj@quicinc.com>
* fixed bugs-2
Signed-off-by: Amit Raj <quic_amitraj@quicinc.com>
* fixed bug-3
Signed-off-by: Amit Raj <quic_amitraj@quicinc.com>
* Added docstring
Signed-off-by: Amit Raj <quic_amitraj@quicinc.com>
* Addressed comments
Signed-off-by: Amit Raj <quic_amitraj@quicinc.com>
* Addressed comments
Signed-off-by: Amit Raj <quic_amitraj@quicinc.com>
* fixed bugs-3
Signed-off-by: Amit Raj <quic_amitraj@quicinc.com>
* ruff check and format
Signed-off-by: Amit Raj <quic_amitraj@quicinc.com>
* Addressed comments-3
Signed-off-by: Amit Raj <quic_amitraj@quicinc.com>
---------
Signed-off-by: Amit Raj <quic_amitraj@quicinc.com>
Signed-off-by: Onkar Chougule <quic_ochougul@quicinc.com>
f"in_features should be perfectly divisible by group_size, got in_features = {self.in_features}, group_size = {self.group_size} while initializing WQLinear_GEMM module"
28
30
)
29
-
ifout_features% (32//self.w_bit) !=0:
31
+
ifout_features% (32//self.bits) !=0:
30
32
raiseValueError(
31
33
f"out_features must be perfectly divisible by number of weights packed into int32 value i.e. 8, got out_features={self.out_features}"
0 commit comments