We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent f6eb4dc commit c3a8a16Copy full SHA for c3a8a16
fastdeploy/model_executor/layers/linear.py
@@ -977,7 +977,12 @@ def __init__(
977
def process_weights_after_loading(self):
978
if self.fd_config.load_config.dynamic_load_weight:
979
return
980
- w = self.kv_b_proj.weight.reshape(
+ w = (
981
+ self.kv_b_proj.weight.transpose([1, 0])
982
+ if self.fd_config.model_config.model_format == "torch"
983
+ else self.kv_b_proj.weight
984
+ )
985
+ w = w.reshape(
986
[
987
self.kv_lora_rank,
988
self.num_heads_per_partition,
0 commit comments