@@ -224,7 +224,7 @@ def __init__(
224
224
rope_freq_base : float = 10000.0 ,
225
225
rope_freq_scale : float = 1.0 ,
226
226
n_gqa : Optional [int ] = None , # (TEMPORARY) must be 8 for llama2 70b
227
- rms_eps_norm : Optional [float ] = None , # (TEMPORARY)
227
+ rms_norm_eps : Optional [float ] = None , # (TEMPORARY)
228
228
verbose : bool = True ,
229
229
):
230
230
"""Load a llama.cpp model from `model_path`.
@@ -287,8 +287,8 @@ def __init__(
287
287
if n_gqa is not None :
288
288
self .params .n_gqa = n_gqa
289
289
290
- if rms_eps_norm is not None :
291
- self .params .rms_eps_norm = rms_eps_norm
290
+ if rms_norm_eps is not None :
291
+ self .params .rms_norm_eps = rms_norm_eps
292
292
293
293
self .last_n_tokens_size = last_n_tokens_size
294
294
self .n_batch = min (n_ctx , n_batch )
@@ -1533,7 +1533,7 @@ def __getstate__(self):
1533
1533
tensor_split = self .tensor_split ,
1534
1534
### TEMPORARY ###
1535
1535
n_gqa = self .params .n_gqa ,
1536
- rms_eps_norm = self .params .rms_eps_norm ,
1536
+ rms_norm_eps = self .params .rms_norm_eps ,
1537
1537
### TEMPORARY ###
1538
1538
### DEPRECATED ###
1539
1539
n_parts = self .n_parts ,
@@ -1559,11 +1559,11 @@ def __setstate__(self, state):
1559
1559
lora_base = state ["lora_base" ],
1560
1560
lora_path = state ["lora_path" ],
1561
1561
tensor_split = state ["tensor_split" ],
1562
- n_gqa = state ["n_gqa" ],
1563
- ### TEMPORARY ###
1564
- rms_eps_norm = state ["rms_eps_norm" ],
1565
1562
verbose = state ["verbose" ],
1566
1563
### TEMPORARY ###
1564
+ n_gqa = state ["n_gqa" ],
1565
+ rms_norm_eps = state ["rms_norm_eps" ],
1566
+ ### TEMPORARY ###
1567
1567
### DEPRECATED ###
1568
1568
n_parts = state ["n_parts" ],
1569
1569
### DEPRECATED ###
0 commit comments