Fill-Mask
Transformers
PyTorch
Safetensors
English
nomic_bert
custom_code
zpn commited on
Commit
f69b8d1
·
1 Parent(s): c93435c

fix: load_return

Browse files
Files changed (1) hide show
  1. modeling_hf_nomic_bert.py +3 -1
modeling_hf_nomic_bert.py CHANGED
@@ -458,7 +458,9 @@ class NomicBertPreTrainedModel(PreTrainedModel):
458
  state_dict = filter_shapes(state_dict, model)
459
 
460
  load_return = model.load_state_dict(state_dict, strict=strict)
461
- logger.warning(load_return)
 
 
462
  return model
463
 
464
  def _set_gradient_checkpointing(self, module, value=False):
 
458
  state_dict = filter_shapes(state_dict, model)
459
 
460
  load_return = model.load_state_dict(state_dict, strict=strict)
461
+ # moe models load new weights
462
+ if getattr(config, "moe_top_k", 0) == 0:
463
+ logger.warning(load_return)
464
  return model
465
 
466
  def _set_gradient_checkpointing(self, module, value=False):