Update modeling_mixformer_sequential.py
Browse filesRemoves print regarding attention_mask to prevent excessive information from being logged.
modeling_mixformer_sequential.py
CHANGED
@@ -756,9 +756,6 @@ class MixFormerSequentialForCausalLM(MixFormerSequentialPreTrainedModel):
|
|
756 |
labels: Optional[torch.LongTensor] = None,
|
757 |
**kwargs,
|
758 |
) -> CausalLMOutputWithPast:
|
759 |
-
if attention_mask is not None and self.training:
|
760 |
-
print("`attention_mask` is not supported during training. Using it might lead to unexpected results.")
|
761 |
-
|
762 |
if past_key_values is None and attention_mask is None:
|
763 |
lm_logits = self.layers(input_ids)
|
764 |
else:
|
|
|
756 |
labels: Optional[torch.LongTensor] = None,
|
757 |
**kwargs,
|
758 |
) -> CausalLMOutputWithPast:
|
|
|
|
|
|
|
759 |
if past_key_values is None and attention_mask is None:
|
760 |
lm_logits = self.layers(input_ids)
|
761 |
else:
|