0

When I'm using the huggingface transformer to train bert, i see, for every batch the input_ids are different. For example for first batch it is torch.Size([16, 171]) and second batch it is torch.Size([16, 450]). What is the reason?

kowser66
  • 125
  • 1
  • 8
  • It is most likely because the sentences in the batches have different lengths in the number of subwords. Do you have a particular example (=code snippet) that you think does not work as you expect? – Jindřich Oct 03 '22 at 08:21

0 Answers0