How can I do a seq2seq task with PyTorch Transformers if I am not trying to be autoregressive?

后端 未结 0 1958
刺人心
刺人心 2020-12-02 15:29

I may be mistaken, but it seems that PyTorch Transformers are autoregressive, which is what masking is for. However, I\'ve seen some implementations where people use just th

相关标签:
回答
  • 消灭零回复
提交回复
热议问题