Loss is “nan” when fine-tuning HuggingFace NLI model (both RoBERTa/BART)

后端 未结 0 1104
青春惊慌失措
青春惊慌失措 2020-12-17 06:13

I\'m using HuggingFace\'s Transformer\'s library and I’m trying to fine-tune a pre-trained NLI model (ynie/roberta-large-snli_mnli_fever_anli_R1_R2_R3-nli) on a

相关标签:
回答
  • 消灭零回复
提交回复
热议问题