BERT: Training Longer and with More Data to Return to SOTA

BERT: Training Longer and with More Data to Return to SOTA

Machine Heart Report Contributors: Si Yuan, Qian Zhang The championship throne of XLNet has not yet warmed up, and the plot has once again taken a turn. Last month, XLNet comprehensively surpassed BERT on 20 tasks, creating a new record for NLP pre-training models and enjoyed a moment of glory. However, now, just a month … Read more