BERT: Pre-Training of Deep Bidirectional Transformers for Language Understanding


New top story on Hacker News: BERT: Pre-Training of Deep Bidirectional Transformers for Language Understanding