Understanding ALBERT in Interviews

Understanding ALBERT in Interviews

Follow the WeChat public account “ML_NLP” Set it as “Starred”, heavy content delivered first time! Source | Zhihu Address | https://zhuanlan.zhihu.com/p/268130746 Author | Mr.robot Editor | Machine Learning Algorithms and Natural Language Processing WeChat Public Account This article has been authorized by the author, and secondary reproduction is prohibited without permission. Interviewer: Do you understand … Read more

Text and Visual: Introduction to Multiple Visual/Video BERT Papers

Text and Visual: Introduction to Multiple Visual/Video BERT Papers

Reprinted from WeChat Official Account: AI Technology Review Author: Yang Xiaofan Since the success of Google’s BERT model in 2018, more and more researchers have drawn on BERT’s ideas for tasks beyond pure text, developing various visual/video (Visual/Video) fusion BERT models.Here we introduce the original VideoBERT paper and six other recent V-BERT papers (sorted in … Read more

Few-Shot NER with Dual-Tower BERT Model

Few-Shot NER with Dual-Tower BERT Model

Delivering NLP technical insights to you every day! Author | SinGaln Source | PaperWeekly This is an article from ACL 2022. The overall idea is to use a dual-tower BERT model to encode text tokens and their corresponding labels separately based on meta-learning, and then perform classification on the output obtained from the dot product … Read more

Exploring the Transformer Model: Understanding GPT-3, BERT, and T5

Exploring the Transformer Model: Understanding GPT-3, BERT, and T5

Author: Dale Markowitz Translation: Wang Kehan Proofreading: He Zhonghua This article is approximately 3800 words long and is recommended to be read in 5 minutes This article introduces the currently most popular language model in natural language processing—the Transformer model. Tags: Natural Language Processing Do you know this saying: When you have a hammer, everything … Read more

A Comprehensive Overview of Named Entity Recognition (NER)

A Comprehensive Overview of Named Entity Recognition (NER)

MLNLP(Machine Learning Algorithms and Natural Language Processing) is one of the largest natural language processing communities in China and abroad, gathering over 500,000 subscribers, with an audience that includes NLP graduate students, university professors, and researchers from enterprises. The vision of the community is to promote communication and progress among the academic and industrial circles … Read more

Sentence-BERT: A Siamese Network for Fast Sentence Similarity Computation

Sentence-BERT: A Siamese Network for Fast Sentence Similarity Computation

Follow the public account “ML_NLP“ Set as “Starred“, delivering heavy content promptly! Author: Shining School: Beijing University of Posts and Telecommunications Original article link: https://www.cnblogs.com/gczr/p/12874409.html 1. Background Introduction   BERT and RoBERTa have achieved SOTA results in regression tasks for sentence pairs, such as text semantic similarity. However, they require feeding both sentences into the network … Read more

Do You Really Need GPT-3? BERT’s MLM Model Also Enables Few-Shot Learning

Do You Really Need GPT-3? BERT's MLM Model Also Enables Few-Shot Learning

Follow the official account “ML_NLP“ Set it as “Starred“, delivering heavy content immediately! Source|PaperWeekly ©PaperWeekly Original · Author|Su Jianlin Unit|Zhuiyi Technology Research Direction|NLP, Neural Networks As we all know, GPT-3 is currently very popular, however, everywhere we see promotions for GPT-3, do readers remember the name of the GPT-3 paper? In fact, the paper is … Read more

Soft-Masked BERT: Latest Integration of Text Correction and BERT

Soft-Masked BERT: Latest Integration of Text Correction and BERT

Follow the public account “ML_NLP“ Set as “Starred“, delivering heavy content to you first! From | Zhihu Address | https://zhuanlan.zhihu.com/p/144995580 Author | Ye Chen Editor | Simple AI Text correction is a technology in the field of natural language processing that detects whether a piece of text contains typos and corrects them. It is generally … Read more

Exclusive: BERT Model Compression Based on Knowledge Distillation

Exclusive: BERT Model Compression Based on Knowledge Distillation

Authors: Siqi Sun, Yu Cheng, Zhe Gan, Jingjing Liu This article is about1800 words, recommended reading time5 minutes. This article introduces the “Patient Knowledge Distillation” model. Data Department THU backend reply“191010”, get the paper address. In the past year, there have been many groundbreaking advances in language model research, such as GPT generating sentences that … Read more

ACL2022 | Dual Tower BERT Model with Label Semantics

ACL2022 | Dual Tower BERT Model with Label Semantics

Delivering NLP technology insights to you every day! Source: Alchemy Notes Author: SinGaln This is a paper from ACL 2022. The overall idea is to use a dual tower BERT model based on meta-learning to encode text tokens and their corresponding labels separately, and then perform a classification task using the output obtained from their … Read more