Building Language Applications with Hugging Face Transformers

Building Language Applications with Hugging Face Transformers

Hugging Face is a chatbot startup based in New York, focusing on NLP technology, with a large open-source community. Especially, the open-source natural language processing and pre-trained model library, Transformers, has been downloaded over a million times and has more than 24,000 stars on GitHub. Transformers provides a large number of state-of-the-art pre-trained language model … Read more

Understanding Huggingface BERT Source Code: Application Models and Training Optimization

Understanding Huggingface BERT Source Code: Application Models and Training Optimization

Follow our public account “ML_NLP“ Set as “Starred“, heavy content delivered first time! Reprinted from | PaperWeekly ©PaperWeekly Original · Author|Li Luoqiu School|Zhejiang University Master’s Student Research Direction|Natural Language Processing, Knowledge Graph Continuing from the previous article, I will record my understanding of the HuggingFace open-source Transformers project code. This article is based on the … Read more

Hugging Face Official Course Launched: Free NLP Training

Hugging Face Official Course Launched: Free NLP Training

Machine Heart reports Editor: Du Wei The Hugging Face NLP course is now live, and all courses are completely free. Those in the NLP field should be very familiar with the renowned Hugging Face, a startup focused on solving various NLP problems that has brought many beneficial technical achievements to the community. Last year, the … Read more

Unlocking Speech Language Model Generation with SpeechGen

Unlocking Speech Language Model Generation with SpeechGen

Machine Heart Column Machine Heart Editorial Team This article presents a unified framework called SpeechGen, which can be used for any speech LM and various speech generation tasks, showing great potential. Paper link: https://arxiv.org/pdf/2306.02207.pdf Demo page: https://ga642381.github.io/SpeechPrompt/speechgen.html Code: https://github.com/ga642381/SpeechGen Introduction and Motivation Large Language Models (LLMs) have garnered significant attention in the field of AI-generated … Read more

Has Prompt Tuning Surpassed Fine Tuning?

Has Prompt Tuning Surpassed Fine Tuning?

MLNLP community is a well-known machine learning and natural language processing community both domestically and internationally, covering NLP graduate and doctoral students, university teachers, and corporate researchers. The community’s vision is to promote communication and progress between the academic and industrial sectors of natural language processing and machine learning, especially for beginners. Reprinted from | … Read more

Overview of Prompt Learning and Tuning

Overview of Prompt Learning and Tuning

↑ ClickBlue Text Follow the Jishi platform Author丨Jarvis73@Zhihu (Authorized) Source丨https://zhuanlan.zhihu.com/p/601905339 Editor丨Jishi Platform Jishi Guide This article summarizes Prompt Learning/Tuning. >> Join the Jishi CV technology exchange group to stay at the forefront of computer vision Since the advent of Self-Attention and Transformer, they have become the new stars in the field of natural language processing. … Read more

Overview of Prompt Engineering

Overview of Prompt Engineering

Abbreviation Explanation: task column CR: Commonsense Reasoning QA: Question Answering SUM: Summarization MT: Machine Translation LCP: Linguistic Capacity Probing GCG: General Conditional Generation CKM: Commonsense Knowledge Mining FP: Fact Probing TC: Text Classification MR: Mathematical Reasoning SR: Symbolic Reasoning AR: Analogical Reasoning Theory: Theoretical Analysis IE: Information Extraction D2T: Data-to-text TAG: Sequence Tagging SEMP: Semantic … Read more

Gzip + kNN Text Classification Beats Transformer with 14 Lines of Code

Gzip + kNN Text Classification Beats Transformer with 14 Lines of Code

A few days ago, the ACL 2023 awards were announced, attracting significant attention. Among the many papers included, one titled “Low-Resource Text Classification: A Parameter-Free Classification Method with Compressors” has started to generate much discussion. This paper was jointly completed by the University of Waterloo and AFAIK, but it is neither an award-winning paper nor … Read more

Training Word Vectors with Word2vec, Fasttext, Glove, Elmo, Bert, and Flair

Training Word Vectors with Word2vec, Fasttext, Glove, Elmo, Bert, and Flair

For all source code in this tutorial, please visit Github: https://github.com/zlsdu/Word-Embedding 1. Word2vec 1. Gensim Library The gensim library provides implementations of the Word2vec cbow model and skipgram model, which can be called directly. Full reference code 2. TensorFlow Implementation of Skipgram Model The skipgram model predicts context words based on a center word; there … Read more

Unveiling Word2Vec: A Small Step in Deep Learning, A Giant Leap in NLP

Unveiling Word2Vec: A Small Step in Deep Learning, A Giant Leap in NLP

Click the “AI Park” above to follow the public account, and choose to add a “star” or “top” Author: Suvro Banerjee Translated by: ronghuaiyang Prelude In NLP today, word vectors are indispensable. Word vectors provide us with a very good vector representation of words, allowing us to represent all words with a fixed-length vector, and … Read more