Understanding Mamba: The Strongest Competitor to Transformers

Understanding Mamba: The Strongest Competitor to Transformers

MLNLP community is a well-known machine learning and natural language processing community both domestically and internationally, covering NLP graduate students, university professors, and corporate researchers.The vision of the community is to promote communication and progress between the academic and industrial communities in natural language processing and machine learning, especially for beginners. Reprinted from | Machine … Read more

What Does ‘GPT’ Mean in ChatGPT?

What Does 'GPT' Mean in ChatGPT?

Writing scripts, creating novels, coding, answering questions… the almost omnipotent ChatGPT has become a frequent topic on hot search lists in recent months. At the end of November last year, ChatGPT was launched and quickly went viral on social media. In just five days, the number of registered users exceeded 1 million; within two months, … Read more

Current Status and Analysis of Pre-trained Models in NLP

Current Status and Analysis of Pre-trained Models in NLP

Author | Wang Zeyang Organization | Niutrans Editor | Tang Li Reprinted from WeChat Official Account | AI Technology Review This article is submitted by Wang Zeyang, a graduate student from the Natural Language Processing Laboratory of Northeast University. Wang Zeyang’s research direction is machine translation. Niutrans, whose core members come from the Natural Language … Read more

Mastering DeepSeek: A Guide to Large Model Learning and Pitfalls

Mastering DeepSeek: A Guide to Large Model Learning and Pitfalls

Preface Happy holidays, everyone. During the New Year, I believe many of you have been inundated with the domestic AI large model DeepSeek. With the rapid development of artificial intelligence, DeepSeek has become a hot large language model (LLM). However, many students still view it as just a “chatbot” or mistakenly believe it to be … Read more

Unlocking the Mystery of 1-Layer Transformer: Attention Mechanism Explained

Unlocking the Mystery of 1-Layer Transformer: Attention Mechanism Explained

Click on the above“Beginner Learning Vision”, select to add Star or Top” Essential insights delivered in real-time This is for academic sharing only and does not represent the views of this public account. Contact for removal in case of infringement.Reprinted from: New Intelligence Source The Transformer architecture has swept across multiple fields including natural language … Read more

Why Transformer-Based Large Models Dominate Development!

Why Transformer-Based Large Models Dominate Development!

In the development of artificial intelligence, the evolution of model architectures is comparable to the process of the human industrial revolution. From the initial traditional neural networks to today’s Transformer architecture, each technological innovation has propelled a leap in AI technology. The emergence of the Transformer model has not only changed the landscape of natural … Read more

Latest Overview of Multimodal Pre-training Models

Latest Overview of Multimodal Pre-training Models

Follow the public account “ML_NLP“ Set as “Starred“, delivering heavy content promptly! Reprinted from | Zhihu Author | Liang Chao Wei from Summer ResortOriginal link | https://zhuanlan.zhihu.com/p/412126626 01 – Background In the traditional NLP unimodal field, the development of representation learning is relatively mature. However, in the multimodal field, due to the scarcity of high-quality … Read more

EHCTNet: Enhanced Hybrid of CNN and Transformer Network for Remote Sensing Image Change Detection

EHCTNet: Enhanced Hybrid of CNN and Transformer Network for Remote Sensing Image Change Detection

Good news! Join Knowledge Planet to read the full PDF version of this article in detail Paper Information Title: EHCTNet: Enhanced Hybrid of CNN and Transformer Network for Remote Sensing Image Change Detection EHCTNet: Enhanced Hybrid of CNN and Transformer Network for Remote Sensing Image Change Detection Authors: Junjie Yang, Haibo Wan, Zhihai Shang Innovations … Read more

Alibaba’s Tora: A Trajectory-Controlled DiT Video Generation Model

Follow our official account to discover the beauty of CV technology This paper shares Tora: Trajectory-oriented Diffusion Transformer for Video Generation, where Alibaba proposes the trajectory-controlled DiT video generation model Tora. Paper link: https://arxiv.org/abs/2407.21705 Project link: https://ali-videoai.github.io/tora_video/ Background Video generation models have recently made significant progress. For example, OpenAI’s Sora and domestic models like Vidu … Read more

Detailed Module Analysis of DETR Structure

Detailed Module Analysis of DETR Structure

Transformers shine in the field of computer vision, and the Detection Transformer (DETR) is a successful application of Transformers in object detection. By utilizing the attention mechanism in Transformers, it effectively models long-range dependencies in images, simplifying the object detection pipeline and constructing an end-to-end object detector. Object detection can be understood as a set … Read more