Thoughts on Upgrading Transformer: Simple Considerations on Multimodal Encoding Positions

Thoughts on Upgrading Transformer: Simple Considerations on Multimodal Encoding Positions

©PaperWeekly Original · Author | Su Jianlin Affiliation | Scientific Space Research Direction | NLP, Neural Networks In the second article of this series, “The Path of Transformer Upgrade: A Rotational Position Encoding that Draws on the Strengths of Many,” the author proposes Rotational Position Encoding (RoPE) — a method to achieve relative position encoding … Read more

Mistral: The Most Powerful Open Source Model

Mistral: The Most Powerful Open Source Model

Author: Jay Chou from Manchester Reviewer: Los Project Address: mistralai/mistral-src: Reference implementation of Mistral AI 7B v0.1 model This article aims to deeply analyze the key improvements of Mistral 7B and Mistral 8X7B. Mistral AI is an AI company co-founded in Paris by three former employees of DeepMind and Meta. In September 2023, Mistral AI … Read more

Language Families and Machine Translation Challenges

Language Families and Machine Translation Challenges

Source | Language Spring and Autumn Current machine translation technology can be divided into two categories: one is Rich Resource NMT, which refers to language pairs with abundant bilingual corpora (such as Chinese – English); the other is Low Resource NMT, which lacks sufficient bilingual corpora (such as Chinese – Hebrew). Current machine translation has … Read more