Understanding Generative AI Technology

Understanding Generative AI Technology

Generative AI is a type of artificial intelligence technology that utilizes machine learning models and deep learning techniques to generate new content by studying patterns in historical data. This content can be text, images, audio, or video. Unlike traditional systems that generate output based on given rules or data, generative AI autonomously creates entirely new … Read more

Explosive! Deepseek-Janus-Pro Can Recognize Image Addresses and Tell Stories

Explosive! Deepseek-Janus-Pro Can Recognize Image Addresses and Tell Stories

This public account mainly focuses on cutting-edge AI technologies such as NLP, CV, LLM, RAG, and Agents, sharing practical industry cases and courses for free, helping you fully embrace AIGC. 1. Janus-Pro Can Perform 5 Tasks 1.1 Image Description 1.2 Location Recognition 1.3 Background Inference 1.4 OCR Text Recognition 1.5 Text-Image Generation 2. Principles of … Read more

DeepSeek Janus Series: Unified Multimodal Understanding and Generation Models

DeepSeek Janus Series: Unified Multimodal Understanding and Generation Models

DeepSeek released the multimodal model Janus-Pro this morning. The Janus series includes three model components: Janus, JanusFlow, and Janus-Pro, significantly enhancing multimodal understanding and visual generation performance. Currently, the code and models are fully open. I tried them today and felt that the results were average (details below), and there is still a certain gap … Read more

How BERT Tokenizes Text

How BERT Tokenizes Text

Follow the official account “ML_NLP“ Set as “Starred“, delivering heavy content promptly! Source | Zhihu Link | https://zhuanlan.zhihu.com/p/132361501 Author | Alan Lee Editor | Machine Learning Algorithms and Natural Language Processing Public Account This article is authorized and reposting is prohibited This article was first published on my personal blog on 2019/10/16 and cannot be … Read more

Pre-training BERT: How TensorFlow Solved It Before Official Release

Pre-training BERT: How TensorFlow Solved It Before Official Release

Edited by Machine Heart Contributors: Siyuan, Wang Shuting This month, Google’s BERT has received a lot of attention, as the research has refreshed the state-of-the-art performance records in 11 NLP tasks with its pre-trained model. The authors of the paper stated that they would release the code and pre-trained model by the end of this … Read more

Beginner’s Guide to Using BERT: Principles and Hands-On Examples

Beginner's Guide to Using BERT: Principles and Hands-On Examples

Author Jay Alammar, Translated by QbitAI | WeChat Official Account QbitAI BERT, as a key player in the field of natural language processing, is an unavoidable topic for NLPer. However, for those with little experience and a weak foundation, mastering BERT can be a bit challenging. Now, tech blogger Jay Alammar has created a “Visual … Read more

Google Automatically Generates Text from Knowledge Graphs

Google Automatically Generates Text from Knowledge Graphs

New Intelligence Report Source: Google AI Editor: LRS [New Intelligence Guide] Based on pre-training experience, more data leads to better performance! Google recently published a paper at NAACL 2021 that can automatically generate text data from knowledge graphs, so there’s no need to worry about insufficient corpora anymore! Large pre-trained natural language processing (NLP) models, … Read more

How to Use BERT and GPT-2 in Your Models

How to Use BERT and GPT-2 in Your Models

Recommended by New Intelligence Source: Zhuanzhi (ID: Quan_Zhuanzhi) Editor: Sanshi [New Intelligence Guide] In the field of NLP, various advanced tools have emerged recently. However, practice is the key, and how to apply them to your own models is crucial. This article introduces this issue. Recently in NLP, various pre-trained language models like ELMO, GPT, … Read more

BERT-of-Theseus: A Model Compression Method Based on Module Replacement

BERT-of-Theseus: A Model Compression Method Based on Module Replacement

©PaperWeekly Original · Author|Su Jianlin School|Zhuiyi Technology Research Direction|NLP, Neural Networks Recently, I learned about a BERT model compression method called “BERT-of-Theseus”, derived from the paper BERT-of-Theseus: Compressing BERT by Progressive Module Replacing. This is a model compression scheme built on the concept of “replaceability”. Compared to conventional methods like pruning and distillation, it appears … Read more

When BERT Meets Knowledge Graphs

When BERT Meets Knowledge Graphs

Author: Gao Kaiyuan School: Shanghai Jiao Tong University Research Direction: Natural Language Processing Zhihu Column: BERT on the Shoulders of Giants Original Article Link: https://zhuanlan.zhihu.com/p/91052495 Introduction In the previous blog, I discussed some knowledge representation learning models. Today, let’s explore the current most popular BERT model and how it develops with the addition of external … Read more