Amazing! LSTM With Only Forget Gate Outperforms Standard LSTM

Amazing! LSTM With Only Forget Gate Outperforms Standard LSTM

Selected from arXiv Authors:Jos van der Westhuizen, Joan Lasenby Compiled by Machine Heart Contributors: Pedro, Lu This paper studies what happens when LSTM only has a forget gate and proposes JANET, with experiments showing that this model outperforms standard LSTM. 1. Introduction Excellent engineers ensure their designs are practical. We now know that the best … Read more

Visualizing LSTM Model Structure

Visualizing LSTM Model Structure

Source: Deep Learning Enthusiasts Author on Zhihu | Master Su Link | https://zhuanlan.zhihu.com/p/139617364 This article is about 3200 words, and it is recommended to read in 5 minutes This article introduces the visualization of the LSTM model structure. Recently, I have been studying the application of LSTM in time series prediction, but I encountered a … Read more

Application of Bi-LSTM + CRF in Text Sequence Labeling

Application of Bi-LSTM + CRF in Text Sequence Labeling

Machine Heart Column Author: Principal Engineer Dong Bingfeng, Chubao AI Lab The input vector X in traditional CRF is generally in the one-hot format of words, which loses a lot of semantic information about words. With the advent of word embedding methods, the word representation in vector form generally performs better than the one-hot representation. … Read more