LZ_Jaja

  博客园  :: 首页  :: 新随笔  :: 联系 :: 订阅 订阅  :: 管理

Link of the Paper: https://arxiv.org/pdf/1409.3215.pdf

Main Points:

  1. Encoder-Decoder Model: Input sequence -> A vector of a fixed dimensionality -> Target sequence.
  2. A multilayered  LSTM: The LSTM did not have difficulty on long sentences. Deep LSTMs significantly outperformed shallow LSTMs.
  3. Reverse Input: Better performance. While the authors do not have a complete explanation to this phenomenon, they believe that it is caused by the introduction of many short term dependencies to the dataset. LSTMs trained on reversed source sentences did much better on long sentences than LSTMs trained on the raw source sentences, which suggests that reversing the input sentences results in LSTMs with better memory utilization.

Other Key Points:

  1. A significant limitation: Despite their flexibility and power, DNNs can only be applied to problems whose inputs and targets can be sensibly encoded with vectors of fixed dimensionality.
posted on 2018-08-09 10:06  LZ_Jaja  阅读(289)  评论(0编辑  收藏  举报