Associative Long Short-Term Memory
ICML'16: Proceedings of the 33rd International Conference on International Conference on Machine Learning - Volume 48(2016)
摘要
We investigate a new method to augment recurrent neural networks with extra memory without increasing the number of network parameters. The system has an associative memory based on complex-valued vectors and is closely related to Holographic Reduced Representations and Long Short-Term Memory networks. Holographic Reduced Representations have limited capacity: as they store more information, each retrieval becomes noisier due to interference. Our system in contrast creates redundant copies of stored information, which enables retrieval with reduced noise. Experiments demonstrate faster learning on multiple memorization tasks.
更多查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络