In this paper we show that a simple beam approximation of the joint distribution between attention and output is an easy, accurate, and efficient attention mechanism for sequence to sequence learning. The method combines the advantage of sharp focus in hard attention and the implementation ease of soft attention. On five translation tasks we show effortless and consistent gains in BLEU compared to existing attention mechanisms.
Yi Ni TohRoger W. RemingtonVanessa G. Lee
Yi Ni TohRoger W. RemingtonVanessa G. Lee
Roger W. SchvaneveldtRebecca L. Gómez
Takaaki HoriHai WangChiori HoriShinji WatanabeBret HarshamJonathan Le RouxJohn R. HersheyYusuke KojiJing YiZhaocheng ZhuTakeyuki Aikawa
Asher CohenRichard IvrySteven W. Keele