Renyu HuHao XuYang XiaoChenjun WuHaiyang Jia
Abstract Machine translation is a classic problem in natural language process (NLP). Recent years, the encoder and decoder through an attention mechanism has become a trend. Google proposed a new simple network architecture, the Transformer using attention mechanisms only rather than CNN or RNN in 2017. However, it may lose some important information (e.g., grammatical component, etc) when using attention mechanism for whole sentence.We propose a new brand model based on transformer using Group attention layers and group position embedding. It absorbs the features of Group-CNN combines the algorithm in computer vision (CV) and NLP. The model not only pays more attention to the ingredients (e.g., subject, predicate and adverbial, etc), but also enhances the connection of phrase. It outperforms SofA Transformer in using more syntactic information.
Liyan KangShaojie HeMingxuan WangFei LongJinsong Su
Ruigang LiangYing CaoPeiwei HuKai Chen