Contemporary Foreign Languages Studies >
Neural Network Models in Natural Language Processing
Natural Language Processing (NLP) is a new interdisciplinary subject to study and process the natural language by computer. In recent years, NLP has developed very rapidly and it attracted great attention from the linguistic community. This paper discusses four types of neural network model in natural language processing: Feed-forward Neural Network model (FNN), Convolutional Neural Network model (CNN), Recurrent Neural Network model (RNN), and Pre-Training model (PT), including the basic principle, structure, algorithm, and mechanism of the model, highlights their application in NLP. The paper points out that although the neural network models have become the mainstream of NLP, but these models still lack interpretability and need to be supported by rule-based language models and statistics-based language models in the futur.
FENG Zhiwei, DING Xiaomei . Neural Network Models in Natural Language Processing[J]. Contemporary Foreign Languages Studies, 2022 , 22(4) : 98 -110 . DOI: 10.3969/j.issn.1674-8921.2022.04.010
| [1] | Bengio Y., R. Ducharme, P. Vincent, et al. 2003. A neural probabilistic language model[J]. Journal of Machine Learning Research (3):1137-1155. |
| [2] | Devlin J., M. W. Chang K. Lee, et al. 2019. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding[C]. Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics. |
| [3] | Goldberg J. 2017. Neural Network Methods for Natural Language Processing[M]. New York: Morgan & Claypool Publishers. |
| [4] | Mikolov T., K. Chen, G. Corrado, et al. 2013. Efficient estimation of word representation in Vector Space[J/OL]. Computer Science (3): 24-30. |
| [5] | Mostafa D., G. Stephan, V. Oriol, U. Jakob, et al. 2018. Universal transformers[J/OL]. [2018-07-10]. https://arxiv.org/abs/1807.03819. |
| [6] | Radford A, K. Narasimhan, T. Salimans, et al. 2018. Improving language understanding by generative pre-training[J/OL]. [2018-10-22]. https://s3-us-west-2.amazonaws.com/openai-assets/researchcovers/languageunsupervised/language understanding paper.pdf. |
| [7] | Vaswani A, N. Shazeer, N. Parmar, et al. 2017. Attention is all you need[C]. Proceedings of Advances in Neural Information Processing Systems. 5998-6008. |
| [8] | 冯志伟. 2011. 计算语言学的历史回顾与现状分析[J]. 外国语(上海外国语大学学报) (1):9-17. |
| [9] | 冯志伟. 2017. 自然语言计算机形式分析的理论与方法[M]. 合肥: 中国科学技术大学出版社. |
| [10] | 冯志伟、 丁晓梅. 2021. 自然语言处理中的语言模型[J]. 外语电化教学 (6):17-24. |
| [11] | 冯志伟、 李颖. 2021. 自然语言处理中的预训练范式[J]. 外语研究 (1): 1-14. |
| [12] | 冯志伟. 2021a. 生成词向量的三种方法[J]. 外语电化教学 (1): 18-26. |
| [13] | 冯志伟. 2021b. 神经网络、深度学习与自然语言处理[J]. 上海师范大学学报(社会科学版):(2):110-122. |
| [14] | 冯洋、 邵晨泽. 2020. 神经机器翻译前沿综述[J]. 中文信息学报 (7):1-18. |
| [15] | 李沐、 刘树杰、 张冬冬、 周明. 2018. 机器翻译[M]. 北京: 高等教育出版社. |
| [16] | 邱锡鹏. 2019. 神经网络与深度学习[M/OL]. [2019-07-20]. http://nndl.github.io. |
/
| 〈 |
|
〉 |