论文列表

FRATERNAL DROPOUT

[J]. arXiv preprint arXiv:1711.00066, 2017.

Regularizing and Optimizing LSTM Language Models

[J]. arXiv preprint arXiv:1708.02182, 2017.

DYNAMIC EVALUATION OF NEURAL SEQUENCE MODELS

[C]. International Conference on Machine Learning. PMLR, 2018: 2766-2775.

Dropout: A Simple Way to Prevent Neural Networks from Overfitting

[J]. The journal of machine learning research, 2014, 15(1): 1929-1958.

Comparison of Modern Stochastic Optimization Algorithms

[J]. University of Edinburgh, Edinburgh, 2014.

No Free Lunch Theorems for Optimization

[J]. IEEE transactions on evolutionary computation, 1997, 1(1): 67-82.

KagNet: Knowledge-Aware Graph Networks for Commonsense Reasoning

[J]. Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), 2019.

Convolutional Recurrent Neural Networks for Text Classification

[C]// 2019 International Joint Conference on Neural Networks (IJCNN). 2019.

Attention-Based Bidirectional Long Short-Term Memory Networks for Relation Classification

[C]// Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers). 2016.

Convolutional Neural Networks for Sentence Classification

[J]. arXiv preprint arXiv:1408.5882, 2014.

Learning to summarize from human feedback

[J]. arXiv preprint arXiv:2009.01325, 2020.