End-to-end response selection based on multi-level context response matching
Tóm tắt
Từ khóa
Tài liệu tham khảo
Bahdanau, D., Cho, K., Bengio, Y., 2014. Neural machine translation by jointly learning to align and translate. arXiv:1409.0473.
Baudiš, P., Pichl, J., Vyskočil, T., Šedivý, J., 2016. Sentence pair scoring: towards unified framework for text comprehension. arXiv:1603.06127.
Ben-Hur, 2001, Support vector clustering, J. Mach. Learn. Res., 2, 125
Bojanowski, 2017, Enriching word vectors with subword information, Trans. Assoc. Comput. Linguist., 5, 135, 10.1162/tacl_a_00051
Boussaha, 2019, Multi-level context response matching in retrieval-based dialog systems
Boussaha, 2019, Towards simple but efficient next utterance ranking
Cao, 2008, Selecting good expansion terms for pseudo-relevance feedback, 243
Chen, 2019, Sequential attention-based network for noetic end-to-end response selection
Chen, 2019, Sequential matching model for end-to-end multi-turn response selection, 7350
Chen, 2018, Neural natural language inference models enhanced with external knowledge, 2406
Chollet, F., et al., 2015. Keras. https://github.com/keras-team/keras.
Chung, 2014, Empirical evaluation of gated recurrent neural networks on sequence modeling
Fader, 2013, Paraphrase-driven learning for open question answering, 1608
Goodfellow, 2014, Generative adversarial nets, 2672
Gunasekara, 2019, Dstc7 task 1: noetic end-to-end response selection
Hannun, A., Case, C., Casper, J., Catanzaro, B., Diamos, G., Elsen, E., Prenger, R., Satheesh, S., Sengupta, S., Coates, A., et al., 2014. Deep speech: scaling up end-to-end speech recognition. arXiv:1412.5567.
He, 2013
Hou, 2018, Sequence-to-sequence data augmentation for dialogue language understanding, 1234
Kadlec, 2015, Improved deep learning baselines for ubuntu corpus dialogs
Kingma, 2015, Adam: a method for stochastic optimization
Krizhevsky, 2012, Imagenet classification with deep convolutional neural networks, 1097
Kummerfeld, J. K., Gouravajhala, S. R., Peper, J., Athreya, V., Gunasekara, C., Ganhotra, J., Patel, S. S., Polymenakos, L., Lasecki, W. S., 2018. Analyzing assumptions in conversation disentanglement research through the lens of a new dataset and model. arXiv:1810.11118.
Kummerfeld, 2019, A large-scale corpus for conversation disentanglement
Kurata, 2016, Labeled data generation with encoder-decoder LSTM for semantic slot filling, 725
LeCun, 1998, Gradient-based learning applied to document recognition, Proc. IEEE, 86, 2278, 10.1109/5.726791
Li, 2016, A diversity-promoting objective function for neural conversation models, 110
Li, 2019, Insufficient data can also rock! Learning to converse using smaller data with augmentation, 6698
Lowe, 2015, The ubuntu dialogue corpus: a large dataset for research in unstructured multi-turn dialogue systems, 285
Lowe, 2017, Training end-to-end dialogue systems with the ubuntu dialogue corpus, Dialogue Discourse, 8, 31, 10.5087/dad.2017.102
Pedregosa, 2011, Scikit-learn: machine learning in Python, J. Mach. Learn. Res., 12, 2825
Qiu, 2017, Alime chat: a sequence to sequence and rerank based chatbot engine, 498
Serban, 2016, Building end-to-end dialogue systems using generative hierarchical neural network models, 3776
Shao, 2017, Generating high-quality and informative conversation responses with sequence-to-sequence models, 2210
Shum, 2018, From eliza to xiaoice: challenges and opportunities with social chatbots, Front. Inf. Technol. Electron. Eng., 19, 10, 10.1631/FITEE.1700826
Song, 2018, An ensemble of retrieval-based and generation-based human-computer conversation systems, 4382
Sordoni, 2015, A hierarchical recurrent encoder-decoder for generative context-aware query suggestion, 553
Strubell, 2019, Energy and policy considerations for deep learning in NLP, 3645
Sutskever, 2014, Sequence to sequence learning with neural networks, 3104
Tan, 2015, Lstm-based deep learning models for non-factoid answer selection
Tenenbaum, 2000, Separating style and content with bilinear models, Neural Comput., 12, 1247, 10.1162/089976600300015349
Theano Development Team, 2016. Theano: a Python framework for fast computation of mathematical expressions. arXiv:1605.02688.
Vaswani, 2017, Attention is all you need, 5998
Vinyals, 2015, A neural conversational model
Wan, 2016, A deep architecture for semantic matching with multiple positional sentence representations, 2835
Wang, 2016, Learning natural language inference with LSTM, 1442
Wang, 2017, Gated self-matching networks for reading comprehension and question answering, 189
Wu, Y., Wu, W., Li, Z., Zhou, M., 2016. Response selection with topic clues for retrieval-based chatbots. arXiv:1605.00090.
Wu, 2017, Sequential matching network: a new architecture for multi-turn response selection in retrieval-based chatbots, 496
Xu, 2017, Incorporating loose-structured knowledge into conversation modeling via recall-gate LSTM, 3506
Yan, 2016, Learning to respond with deep neural networks for retrieval-based human-computer conversation system, 55
Yang, 2018, Response ranking with deep matching networks and external knowledge in information-seeking conversation systems, 245
Yoshino, K., Hori, C., Perez, J., D’Haro, L. F., Polymenakos, L., Gunasekara, C., Lasecki, W. S., Kummerfeld, J., Galley, M., Brockett, C., Gao, J., Dolan, B., Gao, S., Marks, T. K., Parikh, D., Batra, D., 2018. The 7th dialog system technology challenge. arXiv preprint.
Zhang, 2015, Character-level convolutional networks for text classification, 649
Zhang, 2018, Modeling multi-turn conversation with deep utterance aggregation, 3740
Zhou, 2016, Multi-view response selection for human-computer conversation, 372
Zhou, 2018, Multi-turn response selection for chatbots with deep attention matching network, 1118