参考文献を書くのって意外と手間がかかりめんどくさい…
自然言語処理を勉強している学生、社会人の方に向けて、形態素解析ツールと言語モデル論文の一部をまとめました。
2022/11/19 コメントを受けて形態素解析の引用文献を改善
形態素解析
Mecab
Taku Kudo, Kaoru Yamamoto, Yuji Matsumoto: Applying Conditional Random Fields to Japanese Morphological Analysis, Proceedings of the 2004 Conference on Empirical Methods in Natural Language Processing (July 2004), pp.230-237.
(https://aclanthology.org/W04-3230)
Juman++
Arseny Tolmachev, Daisuke Kawahara, Sadao Kurohashi: Juman++: A Morphological Analysis Toolkit for Scriptio Continua, Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing (November 2018): System Demonstrations, pp.54-59.
(https://dx.doi.org/10.18653/v1/D18-2010)
Sudachi
Kazuma Takaoka, Sorami Hisamoto, Noriko Kawahara, Miho Sakamoto, Yoshitaka Uchida, Yuji Matsumoto: Sudachi: a Japanese Tokenizer for Business, LREC 2018: Eleventh International Conference on Language Resources and Evaluation (May 2018), pp.2246-2249.
(https://aclanthology.org/L18-1355)
GINZA
松田寛: GiNZA - Universal Dependenciesによる実用的日本語解析, 自然言語処理, Vol.27, No.3 (2020年9月), pp.695-701.
(https://doi.org/10.5715/jnlp.27.695)
Stanza
Peng Qi, Yuhao Zhang, Yuhui Zhang, Jason Bolton, Christopher D. Manning: Stanza: A Python Natural Language Processing Toolkit for Many Human Languages, 58th Annual Meeting of the Association for Computational Linguistics: Proceedings of the System Demonstration (July 2020), pp.101-108.
(https://dx.doi.org/10.18653/v1/2020.acl-demos.14)
言語モデル
Word2Vec
Mikolov, T., Chen, K., Corrado, G., and Dean, J. (2013). Efficient Estimation of Word Representations in Vector Space. arXiv:1301.3781.
RNN
J. L. Elman. (1990). Finding structure in time, Cognitive science, 14(2), 179-211
LSTM
S. Hochreiter and J. Schmidhuber. (1997). Long short-term memory, Neural computation, 9(8), 1735-1780.
ELMo
Matthew E. Peters, Mark Neumann, Mohit Iyyer, Matt Gardner, Christopher Clark, Kenton Lee, Luke Zettlemoyer. (2018). Deep contextualized word representations, arXiv:1301.05365v2.
Transformer
Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Lukasz Kaiser, and Illia Polosukhin. (2017). Attention is all you need. arXiv:1706.03762.
BERT
Jacob Devlin, Ming-Wei Chang, Kenton Lee, Kristina Toutanova. (2018). BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. arXiv:1810.04805.