4
4

Delete article

Deleted articles cannot be recovered.

Draft of this article would be also deleted.

Are you sure you want to delete this article?

More than 1 year has passed since last update.

自然言語処理で引用文献を楽したい人向けの論文一覧(形態素解析、wrod2vecからBERTまで)

Last updated at Posted at 2022-11-16

参考文献を書くのって意外と手間がかかりめんどくさい…
自然言語処理を勉強している学生、社会人の方に向けて、形態素解析ツールと言語モデル論文の一部をまとめました。

2022/11/19 コメントを受けて形態素解析の引用文献を改善

形態素解析

Mecab

Taku Kudo, Kaoru Yamamoto, Yuji Matsumoto: Applying Conditional Random Fields to Japanese Morphological Analysis, Proceedings of the 2004 Conference on Empirical Methods in Natural Language Processing (July 2004), pp.230-237.

(https://aclanthology.org/W04-3230)

Juman++

Arseny Tolmachev, Daisuke Kawahara, Sadao Kurohashi: Juman++: A Morphological Analysis Toolkit for Scriptio Continua, Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing (November 2018): System Demonstrations, pp.54-59.

(https://dx.doi.org/10.18653/v1/D18-2010)

Sudachi

Kazuma Takaoka, Sorami Hisamoto, Noriko Kawahara, Miho Sakamoto, Yoshitaka Uchida, Yuji Matsumoto: Sudachi: a Japanese Tokenizer for Business, LREC 2018: Eleventh International Conference on Language Resources and Evaluation (May 2018), pp.2246-2249.

(https://aclanthology.org/L18-1355)

GINZA

松田寛: GiNZA - Universal Dependenciesによる実用的日本語解析, 自然言語処理, Vol.27, No.3 (2020年9月), pp.695-701. 

(https://doi.org/10.5715/jnlp.27.695)

Stanza

Peng Qi, Yuhao Zhang, Yuhui Zhang, Jason Bolton, Christopher D. Manning: Stanza: A Python Natural Language Processing Toolkit for Many Human Languages, 58th Annual Meeting of the Association for Computational Linguistics: Proceedings of the System Demonstration (July 2020), pp.101-108.
(https://dx.doi.org/10.18653/v1/2020.acl-demos.14)

言語モデル

Word2Vec

Mikolov, T., Chen, K., Corrado, G., and Dean, J. (2013). Efficient Estimation of Word Representations in Vector Space. arXiv:1301.3781.

RNN

J. L. Elman. (1990). Finding structure in time, Cognitive science, 14(2), 179-211

LSTM

S. Hochreiter and J. Schmidhuber. (1997). Long short-term memory, Neural computation, 9(8), 1735-1780.

ELMo

Matthew E. Peters, Mark Neumann, Mohit Iyyer, Matt Gardner, Christopher Clark, Kenton Lee, Luke Zettlemoyer. (2018). Deep contextualized word representations, arXiv:1301.05365v2.

Transformer

Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Lukasz Kaiser, and Illia Polosukhin. (2017). Attention is all you need. arXiv:1706.03762.

BERT

Jacob Devlin, Ming-Wei Chang, Kenton Lee, Kristina Toutanova. (2018). BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. arXiv:1810.04805.

4
4
2

Register as a new user and use Qiita more conveniently

  1. You get articles that match your needs
  2. You can efficiently read back useful information
  3. You can use dark theme
What you can do with signing up
4
4

Delete article

Deleted articles cannot be recovered.

Draft of this article would be also deleted.

Are you sure you want to delete this article?