GrammarTagger — A Neural Multilingual Grammar Profiler for Language Education
We are happy to announce GrammarTagger, an open-source toolkit for grammatical profiling for language learning. It can analyze text in English and Chinese and show you grammatical items included in the input, along with its estimated difficulty.
Complete Guide to Subword Tokenization Methods in the Neural Era
In deep natural language processing (NLP), the input text is often broken into "subword"—units that are shorter than words. In this article, we'll review common subword tokenization techniques including WordPiece, byte-pair encoding (BPE), and SentencePiece,
Complete Guide to Japanese BERT: Choosing the Right Pretrained Language Model for You
Pretrained language models (PLMs) such as BERT are used for solving more and more NLP applications in many languages, including Japanese. In this post, I'm going to compare various Japanese pretrained BERT models and their task performance and make a specific recommendation as of this writing.
State of Automated Essay Scoring with Pretrained Language Models
Automated essay scoring (AES), where a computer algorithm scores student essays automatically, is an important application of natural language processing for education and language learning. In this article, we'll review some of the recent advances in automated essay scoring.