ABSTRACT

Modern language models in AI are deep neural networks (DNNs) capable of processing sequences in terms of the varying natures of the contexts in which they occur and in ways that did not exist in the past. This chapter explains them in a concrete manner by pointing out their pros and cons and how they can be used for “dialoguing with texts.” More specifically, it shows how DNNs can be used for processing sequences and how large language models rely on attention and context to carry out “masking tasks” successfully.