Understanding Large Language Models

Understanding Large Language Models

Sreerakuvandana Sreerakuvandana (Jain University (Deemed), India), Princy Pappachan (Department of Foreign Languages and Literature, Asia University, Taiwan), and Varsha Arya (Department of Business Administration, Asia University, Taiwan)
Copyright: © 2024 |Pages: 24
DOI: 10.4018/979-8-3693-3860-5.ch001
OnDemand:
(Individual Chapters)
Forthcoming
$37.50
No Current Special Offers
TOTAL SAVINGS: $37.50

Abstract

Large language models (LLMs) are a revolutionary development that allows machines to comprehend and produce text, similar to that of humans on a never-before-seen scale. This chapter examines the basic ideas underlying LLMs with an emphasis on their applications, training approaches, and architecture. Deep neural networks with billions of parameters are used by LLMs, such as the GPT-3 model, to capture complex linguistic patterns and contextual subtleties. Massive datasets, frequently drawn from a variety of online sources, are used in the training process to impart a thorough understanding of language. Consequently, LLMs show remarkable abilities in tasks like question answering, language translation, and text generation. Issues like bias, ethical issues, and interpretability thus become important concerns. So, this chapter outlines the main elements of LLMs, discusses their advantages, reviews current research, and addresses the ethical issues surrounding their application.
Chapter Preview

Complete Chapter List

Search this Book:
Reset