Large Language Models (LLMs) have achieved very good success in natural language processing, but deployment of these powerful models on edge computing devices across all domains presents unique challenges. This paper reviews the state of LLMs in edge computing, focusing on four key aspects: their emerging applications across various sectors, the technical challenges of running LLMs on resource-constrained edge devices, the potential benefits of bringing LLM capabilities closer to data sources, and effective deployment strategies to enable LLMs at the edge. We also discuss on how LLM edge deployment could offer low-latency, privacy-preserving intelligent assistance throughout a range of domains, such as healthcare, IoT, industrial automation, and more. We also look at some techniques and architectures that can overcome the limitations of edge devices, such as cloud-edge collaboration, federated learning, model compression, and on-device inference. This review identifies practical ways to integrate LLMs into edge environments by examining current practices and their trade-offs. It also provides guidance for future research to address the remaining issues in this quickly expanding field.
Abstract views:
Downloads:
hh-index
Citations
inLibrary — is a scientific electronic library built on the paradigm of open science (Open Science), the main tasks of which are the popularization of science and scientific activities, public quality control of scientific publications, the development of interdisciplinary research, a modern institute of scientific review, increasing the citation of Uzbek science and building a knowledge infrastructure.
CONTACTS:
Republic of Uzbekistan, Tashkent, Parkent street 51, floor 2