After the revolutionary introduction of LLMs, there is growing interest in Small Language Models (SLMs). These are models with up to 10 billion parameters, in contrast with LLMs, which can have hundreds of billions or even trillions of parameters.
Training and using LLMs requires enormous amounts of computational resources. In contrast, SLMs are significantly smaller. Therefore, they are less intensive in terms of data processing, hardware, and training time requirements. SLMs also consume less energy, making them more suitable for applications on smaller devices.
SLMs are more accessible to users who want to train and run these models on consumer hardware at the edge of a network, especially for single-purpose devices (e.g. sensors). In addition, SLMs are particularly useful for specific tasks rather than for use as general-purpose tools.