A Look Inside the Lab: What Are the Biggest Open Research Questions in LLMs? What Projects Are You Working on Right Now?
Jan Plogsties: One key question is how to run AI models locally on small devices – on smartphones, in vehicles, or in industrial systems. More specialized accelerators and processors are entering the market, particularly for smartphones and embedded systems. This makes it possible to run LLMs locally – without the cloud and without latency.
However, these models then must operate with significantly fewer parameters and far less computing power. At Fraunhofer, we refer to as this “large models for small devices.” This is particularly exciting for interactive applications: for example, if I give a spoken command to a machine, the response must come almost in real time and cannot first be routed through a large cloud-based model. We are addressing this challenge, among other things, in the Bavarian project DSgenAI. With our model ELMOD, we have already demonstrated that a high-quality German language model can run locally on a smartphone.
Another exciting project we are currently working on is SOOFI. In this project, we are developing a family of language models with sizes of up to 100 billion parameters, together with some of Germany’s leading experts in LLM training. To ensure that the model ultimately delivers what is needed in real‑world applications, we involve industry partners who contribute their requirements directly. With SOOFI, we aim to help establish sovereign AI models in Germany and Europe.