Leading the Way with Embedded AI

Devices and systems that gather and process data locally are the key to greater efficiency, precision and safety in AI-based applications. Embedded artificial intelligence (Embedded AI) is capable of continuously analyzing data streams, generating forecasts and diagnoses, making decisions and triggering appropriate actions. In this way, it automatically monitors and controls the smallest details or complex processes.

The ingredients for a successful Embedded AI product: a lean AI model and efficient hardware – both ideally tailored to the task to be solved. What truly perfects the solution, however, is a system integration process that targets the optimal interplay between algorithms and hardware while keeping the overall system requirements of the end product in mind.

AI without the cloud: intelligence moves closer to the action

Embedded AI processes data directly where it is generated – in the end devices at the edge. It therefore works completely locally and does not require wireless transmission of sensor data to the cloud. The usual data traffic to cloud servers is eliminated and the result is impressive:

Best energy efficiency

Energy for data transmission and server operation is no longer required. Specialized Edge AI hardware architectures and energy-optimized neural networks reduce the power consumption of AI to a minimum – enabling applications in the microwatt range.

Less latency

Transmission delays caused by poor network coverage or low data rates vanish with Embedded AI. At the same time, the local AI execution boosts system response times – ideal for time-critical applications.

More security

On-device processing prevents data leaks, as sensor, audio and video data stay out of the cloud. Users and operators keep full control over their data, while the risk of data and devices being manipulated is drastically reduced.

Optimally combining algorithms and hardware

When implementing Embedded AI solutions, we concentrate on perfectly aligning the AI model with the chosen hardware. Using these technologies, we make Edge AI more efficient and powerful:

NAS

With our toolset for neural architecture search (NAS), we automate the search for the optimal architecture of neural networks and thus accelerate the development of your Embedded AI products.

Different network architectures are systematically generated in a defined search space and compared in terms of accuracy, computational complexity and model size until the optimal configuration is found.

Our search process is "hardware-aware", delivering an optimized neural network that is ideally tailored to your hardware.

Hardware-aware training

With our hardware-aware training approach, we optimize neural networks even further for use on a specific hardware platform.

We simulate typical hardware effects such as quantization, noise, distortion and other effects of analog hardware or deliberately injected errors already during training.

The result: robust models that retain their accuracy even under real operating conditions.

Distributed AI

When more compute power is required, we spread artificial intelligence across many devices. With Distributed AI, we split a neural network into several sub-models that run in parallel on multiple IoT or edge devices.

In this way, we bundle compute power and enable complex inference – directly at the edge of the network and without any loss of accuracy.

Self-organized ad-hoc networks and orchestrated pipeline architectures ensure reliable data exchange between the devices, reduce latencies and increase both throughput and reliability.

Your path to the ideal software-hardware symbiosis for Embedded AI

Initiation, Conception, Realization
© Fraunhofer IIS
Workshop for needs analysis Specification and design Implementation, deployment, testing and benchmarking
  • Gaining certainty about application possibilities and feasibility
  • Identifying benefits and potential for improvement
  • Establishing a shared understanding of the final product as basis for technology selection
  • Defining measurable goals (KPIs)
  • Designing the Embedded AI solution
  • Creating an implementation roadmap
  • AI model and prototype development
  • Trial installation in the lab and in the operational environment
  • Validating target KPIs

AI in the device, added value in the field

Reliable efficiency for networks, factories, vehicles and more

Communication systems

Embedded AI-optimized transceiver systems process data packets and analyze signal quality. When necessary, they adapt the transmission and thus improve latency, load and resilience to interference.

Industrial automation

Sensor-based Embedded AI devices analyze machine data directly on site, detect wear, impending failures or quality losses early and reduce downtime.

Robotics

With Embedded AI, autonomous mobile robots process image and sensor data locally to identify obstacles and adjust motion in real time – enabling collision-free, efficient operation without cloud dependency.

Automotive

In-vehicle Edge AI modules process camera, radar and lidar signals. They detect pedestrians, traffic signs and dangerous situations and support assistance systems with minimal latency.

Digital health

Wearables and medical devices use Edge AI for on-device analysis of vital parameters. They detect irregular heart rhythms or breathing and enable privacy-compliant, proactive health monitoring.

Reference projects

  • 6G LINO – 6G Laboratory In Orbit

    In the 6G LINO project, Fraunhofer IIS is working with partners from the European satellite and mobile communications industry to develop a test infrastructure for testing 6G mobile communications via a LEO satellite. The project, which is funded by the European Space Agency (ESA), involves setting up a complete transmission chain from the ground station via a satellite in CubeSat format to the receiver on the ground.

    Various features are being implemented and demonstrated in the project, including a mobile base station (gNB) on the satellite, handover between terrestrial and non-terrestrial network, options for testing future 6G waveform adaptations and spectrum sensing with AI. To this end, an AI-based solution will be integrated on the satellite to examine the utilization of frequencies in orbit, for example to optimize the use of the spectrum (dynamic satellite-terrestrial spectrum management and sharing in S-band). Fraunhofer IIS is working on implementing the planned functions and preparing everything for their deployment on the satellite.

  • GAIA Initiative – Guardian of the Wild using Artificial Intelligence Applications

    As part of the GAIA Initiative, Fraunhofer IIS and its partners have developed a high-tech early warning system that uses animal transmitters to detect ecological changes and critical events in the environment. The initiative comprises the two sub-projects GAIA-Sat-IoT and SyNaKI. They were funded by the Federal Ministry for Economic Affairs and Climate Action on the basis of a decision by the German Bundestag.

    In the GAIA-Sat-IoT project, intelligent camera tags with integrated AI and satellite-based IoT communication were developed to collect ecological data and for the early detection of wildlife diseases.

    In the SyNaKI project, Distriubted AI was used to create virtual swarm intelligence based on the interaction of animals and microprocessors in digital networks. 

  • KI-FLEX – Reconfigurable Hardware Platform for AI-based Sensor Data Processing for Autonomous Driving

    In the KI-FLEX project, funded by the German Federal Ministry of Education and Research, eight project partners developed a high-performance, energy-efficient hardware platform and the associated software framework for autonomous driving. The KI-FLEX platform is designed to reliably and quickly process and merge data from laser, camera and radar sensors in the car. Artificial intelligence (AI) methods are used for this purpose. The vehicle thus always has an accurate picture of the actual traffic conditions, can locate its own position in this environment, and on the basis of this information, make the right decision in every driving situation.

    Fraunhofer IIS's contribution was the development of a flexible DLI accelerator core for the multi-core deep learning accelerator, which is integrated together with other DLI accelerators into a flexible, future-proof ASIC. The architecture of the ASIC is designed in such a way that future improvements of neural network (NN)architectures, i.e. newly emerging NN types and concepts, can still be realized with it. For this purpose, critical areas were specifically designed to be reconfigurable in order to build a bridge from the rigidity of an ASIC to the flexibility of an FPGA.

  • LODRIC – Low-Power Digital Deep Learning Inference Chip

    The LODRIC project, funded by the Federal Ministry of Education and Research, extended the successful collaboration of the consortium, consisting of Friedrich-Alexander-Universität Erlangen-Nürnberg (FAU) and Fraunhofer IIS, from the previous project Lo3-ML.

    Its continuation is about the development of a design methodology for low-power digital AI chips with embedded non-volatile memory elements and its prototypical application on the basis of three different applications. Thereby, the main innovation of the project Lo3-ML, namely the development of data flow-oriented computer architectures in combination with distributed, non-volatile weight memory and strongly (ternary) quantized weights shall be taken up and specifically methodically developed further.

    Fraunhofer IIS is represented with three disciplines: medical technology, digital circuit design and Embedded AI. The latter will expand its competencies in the area of hardware-aware training. In this context, a tool chain specific to accelerator technology will be further developed, which on the one hand achieves a significant reduction (optimization) of the neural network and on the other hand maintains its accuracy despite high quantization of the neuron weights through iterative retraining.

  • MANOLO – Trustworthy Efficient AI for Cloud-Edge Computing

    The vision of MANOLO is to deliver a complete and trustworthy stack of algorithms and tools to help AI systems reach better efficiency and seamless optimization in their operations. The focus is on the energy-efficient training of AI models with quality-checked data and the execution of resource-efficient AI models on a wide range of devices for use on the edge and in the cloud. MANOLO is funded by the Horizon Europe framework programme of the European Union.

    In the project, Fraunhofer IIS is focusing on bringing AI applications to the edge. To this end, algorithms and tools are being developed that search for and optimize suitable neural networks automatically (Neural Architecture Search, NAS).

You may also be interested in

 

Spiking neural networks

The next generation of neural networks

 

Neuromorphic computing

Energy-efficient smart chip design

 

Artificial intelligence

Go back to business area "Artificial intelligence"