web hit counter

PDigit's AI PORTFOLIO

PDigit

Blending AI efficiency with human interpretation

EDGE AI Embedded Computing

Example of an M2 TPU Processing unit
Example of an M2 TPU Processing unit

See also Small LLM and TinyML


The Evolution of Edge Computing: Fine-Tuning AI Models in Real-Time

NOTE: We here refer to Edge AI in an embedded context The term “Edge Computing” can also refer to distributed computing but has quite a different meaning. See Wikipedia for the “distributed” definition.

Edge Computing represents a significant leap in data processing technology, particularly in its ability to fine-tune AI models in real-time. This process involves making incremental adjustments to AI models, allowing them to adapt more effectively to specific tasks. Crucial in dynamic environments, Edge Computing enables AI models to continuously learn and evolve, contrasting with traditional fixed models

This technology is transformative, bringing data processing closer to its source, thus enhancing efficiency and leading to smarter technological solutions. Key benefits include reduced latency, improved data management, and enhanced security
Practical examples (2024/2025) include NVIDIA Jetson Nano-Orin, axelera.ai Metis, Raspberry Pi 5 + Pi AI HAT/HAT+, Google Coral TPUs and more announced, showcasing Edge Computing’s versatility, STM32 AI (* and many new chips coming out continuously)

References

(external Links)

Advantages of Edge AI

Reduced Latency and Improved Speed

By processing data close to its source, Edge Computing minimizes latency, essential for real-time applications like autonomous vehicles and smart city infrastructures.

Enhanced Data Management and Privacy

Local data processing reduces reliance on cloud storage, enhancing data privacy and meeting data protection regulations.

Scalability and Cost-Effectiveness

It allows for efficient scalability without major infrastructure overhauls, exemplified by specialized hardware like the Axelera Metis TPU, NVIDIA Jetso Orin*

Resilience and Continuous Operations

Its independence from central servers ensures continuous operation, even in unstable environments.

Benchmarking and Profiling Performances

HW Benchmarking
Benchmarking chip performances on AI tasks

Typ Measurements

  • TOPS, TOPS/W
  • Inference Latency and accuracy on common (open src) models
  • FPS

See also - Edge Benchmarking Projects


Small LLM and TinyML

In these scenarios, small LLMs offer a balance between the advanced capabilities of AI language processing and the constraints of specific applications, environments, or resources. They represent a tailored approach, where the size and complexity of the model are matched to the specific needs and limitations of the use case.

TinyML typical Applications

Embedded Applications (Small LLM & Edge Computing) Summary

Edge Computing has diverse industry applications, from healthcare’s real-time patient monitoring to predictive maintenance in manufacturing and enhanced personalized on-site services, as well as remote monitoring stations.