Article

Designing efficient and reliable power architectures for edge AI

Nishant Nishant
circuit board with AI chip
Board-level power is a design constraint for edge AI applications
KEY TAKEAWAYS:
  • Practical power architecture for AI on MCUs
  • Budgeting for worst-case inference workloads
  • Compare discrete versus module-based approaches
  • Wide bandgap’s role in achieving high-efficiency conversion

AI at the edge can take the form of battery-powered sensor nodes running compact models on a modest microcontroller (MCU), to powerful gateways with multiple processors running inferences at trillions of operations per second (TOPS) to capture and analyze high-resolution video.

In simple applications, an MCU may only wake periodically to execute a few thousand multiply–accumulate operations before returning to a deep sleep mode to preserve energy. Traditionally, these systems can run for months or even years on primary cell batteries, but adding AI inferencing will place new demands on the system’s operational profile. Understanding those demands will help engineers estimate the batteries’ usable lifetime.

More complex applications may include a vision accelerator, CPU, memory subsystem and networking stack, operating continuously inside an enclosure in a harsh industrial environment. These systems need a constant supply of reliable energy. Inferencing workloads will be dependent on the scene being analyzed, introducing variability and uncertainty.

Across this spectrum, two trends are clear. First, AI inference is increasingly duty cycle intensive rather than occasional. Second, model complexity and input data rates are rising faster than the mechanical envelope and cooling budgets.

As a result, efficiency in power delivery and thermal management are now critical to system performance, perhaps more critical than computing throughput or memory bandwidth.

Designing edge AI power domains

MCU-centric designs operate under challenging constraints. Battery life, quiescent current and cost dominate, while peak power remains relatively modest. A common approach includes:

  • A high-efficiency buck or buck-boost converter for primary cell or Li-ion battery power.
  • One or more downstream rails for the MCU, radio and sensors.
  • Optional on-chip DC-to-DC conversion within the microcontroller.

Low-dropout regulators are typically reserved for noise-sensitive domains, such as RF or analog front ends, rather than primary power conversion, due to their lower efficiency.

When AI functions are introduced, such as anomaly detection or basic vision processing, the overall hardware platform may remain similar in simple cases, but the power profile changes significantly. Engineers can expect burst current to increase during inference, which can expose weaknesses in decoupling, grounding or regulator response. Ensuring that the power tree can support these bursts without degrading RF sensitivity or sensor accuracy is essential.

Power budgeting for worst-case AI workloads

Datasheets for incumbent commodity components are unlikely to have been updated to include thermal power and current figures for real AI workloads. For connected devices running AI at the edge, engineers must now factor in software updates to deploy patches and evolving models. This can present entirely new power profiles, so it is advisable to base designs on measured worst-case scenarios.

  • Define the most demanding expected workload, including input resolution, model size, concurrency and data rates.
  • Measure rail-by-rail current under elevated ambient conditions, capturing both steady state and transient behavior.
  • Add margin for process variation, component tolerances, and future model growth. This is often 20 to 30 percent above measured peaks for critical rails.

Thermal analysis should proceed in parallel. While wide bandgap devices can tolerate higher junction temperatures, system-level constraints such as PCB materials, magnetics and nearby components will limit overall thermal performance. Validating simulations with real hardware under worst-case workloads helps prevent issues such as thermal throttling or derating in environments with high ambient temperatures.

Discrete vs module power design

As power density increases, the choice between discrete power stages and integrated modules becomes more important. Both approaches can leverage GaN devices for improved efficiency and switching performance, but they differ in design complexity, flexibility and time to market.

Table 1: Key trade offs for edge AI power stages

Consideration Discrete WBG Module-based WBG
Power range Suitable for applications from tens of watts up. Scalable via paralleling when layout is carefully controlled. Optimized for medium to high power with tightly integrated half- or full-bridge stages.
Design flexibility High flexibility in topology, component selection and mechanical placement. Good for platform designs. Fixed internal topology and pinout, which speeds development but limits customization.
Layout and EMI Requires design expertise to manage parasitics, ringing and EMI at high edge rates. Vendor-engineered layout reduces parasitics, easing EMI compliance and shortening validation.
Thermal behavior Large footprint can aid heat spreading. Easier to adapt to different heatsink schemes. Higher power density. May demand more advanced cooling but offers well characterized thermal paths.
Cost and time to market Lower device cost but higher engineering and verification effort. Best for experienced power teams and higher volumes. Higher unit cost offset by shorter design cycles. Attractive when schedules are tight or volumes moderate.

For many AI-heavy gateways, a hybrid approach is effective. A power module at the front end can reduce design risk and accelerate development. Discrete multiphase point-of-load stages can then be used to optimize for specific load requirements.

System-level power and thermal design for edge AI

The most successful edge AI platforms are those where power architecture, thermal design and workload planning evolve together. Power telemetry, when available, can inform thermal management strategies and even influence how and when models are executed.

Firmware can further optimize system behavior by using techniques such as dynamic voltage and frequency scaling, workload scheduling and aggressive low-power states. This requires that the underlying power system remains stable across these transitions.

By starting with realistic worst-case inference scenarios, selecting appropriate power architectures for both accelerators and MCUs, and choosing the right mix of discrete and module-based approaches, design teams can maximize usable AI performance within fixed thermal and mechanical constraints. The result is longer system life, improved reliability and greater headroom for future model complexity.

 

About Author

Nishant Nishant
Avnet Staff

We use Avnet Staff as a collective byline when our team of editors and writers collaborate on the co...

Helpful Links

Marketing Content Spots

Related Articles

Related Articles
person holding AI info in hands
AI coding assistants in the embedded domain
By Philip Ling   -   February 26, 2026
More engineers are being exposed to AI agents designed to help them be more productive. The technology is rapidly catching up with expectations.
Industrial automation using analog sensor data for AI systems
Optimized analog front-end design for edge AI
By Avnet Staff   -   November 21, 2025
Connected applications are in transition, from the internet of things (IoT) to the artificial intelligent internet of things (AIoT). Real-world data is now doing double-duty. How does that impact analog design?

Related Events

Related Events
What, Why & How of Vision AI at the Edge
Date: April 23, 2021
Location: On Demand

designing-efficient-and-reliable-power-architectures-for-edge-ai