Itinai.com llm large language model graph clusters multidimen f45b3cbc 46c3 4e70 9028 e654e8394d2d 2
Itinai.com llm large language model graph clusters multidimen f45b3cbc 46c3 4e70 9028 e654e8394d2d 2

IBM and ETH Zürich Develop Analog Foundation Models to Enhance In-Memory AI Hardware Performance

Overview of Analog Foundation Models

IBM researchers, in collaboration with ETH Zürich, have introduced a new class of Analog Foundation Models (AFMs) aimed at addressing the noise issues inherent in Analog In-Memory Computing (AIMC) hardware. AIMC has the potential to significantly enhance efficiency by enabling the execution of models with a billion parameters in a compact footprint suitable for embedded or edge devices. However, noise has been a critical barrier, as matrix-vector multiplications performed directly within non-volatile memory (NVM) devices often result in non-deterministic errors that hinder the performance of existing models.

The Importance of Analog Computing for LLMs

Unlike traditional computing methods using GPUs or TPUs, AIMC performs matrix-vector multiplications directly within memory arrays, eliminating the von Neumann bottleneck and significantly improving throughput and power efficiency. Previous studies indicated that combining AIMC with 3D NVM and Mixture-of-Experts (MoE) architectures could theoretically support trillion-parameter models on compact accelerators, making large-scale AI feasible beyond data centers.

Challenges in Implementing AIMC

The primary challenge in utilizing AIMC is the presence of noise. AIMC computations are affected by device variability, DAC/ADC quantization, and runtime fluctuations, which can degrade model accuracy. Unlike quantization on GPUs, where errors are predictable, analog noise is stochastic and unpredictable. While earlier research adapted smaller networks like CNNs and RNNs (less than 100M parameters) to tolerate such noise, LLMs with billions of parameters have struggled under AIMC constraints.

Addressing Noise with Analog Foundation Models

The IBM team has developed AFMs that incorporate hardware-aware training to prepare LLMs for analog execution. Their training pipeline includes:

  • Noise injection during training to simulate AIMC randomness.
  • Iterative weight clipping to stabilize distributions within device limits.
  • Learned static input/output quantization ranges aligned with real hardware constraints.
  • Distillation from pre-trained LLMs using 20B tokens of synthetic data.

These methods, implemented with AIHWKIT-Lightning, enable models like Phi-3-mini-4k-instruct and Llama-3.2-1B-Instruct to maintain performance comparable to weight-quantized 4-bit / activation 8-bit baselines under analog noise. Evaluations across reasoning and factual benchmarks indicate that AFMs outperform both quantization-aware training (QAT) and post-training quantization (SpinQuant).

Compatibility with Digital Hardware

Interestingly, AFMs also demonstrate strong performance on low-precision digital hardware. Because AFMs are trained to withstand noise and clipping, they manage simple post-training round-to-nearest (RTN) quantization more effectively than existing methods. This adaptability makes them valuable not only for AIMC accelerators but also for standard digital inference hardware.

Scalability of Performance

Yes, performance can scale with increased compute at inference time. Researchers tested compute scaling on the MATH-500 benchmark, generating multiple answers per query and selecting the best using a reward model. AFMs exhibited better scaling behavior than QAT models, with accuracy gaps diminishing as more inference compute was allocated. This aligns with AIMC’s strengths in low-power, high-throughput inference rather than training.

Future Implications for AIMC

This research represents the first systematic demonstration that large LLMs can be adapted to AIMC hardware without significant accuracy loss. While training AFMs is resource-intensive and reasoning tasks like GSM8K still reveal accuracy gaps, the findings mark a significant milestone. The combination of energy efficiency, robustness to noise, and compatibility with digital hardware positions AFMs as a promising avenue for scaling foundation models beyond the limitations of GPU technology.

Conclusion

In summary, the introduction of Analog Foundation Models by IBM and ETH Zürich offers a groundbreaking approach to overcoming the challenges posed by noise in analog computing. By enhancing the performance of large language models on compact hardware, these innovations pave the way for more efficient AI solutions in various applications. As the technology matures, it holds the potential to transform how we approach AI, making it more accessible and effective across different platforms.

FAQ

  • What are Analog Foundation Models? Analog Foundation Models are a new class of AI models designed to operate efficiently in Analog In-Memory Computing environments, addressing noise issues that affect model accuracy.
  • How do AFMs improve model performance? AFMs utilize hardware-aware training techniques that prepare models for the unique challenges of analog execution, such as noise and variability.
  • What are the advantages of AIMC over traditional computing methods? AIMC eliminates the von Neumann bottleneck, improving throughput and power efficiency, making it suitable for large-scale AI applications.
  • Can AFMs be used with digital hardware? Yes, AFMs are compatible with low-precision digital hardware and can perform effectively even under standard digital inference conditions.
  • What future implications do AFMs have for AI technology? AFMs represent a significant advancement in scaling AI models, potentially leading to more energy-efficient and robust AI solutions that can operate beyond the limitations of current GPU technology.
Itinai.com office ai background high tech quantum computing 0002ba7c e3d6 4fd7 abd6 cfe4e5f08aeb 0

Vladimir Dyachkov, Ph.D
Editor-in-Chief itinai.com

I believe that AI is only as powerful as the human insight guiding it.

Unleash Your Creative Potential with AI Agents

Competitors are already using AI Agents

Business Problems We Solve

  • Automation of internal processes.
  • Optimizing AI costs without huge budgets.
  • Training staff, developing custom courses for business needs
  • Integrating AI into client work, automating first lines of contact

Large and Medium Businesses

Startups

Offline Business

100% of clients report increased productivity and reduced operati

AI news and solutions