Author
Listed:
- An Chen
(IBM Research – Almaden)
- Stefano Ambrogio
(IBM Research – Almaden)
- Pritish Narayanan
(IBM Research – Almaden)
- Atsuya Okazaki
(IBM Research – Tokyo)
- Charles Mackin
(IBM Research – Almaden)
- Andrea Fasoli
(IBM Research – Almaden)
- Malte J. Rasch
(IBM T. J. Watson Research Center – Yorktown Heights)
- Alexander Friz
(IBM Research – Almaden)
- Jose Luquin
(IBM Research – Almaden)
- Takeo Yasuda
(IBM Research – Tokyo)
- Masatoshi Ishii
(IBM Research – Tokyo)
- Takuto Kanamori
(IBM Research – Tokyo)
- Kohji Hosokawa
(IBM Research – Tokyo)
- Timothy Philicelli
(IBM Albany NanoTech – Albany)
- Seiji Munetoh
(IBM Research – Tokyo)
- Vijay Narayanan
(IBM T. J. Watson Research Center – Yorktown Heights)
- Hsinyu Tsai
(IBM Research – Almaden)
- Geoffrey W. Burr
(IBM Research – Almaden)
Abstract
A Lite Bidirectional Encoder Representations from Transformers model is demonstrated on an analog inference chip fabricated at 14nm node with phase change memory. The 7.1 million unique analog weights shared across 12 layers are mapped to a single chip, accurately programmed into the conductance of 28.3 million devices, for this first analog hardware demonstration of a meaningfully large Transformer model. The implemented model achieved near iso-accuracy on the General Language Understanding Evaluation benchmark of seven tasks, despite the presence of weight-programming errors, hardware imperfections, readout noise, and error propagation. The average hardware accuracy was only 1.8% below that of the floating-point reference, with several tasks at full iso-accuracy. Careful fine-tuning of model weights using hardware-aware techniques contributes an average hardware accuracy improvement of 4.4%. Accuracy loss due to conductance drift – measured to be roughly 5% over 30 days – was reduced to less than 1% with a recalibration-based “drift compensation” technique.
Suggested Citation
An Chen & Stefano Ambrogio & Pritish Narayanan & Atsuya Okazaki & Charles Mackin & Andrea Fasoli & Malte J. Rasch & Alexander Friz & Jose Luquin & Takeo Yasuda & Masatoshi Ishii & Takuto Kanamori & Ko, 2025.
"Demonstration of transformer-based ALBERT model on a 14nm analog AI inference chip,"
Nature Communications, Nature, vol. 16(1), pages 1-11, December.
Handle:
RePEc:nat:natcom:v:16:y:2025:i:1:d:10.1038_s41467-025-63794-4
DOI: 10.1038/s41467-025-63794-4
Download full text from publisher
Corrections
All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:nat:natcom:v:16:y:2025:i:1:d:10.1038_s41467-025-63794-4. See general information about how to correct material in RePEc.
If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.
We have no bibliographic references for this item. You can help adding them by using this form .
If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.
For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: Sonal Shukla or Springer Nature Abstracting and Indexing (email available below). General contact details of provider: http://www.nature.com .
Please note that corrections may take a couple of weeks to filter through
the various RePEc services.