Author
Listed:
- Dongqi Cai
(Beijing University of Posts and Telecommunications
University of Cambridge)
- Shangguang Wang
(Beijing University of Posts and Telecommunications)
- Chen Peng
(Beijing University of Posts and Telecommunications)
- Zeling Zhang
(Beijing University of Posts and Telecommunications)
- Zhenyan Lu
(Beijing University of Posts and Telecommunications
Pengcheng Laboratory)
- Tao Qi
(Beijing University of Posts and Telecommunications)
- Nicholas D. Lane
(University of Cambridge
Flower Labs)
- Mengwei Xu
(Beijing University of Posts and Telecommunications)
Abstract
Forgetting is inevitable in human memory. Recently, multimodal embedding models have been proposed to vectorize multimodal reality into a unified embedding space. Once generated, these embeddings allow mobile users to quickly retrieve relevant information, effectively augmenting their memory. However, as the model’s capacity increases, its resource consumption also rises. The resulting slow throughput and significant computational resource requirements hinder its deployment on mobile devices. In this paper, we present Reminisce, an efficient on-device multimodal embedding system that enables high-throughput embedding and precise retrieval on resource-constrained mobile devices. The core design draws inspiration from the memory functions of the human brain, utilizing coarse-grained embeddings to identify likely candidates, which are then refined through query-driven fine-grained retrieval. A series of algorithm-hardware orchestrated optimizations automatically navigates this process and strengthen the embedding quality. Experiments show that Reminisce provides high-quality embedding representation with high throughput while operating silently in the background with negligible memory usage and reduced energy consumption.
Suggested Citation
Dongqi Cai & Shangguang Wang & Chen Peng & Zeling Zhang & Zhenyan Lu & Tao Qi & Nicholas D. Lane & Mengwei Xu, 2025.
"Ubiquitous memory augmentation via mobile multimodal embedding system,"
Nature Communications, Nature, vol. 16(1), pages 1-12, December.
Handle:
RePEc:nat:natcom:v:16:y:2025:i:1:d:10.1038_s41467-025-60802-5
DOI: 10.1038/s41467-025-60802-5
Download full text from publisher
Corrections
All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:nat:natcom:v:16:y:2025:i:1:d:10.1038_s41467-025-60802-5. See general information about how to correct material in RePEc.
If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.
We have no bibliographic references for this item. You can help adding them by using this form .
If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.
For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: Sonal Shukla or Springer Nature Abstracting and Indexing (email available below). General contact details of provider: http://www.nature.com .
Please note that corrections may take a couple of weeks to filter through
the various RePEc services.