[Paper] AME: An Efficient Heterogeneous Agentic Memory Engine for Smartphones

Published: (November 24, 2025 at 10:03 AM EST)
2 min read
Source: arXiv

Source: arXiv

Abstract

On‑device agents on smartphones increasingly require continuously evolving memory to support personalized, context‑aware, and long‑term behaviors. To meet both privacy and responsiveness demands, user data is embedded as vectors and stored in a vector database for fast similarity search. However, most existing vector databases target server‑class environments. When ported directly to smartphones, two gaps emerge:

  1. (G1) Hardware mismatch – mobile SoC constraints differ from vector‑database assumptions, including tight bandwidth budgets, limited on‑chip memory, and stricter data‑type and layout constraints.
  2. (G2) Workload mismatch – on‑device usage resembles a continuously learning memory, where queries must coexist with frequent inserts, deletions, and ongoing index maintenance.

To address these challenges, we propose AME, an on‑device Agentic Memory Engine co‑designed with modern smartphone SoCs. AME introduces two key techniques:

  1. A hardware‑aware, high‑efficiency matrix pipeline that maximizes compute‑unit utilization and exploits multi‑level on‑chip storage to sustain high throughput.
  2. A hardware‑ and workload‑aware scheduling scheme that coordinates querying, insertion, and index rebuilding to minimize latency.

We implement AME on Snapdragon 8‑series SoCs and evaluate it on HotpotQA. In our experiments, AME improves query throughput by up to 1.4× at matched recall, achieves up to faster index construction, and delivers up to higher insertion throughput under concurrent query workloads.

Subjects

  • Distributed, Parallel, and Cluster Computing (cs.DC)

Citation

Submission History

  • v1 – Qingyu Ma, Mon, 24 Nov 2025 15:03:06 UTC (621 KB)
Back to Blog

Related posts

Read more »

It’s code red for ChatGPT

A smidge over three years ago, OpenAI threw the rest of the tech industry into chaos. When ChatGPT launched, even billed as a 'low-key research preview,' it bec...