..., storage, and precision. Its GPU-and EdgeTPU-accelerated inference delivers embeddings in milliseconds, under 15 ms for 256 tokens on EdgeTPU, while quantization-aware training keeps memory usage under 200 MB without compromising quality. This makes it ideal for real-time, on-device tasks such as semantic search, retrieval-augmented generation (RAG), classification, clustering, and similarity detection, whether for personal file search, mobile chatbots, or custom domain use.