Towards edge-caching for image recognition

With the available sensors on mobile devices and their improved CPU and storage capability, users expect their devices to recognize the surrounding environment and to provide relevant information and/or content automatically and immediately. For such classes of real-time applications, user perceptio...

Full description

Saved in:
Bibliographic Details
Published in:2017 IEEE International Conference on Pervasive Computing and Communications Workshops (PerCom Workshops) pp. 593 - 598
Main Authors: Drolia, Utsav, Guo, Katherine, Jiaqi Tan, Gandhi, Rajeev, Narasimhan, Priya
Format: Conference Proceeding
Language:English
Published: IEEE 01-03-2017
Subjects:
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:With the available sensors on mobile devices and their improved CPU and storage capability, users expect their devices to recognize the surrounding environment and to provide relevant information and/or content automatically and immediately. For such classes of real-time applications, user perception of performance is key. To enable a truly seamless experience for the user, responses to requests need to be provided with minimal user-perceived latency. Current state-of-the-art systems for these applications require offloading requests and data to the cloud. This paper proposes an approach to allow users' devices and their onboard applications to leverage resources closer to home, i.e., resources at the edge of the network. We propose to use edge-servers as specialized caches for image-recognition applications. We develop a detailed formula for the expected latency for such a cache that incorporates the effects of recognition algorithms' computation time and accuracy. We show that, counter-intuitively, large cache sizes can lead to higher latencies. To the best of our knowledge, this is the first work that models edge-servers as caches for compute-intensive recognition applications.
DOI:10.1109/PERCOMW.2017.7917629