Publications

Detailed Information

SmrtFridge: IoT-based, user interaction-driven food item & quantity sensing

Cited 0 time in Web of Science Cited 7 time in Scopus
Authors

Sharma, Amit; Misra, Archan; Subramaniam, Vengateswaran; Lee, Youngki

Issue Date
2019-11
Publisher
Association for Computing Machinery, Inc
Citation
SenSys 2019 - Proceedings of the 17th Conference on Embedded Networked Sensor Systems, pp.245-257
Abstract
© 2019 ACM.We present SmrtFridge, a consumer-grade smart fridge prototype that demonstrates two key capabilities: (a) identify the individual food items that users place in or remove from a fridge, and (b) estimate the residual quantity of food items inside a refrigerated container (opaque or transparent). Notably, both of these inferences are performed unobtrusively, without requiring any explicit user action or tagging of food objects. To achieve these capabilities, SmrtFridge uses a novel interaction-driven, multi-modal sensing pipeline, where Infrared (IR) and RGB video sensing, triggered whenever a user interacts naturally with the fridge, is used to extract a foreground visual image of the food item, which is then processed by a state-of-the-art DNN classifier. Concurrently, the residual food quantity is estimated by exploiting slight thermal differences, between the empty and filled portions of the container. Experimental studies, involving 12 users interacting naturally with 19 common food items and a commodity fridge, show that SmrtFridge is able to (a) extract at least 75% of a food item's image in over 97% of interaction episodes, and consequently identify the individual food items with precision/recall values of ~ 85%, and (b) perform robust coarse-grained (3 level) classification of the residual food quantity with an accuracy of ~ 75%.
URI
https://hdl.handle.net/10371/179320
DOI
https://doi.org/10.1145/3356250.3360028
Files in This Item:
There are no files associated with this item.
Appears in Collections:

Altmetrics

Item View & Download Count

  • mendeley

Items in S-Space are protected by copyright, with all rights reserved, unless otherwise indicated.

Share