Abstract
In RadarCat we present a small, versatile radar-based system for material and object classification which enables new forms of everyday proximate interaction with digital devices. We demonstrate that we can train and classify different types of materials and objects which we can then recognize in real time. Based on established research designs, we report on the results of three studies, first with 26 materials (including complex composite objects), next with 16 transparent materials (with different thickness and varying dyes) and finally 10 body parts from 6 participants. Both leave one-out and 10-fold cross-validation demonstrate that our approach of classification of radar signals using random forest classifier is robust and accurate. We further demonstrate four working examples including a physical object dictionary, painting and photo editing application, body shortcuts and automatic refill based on RadarCat. We conclude with a discussion of our results, limitations and outline future directions.
Original language | English |
---|---|
Title of host publication | Proceedings of the 29th Annual Symposium on User Interface Software and Technology (UIST '16) |
Publisher | ACM |
Pages | 833-841 |
Number of pages | 9 |
ISBN (Print) | 9781450341899 |
DOIs | |
Publication status | Published - 16 Oct 2016 |
Event | 29th ACM User Interface Software and Technology Symposium - Hitotsubashi Hall, National Center of Sciences Building, Tokyo, Japan Duration: 16 Oct 2016 → 19 Oct 2016 Conference number: 29 http://uist.acm.org/uist2016/ |
Conference
Conference | 29th ACM User Interface Software and Technology Symposium |
---|---|
Abbreviated title | UIST |
Country/Territory | Japan |
City | Tokyo |
Period | 16/10/16 → 19/10/16 |
Internet address |
Keywords
- Context-aware interaction
- Machine learning
- Material classification
- Object recognition
- Ubiquitous computing