Usable and Ubiquitous Privacy-Aware Sensing Devices
dc.contributor.author | Iravantchi, Yasha | |
dc.date.accessioned | 2025-05-12T17:40:11Z | |
dc.date.available | 2025-05-12T17:40:11Z | |
dc.date.issued | 2025 | |
dc.date.submitted | 2025 | |
dc.identifier.uri | https://hdl.handle.net/2027.42/197264 | |
dc.description.abstract | The proliferation of smart devices is bringing us closer to a future where everyday objects can monitor users, anticipate their needs, and track health metrics. However, privacy concerns have significantly hindered the adoption of information-rich sensors, such as microphones and cameras, particularly in sensitive home environments like bedrooms and bathrooms---locations where critical self-care behaviors and health events, such as falls, are most likely to occur. To address these concerns, ubiquitous sensing technologies must be designed with principles from the usable privacy community, ensuring privacy guarantees that promote adoption and maximize real-world impact. This dissertation presents a Privacy by Design approach to sensor-level privacy across three key domains. First, I introduce privacy-preserving microphones that do not capture speech frequencies but instead leverage inaudible ultrasound to outperform traditional microphones in acoustic event recognition. These microphones can also transform captured signals into locality sensitive hashes, ensuring that privacy-invasive raw audio cannot be reconstructed. In real-world deployments, these privacy-aware microphones demonstrate on-device detection of urinary voiding---an important kidney health metric---using only inaudible frequencies. Second, I explore privacy-preserving cameras that utilize thermal imaging to sanitize personally identifiable information on-device. This approach enables critical machine learning and computer vision applications, such as fall detection, without compromising performance. Finally, I present novel privacy-aware sensing techniques that inherently prevent the capture of sensitive information, such as airborne sound, by leveraging alternative signal sources like surface-acoustic waves. These methods enable activity recognition in the home without invasive data collection. Together, these contributions demonstrate how privacy-aware sensing can bridge the gap between user privacy and real-world sensing applications, paving the way for safer, more adoptable smart home technologies. | |
dc.language.iso | en_US | |
dc.subject | ubiquitous computing | |
dc.subject | privacy | |
dc.subject | sensing | |
dc.title | Usable and Ubiquitous Privacy-Aware Sensing Devices | |
dc.type | Thesis | |
dc.description.thesisdegreename | PhD | |
dc.description.thesisdegreediscipline | Computer Science & Engineering | |
dc.description.thesisdegreegrantor | University of Michigan, Horace H. Rackham School of Graduate Studies | |
dc.contributor.committeemember | Sample, Alanson | |
dc.contributor.committeemember | Kratz, Anna L | |
dc.contributor.committeemember | Abowd, Gregory | |
dc.contributor.committeemember | Banovic, Nikola | |
dc.contributor.committeemember | Shin, Kang Geun | |
dc.subject.hlbsecondlevel | Computer Science | |
dc.subject.hlbtoplevel | Engineering | |
dc.contributor.affiliationumcampus | Ann Arbor | |
dc.description.bitstreamurl | http://deepblue.lib.umich.edu/bitstream/2027.42/197264/1/yiravan_1.pdf | |
dc.identifier.doi | https://dx.doi.org/10.7302/25690 | |
dc.identifier.orcid | 0000-0001-5269-9579 | |
dc.identifier.name-orcid | Iravantchi, Yasha; 0000-0001-5269-9579 | en_US |
dc.working.doi | 10.7302/25690 | en |
dc.owningcollname | Dissertations and Theses (Ph.D. and Master's) |
Files in this item
Remediation of Harmful Language
The University of Michigan Library aims to describe its collections in a way that respects the people and communities who create, use, and are represented in them. We encourage you to Contact Us anonymously if you encounter harmful or problematic language in catalog records or finding aids. More information about our policies and practices is available at Remediation of Harmful Language.
Accessibility
If you are unable to use this file in its current format, please select the Contact Us link and we can modify it to make it more accessible to you.