{"id":988,"date":"2025-03-02T23:14:49","date_gmt":"2025-03-02T23:14:49","guid":{"rendered":"https:\/\/mikewinters.io\/?p=988"},"modified":"2025-03-02T23:17:57","modified_gmt":"2025-03-02T23:17:57","slug":"whispering-wearables","status":"publish","type":"post","link":"https:\/\/mikewinters.io\/?p=988","title":{"rendered":"Whispering Wearables"},"content":{"rendered":"\n<p>Over the past few years, I\u2019ve been exploring the intersection of sound technologies, sensor data, and accessible design\u2014a journey that began with my early work on sonification and tongue gestures. One project, in particular, has left a lasting impression on me: our Silent Speech initiative. This work not only pushed the boundaries of non-invasive, hands-free interaction but also provided a fresh perspective on how subtle sensor signals can be transformed into meaningful communication.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">Silent Speech Research<\/h2>\n\n\n\n<p>During my time as a Post Doc in the Brain-Computer Interface group at Microsoft Research, I had the opportunity to lead a grant that brought together product and research teams. This collaborative effort not only paved the way for innovative breakthroughs but also funded a PhD internship through the Microsoft Research internship program. One of the key contributors was PhD intern <a href=\"https:\/\/sites.google.com\/view\/tanmaysrivastava\/home\">Tanmay Srivastava<\/a>, who focused on leveraging IMU-based silent speech detection. His work was instrumental in capturing and classifying the subtle neuromuscular signals associated with silent, closed-mouth articulations.<\/p>\n\n\n\n<p>Our system was built using off-the-shelf sensors\u2014IMUs, EEG, and PPG\u2014to capture these signals and translate them into commands. While the technical details of sensor fusion and gesture recognition are described in our publication, what I find most fascinating is how this work extends the idea of using sensor data to produce sound\u2014a concept I\u2019ve explored before in my sonification projects.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">Integrating with LLMs: A Quiet Revolution<\/h2>\n\n\n\n<p>One of the most memorable aspects of the project was our integration with GPT-4. At a time when access to GPT-4 was highly restricted\u2014requiring special APIs\u2014we managed to connect our silent speech interface to this powerful language model. The concept was both straightforward and transformative: enable users to interact with a large language model through silent articulation. In essence, you could \u201cspeak\u201d to the system and receive context-aware responses without producing a single audible word.<\/p>\n\n\n\n<p>This integration wasn\u2019t merely a technical milestone; it also signaled a new way of thinking about communication\u2014one that is discreet, accessible, and profoundly inclusive.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">Why it Matters<\/h2>\n\n\n\n<p>At its core, the Silent Speech project is about more than just pushing technological boundaries\u2014it\u2019s about creating a meaningful impact on everyday communication. By transforming minimal sensor readings into actionable commands, we open up new channels for interaction in environments where traditional voice interfaces may fall short. Whether in noisy public spaces, quiet settings, or for individuals with speech impairments, this technology has the potential to offer a truly accessible alternative.<\/p>\n\n\n\n<p>For me, this work resonates with my broader interests in novel sound technologies and accessible design. Just as my earlier projects on sonification sought to make data audible in innovative ways, this initiative reimagines communication\u2014allowing technology to be both intuitive and inclusive. It demonstrates that by harnessing the quiet potential of our bodies, we can bridge gaps between technology and human need in unexpected and powerful ways.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">Final Thoughts<\/h2>\n\n\n\n<p>Reflecting on the journey of the Silent Speech project, I am struck by the collaborative spirit that drove its success. From the grant that united product and research teams to the contributions of talented individuals like Tanmay Srivastava, every step of this process reaffirmed my belief that innovation thrives at the intersection of diverse expertise and shared vision.<\/p>\n\n\n\n<p>Integrating our silent speech interface with GPT-4 was not just a technical achievement\u2014it was a conceptual leap that reshaped how I think about communication. In a world where speaking aloud isn\u2019t always possible or desirable, enabling silent, thoughtful interaction opens up countless possibilities. As I continue to explore new frontiers in sound, sensor data, and inclusive design, I remain excited about the future. This project is a testament to the idea that sometimes, the most profound breakthroughs come not from loud proclamations, but from quiet revolutions.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">Publications<\/h2>\n\n\n\n<p>Srivastava, T., Winters, R. M., Gable, T., Wang, Y. T.; LaScala, T.; Tashev, I. J. (2024). <a href=\"https:\/\/mikewinters.io\/wp-content\/uploads\/2025\/02\/Whispering-Wearables-ICMI2025.pdf\">Whispering Wearables: Multimodal Approach to Silent Speech Recognition with Head-Worn Devices<\/a>. <em>ICMI &#8217;24: International Conference on Multimodal Interaction, 214-223<\/em>.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>Over the past few years, I\u2019ve been exploring the intersection of sound technologies, sensor data, and accessible design\u2014a journey that began with my early work on sonification and tongue gestures. One project, in particular, has left a lasting impression on me: our Silent Speech initiative. This work not only pushed the boundaries of non-invasive, hands-free [&hellip;]<\/p>\n","protected":false},"author":1,"featured_media":992,"comment_status":"closed","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[23,38,21,19,15],"tags":[46,42,47,41,43],"class_list":["post-988","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-accessibility","category-brain-computer-interface","category-industry","category-leadership","category-teaching","tag-accessibility","tag-grants","tag-hands-free","tag-multimodal","tag-publication","has-thumbnail"],"_links":{"self":[{"href":"https:\/\/mikewinters.io\/index.php?rest_route=\/wp\/v2\/posts\/988","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/mikewinters.io\/index.php?rest_route=\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/mikewinters.io\/index.php?rest_route=\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/mikewinters.io\/index.php?rest_route=\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/mikewinters.io\/index.php?rest_route=%2Fwp%2Fv2%2Fcomments&post=988"}],"version-history":[{"count":3,"href":"https:\/\/mikewinters.io\/index.php?rest_route=\/wp\/v2\/posts\/988\/revisions"}],"predecessor-version":[{"id":991,"href":"https:\/\/mikewinters.io\/index.php?rest_route=\/wp\/v2\/posts\/988\/revisions\/991"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/mikewinters.io\/index.php?rest_route=\/wp\/v2\/media\/992"}],"wp:attachment":[{"href":"https:\/\/mikewinters.io\/index.php?rest_route=%2Fwp%2Fv2%2Fmedia&parent=988"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/mikewinters.io\/index.php?rest_route=%2Fwp%2Fv2%2Fcategories&post=988"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/mikewinters.io\/index.php?rest_route=%2Fwp%2Fv2%2Ftags&post=988"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}