NSF
Award Abstract #2032125

RAPID: Human Sound Localization and Analytics

See grant description on NSF site

Program Manager:

Alexander Sprintson

Active Dates:

Awarded Amount:

$100,000

Investigator(s):

Lili Qiu

Awardee Organization:

University of Texas at Austin
Texas

Directorate

Computer and Information Science and Engineering (CISE)

Abstract:

COVID-19 is spreading at an unprecedented rate resulting in the death of so many people all over the world. Social distancing is so far the most effective method to limit its spread. However, manually enforcing social distancing is not only labor-intensive but also error-prone and even dangerous due to possible physical contact. This project proposes to develop techniques and mobile systems that localize human sound such as cough and voice and alarm a user when someone is within the social distance. If successful, this work will significantly advance the state-of-the-art in wireless sensing and localization. To maximize the impact, the researchers will collaborate with industry and local community and release software to the public. The research outcome will also be incorporated into the graduate and undergraduate curriculum. The proposed research aims to develop algorithms and systems to localize uncontrolled and unknown human sound. A unique advantage is that it does not require cooperation from other phones and whoever uses it can immediately benefit from it. It exploits the phone mobility as the user moves to enable localization. The multi-resolution analysis will be performed on low-frequency voice signals to further enhance accuracy. When another phone is cooperating, it will further leverage the time of flight between the two phones to improve the performance.This award reflects NSF's statutory mission and has been deemed worthy of support through evaluation using the Foundation's intellectual merit and broader impacts review criteria.

Back to Top