|Speaker :||Prof Jain|
|Old Dominion University|
|Time:||2:00 pm - 3:00 pm|
|Location:||LINCS Seminars room|
Modern cities are alive with sensors, such as smartphones, wearables, vehicles, and cameras. Realizing our plans for smart environments of the future necessitates a radical transformation in the way these devices perceive and interact with our highly mobile environments. Smart city services are dependent on large-scale and low-cost data acquisition. To achieve this, we rehash these city-wide devices in an attempt towards developing a mobility-aware multimodal sensing framework. Cities contain many moving objects and we show that being aware of this mobility enables efficient sensor reuse even when monitoring moving objects/people. Such a synergy has the potential to change the way our cities function; from optimizing transportation to provisioning city services, and further to legal and economic framework, and regulations. My research leverages existing ubiquitous devices and expands their role to innovate new services, ranging from large-scale video analytics to pedestrian safety. We innovate on smartphone sensors and fitness trackers for in-street detection. Through unobtrusive in-the-wild sensing we gauge pedestrian risk and enable city-scale pedestrian safety services. Infrastructure cameras, often mounted for surveillance and traffic monitoring, are among other prevalent devices in our cities that are underutilized. We devise mobility-aware scheduling for steerable infrastructure cameras to capture views for multiple concurrent analytics applications. We show how commonplace sensors can be used beyond their intended purpose. In building a framework involving widespread heterogeneous sensors and mobility-awareness, we introduce new avenues for smart city analytics. We expand this vision of sensor reuse to healthcare technology to enable continuous monitoring of physical performance during workouts. We also explore novel sensors for early diagnosis of lung and eye disorders.