SDG 2: Zero Hunger - End hunger, achieve food security and improved nutrition and promote sustainable agriculture
I envision a future where we are surrounded by small robots doing various tasks such as pollinating flowers, searching for survivors, carrying payloads, entertaining us and even as our personal pets/companions. To advance autonomy on these tiny mobile robots with on-board sensing and computation, we draw inspiration from nature’s experts – insects, birds and other small beings. My doctoral work culminated in building the world’s first prototype of a RoboBeeHive that involved hummingbird-sized nano-quadrotors capable of pollinating flowers, where all the sensing and computation are performed on-board. My overarching goal to advance autonomy of small mobile robots is built upon four major thrusts: Active perception, interactive perception, novel perception and novel sensing. All these thrusts embrace the perception and action synergy pair to simplify perception problems by controlling one’s movement.
Active perception utilizes the movement of the agent or a part to simplify perception problems. Interactive perception takes it one step further by selectively interacting with the environment to obtain selective cues for better understanding. Novel perception involves building mathematical models based on statistics of data such as neural network uncertainty to solve problems in a minimal manner. For example, one could look at uncertainty of optical flow to dodge obstacles rather than building a depth map, thereby improving agility. Finally, novel sensing involves formulating solutions to robotics problems using novel sensors such as event cameras or building passive structures such as coded apertures to view the world in a “new way” that can enable solving robotics problems in a better way.
All in all, these directions will enable us to build the next generation of robots that will pollinate our plants, save us during disasters, be our pets and much more. Come join us in fulfilling this direction.