The VEMI lab at the University of Maine earned third place in a national competition and an invitation to the White House for developing an inclusive smartphone software platform that will provide navigational assistance to people with visual impairment and seniors who want to use self-driving cars for ride-sharing and hailing services.
VEMI will receive $300,000 for winning the prize in the second phase of the U.S. Department of Transportation’s Inclusive Design Challenge for its Autonomous Vehicle Assistant (AVA) smartphone technology. VEMI leads the group designing the AVA platform, known as the Autonomous Vehicle Research Group (AVRG), which also includes collaborators from Northeastern University and Colby College.
For its challenge, the DOT sought proposals for accessible and inclusive design solutions that would help people with disabilities use autonomous vehicles for employment and essential services. VEMI was invited to participate in Stage II of the challenge after being named a semifinalist in the first phase.
First- and second-place winners in the competition were Purdue University and AbleLink Smart Living Technologies, respectively.
The prizes in the Inclusive Design Challenge were announced July 26 as part of the DOT’s celebration for the 32nd Anniversary of the Americans with Disabilities Act, known as the ADA.
How can the AVA software help blind and partially sighted users?
AVA will help users request a vehicle, find it, enter it, exit it and travel to their chosen destination. It provides a multisensory interface that offers users guidance through audio and haptic, or touch-based, feedback and high-contrast visual cues.
What functions are utilized in the software?
To provide the functionality, researchers utilized GPS, LiDAR, Gyroscope and Accelerometer technology; real-time computer vision via the smartphone camera; machine learning; artificial intelligence and other software.
How does the AVA software work?
Users will create a profile in AVA that reflects their accessibility needs and existing methods of navigation so the software can find suitable transportation for them.
When the vehicle arrives, AVA will guide the user to it using the camera and augmented reality (AR), which provides an overlay of the environment by superimposing high-contrast lines over the image on the smartphone screen to highlight the path, and verbal guidance such as compass directions, street names, addresses, nearby landmarks and other indicators.
The software also will pinpoint environmental hazards, including low-contrast curbs, traffic cones and overhanging obstructions like branches and guy wires, by emphasizing them with contrasting lines and vibrations when users approach them.
It will then help users find the door handle to enter the vehicle awaiting them. It also uses the same functions to help the user when exiting their vehicle to find their destination.
Is there a learning curve using the software?
AVA will offer accessible modules with simulations that train users not only how to use the application, but also training on how to interact with ride sharing and hailing services with self-driving vehicles when a person is no longer available to provide assistance. For future projects, researchers plan to develop additional software that will allow riders to use the technology to connect with the vehicle control systems while riding. These tools will include multisensory maps, context-aware gesture interactions and application programming interfaces, all of which will support in-cabin accessibility.
Watch the video presentation about AVA prepared by VEMI Lab researchers for the DOT’s Inclusive Design Challenge to learn more.
The AVA project builds on a National Science Foundation grant led by Giudice and Corey on trust building and human-vehicle collaboration with autonomous vehicles, as well as a seed grant-funded, joint effort between UMaine and Northeastern University to improve accessibility, safety and situational awareness within self-driving vehicles. Research on both projects aims to develop a new model of human-AI vehicle interaction to ensure people with visual impairments and seniors can better understand what their autonomous vehicle does during their travels, and so the vehicles can effectively communicate with them — work that will be instrumental for informing future AVA development on this Inclusive Design Challenge prize.