Future connected-car technology merges real and virtual worlds to help drivers `see the invisible’
Yokohama, Japan (4 January, 2019) At the upcoming CES trade show, Nissan will unveil its future vision for a vehicle that helps drivers “see the invisible” by merging both real and virtual worlds, creating the ultimate connected-car experience.
Invisible-to-Visible, or I2V, is a future technology created through Nissan Intelligent Mobility, the company’s vision for changing how cars are powered, driven and integrated into society.
I2V will demonstrate the future of driving to CES visitors through an interactive, three-dimensional immersion experience at Nissan’s display.
I2V will support drivers by merging information from sensors outside and inside the vehicle with data from the cloud. This enables the system not only to track the vehicle’s immediate surroundings but also to anticipate what’s ahead – even showing what’s behind a building or around the corner.
To make driving more enjoyable, guidance is given in an interactive, human-like way, such as through avatars that appear inside the car.
By tapping into the virtual world, I2V opens up endless possibilities for service and communication – making driving more convenient, comfortable and exciting.
“By helping you see the invisible, I2V enhances your confidence and makes driving more enjoyable,” said Tetsuro Ueda, an expert leader at the Nissan Research Center.
“The interactive features create an experience that’s tailored to your interests and driving style so that anyone can enjoy using it in their own way.”
Connected driving like never before
I2V is powered by Nissan’s Omni-Sensing technology, which acts as a hub gathering real-time data from the traffic environment and from the vehicle’s surroundings and interior.
Nissan’s SAM (Seamless Autonomous Mobility) technology analyses the road environment through relevant real-time information, and the ProPILOT semi-autonomous driver support system provides information about the car’s surroundingsThe technology maps a 360-degree virtual space around the car to provide information about things like road and intersection status, visibility, signage or nearby pedestrians.
It can also monitor the people inside the vehicle by using interior sensors to better anticipate when they may need assistance with finding something or a coffee break to stay alert.
I2V can also connect drivers and passengers to people in the Metaverse virtual world. This makes it possible for family, friends or others to appear inside the car as three-dimensional, augmented-reality avatars to provide company or assistance.
More support during autonomous or manual driving
During autonomous driving, I2V can make the time spent in a car more comfortable and enjoyable. For example, when driving in the rain, the scenery of a sunny day can be projected inside the vehicle.
When visiting a new place, the system can search within the Metaverse for a knowledgeable local guide who can communicate with people in the vehicle in real time.
Information provided by the guide can be collected with Omni-Sensing and stored in the cloud so that others visiting the same area can access the useful guidance. It can also be used by the onboard artificial intelligence system to provide a more efficient drive through local areas.
During manual driving, I2V provides information from Omni-Sensing as an overlay in the driver’s full field of view. The information helps drivers assess and prepare for things like corners with poor visibility, irregular road surface conditions or oncoming traffic.
The driver can also book a professional driver from the Metaverse to get personal instruction in real time. The professional driver appears as a projected avatar or as a virtual chase car in the driver’s field of vision to demonstrate the best way to drive.
Not only will I2V display information about congestion and estimated travel time; it can communicate unique details to better inform the driver about the road ahead, removing unknowns to ease driving stress.
I2V will give alternative suggestions, even down to the best-moving lane in heavy traffic, by using real-time local data mapping via Omni-Sensing.
Upon arrival at a destination, I2V can access SAM to scan for parking spaces and park the vehicle for the driver in situations requiring difficult parking maneuvers.
I2V demonstration at Nissan’s display
Visitors to CES can experience I2V at Nissan’s display by putting on a pair of augmented-reality goggles and stepping inside a demonstration cockpit featuring three-dimensional interfaces and displays.
Users are guided through scenarios including a tour of a city, receiving help to find an open parking space at a busy shopping centre, seeing a rainy day outside change to a sunny day inside the car, chasing a professional driver avatar to improve driving skills and exploring how I2V can see through buildings and around corners.CES 2019 runs from January 8 to 11 at the Las Vegas Convention Centre.