An autonomous vehicle drives down a lonely stretch of highway. Suddenly, a massive tornado appears in the distance. What does the driverless vehicle do next?
This is just one of the scenarios that Waymo can simulate in the âhyper realisticâ virtual world that it has just created with help from Googleâs DeepMind. Waymoâs World Model is built using Genie 3, Googleâs new AI world model that can generate virtual interactive spaces with text or images as prompts. But Genie 3 isnât just for creating bad knockoffs of Nintendo games; it can also build photorealistic and interactive 3D environments âadapted for the rigors of the driving domain,â Waymo says.
Simulation is a critical component in autonomous vehicle development, enabling developers to test their vehicles in a variety of settings and scenarios, many of which may only come up in the rarest of occasions â without any physical risk of harming passengers or pedestrians. AV companies use these virtual environments to run through a battery of tests, racking up millions â or even billions â of miles in the process, in the hopes of better training their vehicles for any possible âedge caseâ that they may encounter in the real world.
What kind of edge cases is Waymo testing? In addition to the aforementioned tornado, the company can also simulate a snow-covered Golden Gate Bridge, a flooded suburban cul-de-sac with floating furniture, a neighborhood engulfed in flames, or even an encounter with a rogue elephant. In each scenario, the Waymo robotaxiâs lidar sensors generate a 3D rendering of the surrounding environment, including the obstacle in the road.
âThe Waymo World Model can generate virtually any sceneâfrom regular, day-to-day driving to rare, long-tail scenariosâacross multiple sensor modalities,â the company says in a blog post.
Waymo says Genie 3 is ideal for creating virtual worlds for its robotaxis, citing three unique mechanisms: driving action control, scene layout control, and language control. Driving action control allows developers to simulate âwhat ifâ counterfactuals, while scene layout control enables customization of the road layouts, like traffic signals and other road user behavior. Waymo describes language control as its âmost flexible toolâ that allows for time-of-day and weather condition adjustment. This is especially helpful if developers are trying to simulate low-light or high-glare conditions, in which the vehicleâs various sensors may have difficulty seeing the road ahead.
The Waymo World Model can also take real-world dashcam footage and transform it into a simulated environment, for the âhighest degree of realism and factualityâ in virtual testing, the company says. And it can create longer simulated scenes, such as ones that run at 4X speed playback, without sacrificing image quality or computer processing.
âBy simulating the âimpossible,â we proactively prepare the Waymo Driver for some of the most rare and complex scenarios,â the company says in its blog post.
Read the full article here