READ: Large-Scale Neural Scene Rendering for Autonomous Driving

05/11/2022
by   Zhuopeng Li, et al.
27

Synthesizing free-view photo-realistic images is an important task in multimedia. With the development of advanced driver assistance systems (ADAS) and their applications in autonomous vehicles, experimenting with different scenarios becomes a challenge. Although the photo-realistic street scenes can be synthesized by image-to-image translation methods, which cannot produce coherent scenes due to the lack of 3D information. In this paper, a large-scale neural rendering method is proposed to synthesize the autonomous driving scene (READ), which makes it possible to synthesize large-scale driving scenarios on a PC through a variety of sampling schemes. In order to represent driving scenarios, we propose an ω rendering network to learn neural descriptors from sparse point clouds. Our model can not only synthesize realistic driving scenes but also stitch and edit driving scenes. Experiments show that our model performs well in large-scale driving scenarios.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset