Most LiDAR point cloud processing techniques continue to gather more data as the data is available. This is also typical in most imaging systems, especially visible light camera systems. We propose a computationally efficient solution where data only continues to be processed if the data has changed. Once points are received by the LiDAR hardware driver, a sensor frame spatial event filter is used to compare a previous point with the most recent point obtained from that same coordinate in the LiDAR's receptor array. The output of the event filter then fills an array of events, or event map, that will be accessible by a layer of neurons that can be implemented in a GPU. The operations per point are compared between this event-based solution and other similar solutions. We show the event-based solution's efficiency can be better, according to how much the scene is changing and how many post-processing steps are involved. Point cloud data is collected from a LiDAR mounted on a vehicle driving in paved road conditions to illustrate the concept.
|