Image Segmentation
PyTorch
suwesh commited on
Commit
3f01960
·
verified ·
1 Parent(s): 767c0b7

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +16 -1
README.md CHANGED
@@ -2,4 +2,19 @@
2
  license: osl-3.0
3
  ---
4
  Abstract:
5
- Autonomous driving when applied for high-speed racing aside for urban environments presents unique challenges due to dynamic nature of racing circuits and the need for optimal future planning at high speeds. While simulator-based approaches, CARLA, Air Sim, TORCS provide clean environment data to models but transfer learning to real-world is a hard problem. In this paper, we propose leveraging LiDAR data obtained from the real-world to train a deep learning network to understand and predict the future states of the perceived scenes. The proposed neural network, named Perception Pyramid Network (PPN), takes a sequence of past environment scans plus the current scan, and learns how the environment evolves over a sequence of future timesteps. Due to limitation of processing raw point clouds in directly capturing spatial relationships between points the 3D point clouds obtained from LiDAR sweeps are converted into 2D Bird’s Eye View map, which encodes information in each grid cell. PPN is a type of encoder-decoder network that extracts these features, across both space and time, in a hierarchical fashion resembling the pyramid shape. The network is trained with a combination of loss functions and exhibits a real-time inference frequency of ~16 Hz on an NVIDIA Tesla P100 GPU. Implementation is available at: https://github.com/suwesh/Parallel-Perception-Network.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2
  license: osl-3.0
3
  ---
4
  Abstract:
5
+ Autonomous driving when applied for high-speed racing aside from urban environments
6
+ presents challenges in scene understanding due to rapid changes in the track environment.
7
+ Traditional sequential network approaches might struggle to keep up with the real-time
8
+ knowledge and decision-making demands of an autonomous agent which covers large
9
+ displacements in a short time. This paper proposes a novel baseline architecture for
10
+ developing sophisticated models with the ability of true hardware-enabled parallelism
11
+ to achieve neural processing speeds to mirror the agent’s high velocity. The proposed
12
+ model, named Parallel Perception Network (PPN) consists of two independent neural
13
+ networks, a segmentation and a reconstruction network running in parallel on separate
14
+ accelerated hardware. The model takes raw 3D point cloud data from the LiDAR sensor as
15
+ input and converts them into a 2D Bird’s Eye View Map on both devices. Each network
16
+ extracts its input features along space and time dimensions independently and produces
17
+ outputs in parallel. Our model is trained on a system with 2 NVIDIA T4 GPUs with a
18
+ combination of loss functions including edge preservation, and shows a 1.8x speed up in
19
+ model inference time compared to a sequential configuration. Implementation is available
20
+ at: https://github.com/suwesh/Parallel-Perception-Network.