Authors:
Haruka Iesaki
1
;
Tsubasa Hirakawa
1
;
Takayoshi Yamashita
1
;
Hironobu Fujiyoshi
1
;
Yasunori Ishii
2
;
Kazuki Kozuka
2
and
Ryota Fujimura
2
Affiliations:
1
Computer Science, Chubu University, 1200 Matsumoto-cho, Kasugai, Aichi, Japan
;
2
Panasonic Corporation, Japan
Keyword(s):
Path Prediction, Visual Forecasting, Predictions by Dashcams, Convolutional LSTM.
Abstract:
Autonomous cars need to understand the environment around it to avoid accidents. Moving objects like pedestrians and cyclists affect to the decisions of driving direction and behavior. And pedestrian is not always one-person. Therefore, we must know simultaneously how many people is in around environment. Thus, path prediction should be understanding the current state. For solving this problem, we propose path prediction method consider the moving context obtained by dashcams. Conventional methods receive the surrounding environment and positions, and output probability values. On the other hand, our approach predicts probabilistic paths by using visual information. Our method is an encoder-predictor model based on convolutional long short-term memory (ConvLSTM). ConvLSTM extracts visual information from object coordinates and images. We examine two types of images as input and two types of model. These images are related to people context, which is made from trimmed people’s positio
ns and uncaptured background. Two types of model are recursively or not recursively decoder inputs. These models differ in decoder inputs because future images cannot obtain. Our results show visual context includes useful information and provides better prediction results than using only coordinates. Moreover, we show our method can easily extend to predict multi-person simultaneously.
(More)