Description: Description: http://www.cs.iupui.edu/~jzheng/incarv2.jpgIntelligent Vehicles

 

 

Pedestrian detection has been intensively studied based on appearances for driving safety. Only a few works have explored between-frame optical flow as one of features for human classification. In this paper, however, a new point of view is taken to watch a longer period for non-smooth movement. We explore the pedestrian detection purely based on motion, which is common and intrinsic for all pedestrians regardless of their shape, color, background, etc. We found unique motion characteristics of humans different from rigid object motion caused by vehicle motion.

My Publications:

Mehmet Kilicarslan, Jiang Yu Zheng, Aied Algarni, Pedestrian detection from non-smooth motion, IEEE Intelligent Vehicles Symposium (IV), 2015, pp. 487-492

Mehmet Kilicarslan, Jiang Yu Zheng, Detecting walking pedestrians from leg motion in driving video, 2014 IEEE 17th International Conference on Intelligent Transportation Systems (ITSC), pp. 2924-2929

 

 

Monocular camera based bicyclist detection in naturalistic driving video is a very challenging problem due to the high variance of the bicyclist appearance and complex background of naturalistic driving environment. In this paper, we propose a two-stage multi-modal bicyclist detection scheme to efficiently detect bicyclists with varied poses for further behavior analysis. A new motion based region of interest (ROI) detection is first applied to the entire video to refine the region for sliding-window detection. Then an efficient integral feature based detector is applied to quickly filter out the negative windows. Finally, the remaining candidate windows are encoded and tested by three pre-learned pose-specific detectors. The experimental results on our TASI 110 car naturalistic driving dataset show the effectiveness and efficiency of the proposed method. The proposed method outperforms the traditional methods.

My Publications:

Kai Yang, Chao Liu, Jiang Yu Zheng, Lauren Christopher, Yaobin Chen, Bicyclist detection in large scale naturalistic driving video, 2014 IEEE 17th International Conference on Intelligent Transportation Systems (ITSC), 2015, pp. 1638-1643

 

Pedestrian Automatic Emergency Braking (PAEB) for helping avoiding/mitigating pedestrian crashes has been equipped on some passenger vehicles. Since approximately 70% pedestrian crashes occur in dark conditions, one of the important components in the PAEB evaluation is the development of standard testing at night. The test facility should include representative low-illuminance environment to enable the examination of the sensing and control functions of different PAEB systems. The goal of this research is to characterize and model light source distributions and variations in the low-illuminance environment and determine possible ways to reconstruct such an environment for PAEB evaluation. This paper describes a general method to collect light sources and illuminance information by processing large amount of potential collision locations at night from naturalistic driving video data. This study was conducted in four steps. (1) Gather night driving video collected from Transportation Active Safety Institute (TASI) 110 car naturalistic driving study, particularly emphasizing locations with potential pedestrian collision. (2) Generate temporal video profile as a compact index toward large volumes of video, (3) Identify light fixtures by removing dynamic vehicle head lighting in the profile and stamp them with their Global Positioning System (GPS) coordinates. (4) Find the average distribution and intensity of illuminants by grouping lighting component information around the potential collision locations. The resulting lighting model and setting can be used for lighting reconstruction at PAEB testing site.

My Publications:

Libo Dong, Stanley Chien, Jiang-Yu Zheng, Yaobin Chen, Rini Sherony, Hiroyuki Takahashi, Modeling of Low Illuminance Road Lighting Condition Using Road Temporal Profile, SAE Technical Paper, 2016, pp. 1638-1643

 

This work models various dangerous situations that may happen to a driving vehicle on road in probability, and determines how such events are mapped to the visual field of the camera. Depending on the motion flows detected in the camera, our algorithm will identify the potential dangers and compute the time to collision for alarming. The identification of dangerous events is based on the location-specific motion information modeled in the likelihood probability distributions. The originality of the proposed approach is at the location dependent motion modeling using the knowledge of road environment. This will link the detected motion to the potential danger directly for accident avoidance. The mechanism from visual motion to the dangerous events omits the complex shape recognition so that the system can response without delay.

My Publications:

Mehmet Kilicarslan, Jiang Yu Zheng, Modeling Potential Dangers in Car Video for Collision Alarming, IEEE International Conference on Vehicle Electronics and safety 2012, 1-6

Mehmet Kilicarslan,Jiang Yu Zheng, Towards Collision Alarming based on Visual Motion. IEEE International Conference on Intelligent Transportation Systems 2012: 1-6

 

Nowadays, many vehicles are equipped with a vehicle borne camera system for monitoring drivers’ behavior, accident investigation, road environment assessment, and vehicle safety design. Huge amount of video data is being recorded daily. Analyzing and interpreting these data in an efficient way has become a non-trivial task. As an index of video for quick browsing, this work maps the video into a temporal image of reduced dimension with as much intrinsic information as possible observed on the road. The perspective projection video is converted to a top-view temporal profile that has precise time, motion, and event information during the vehicle driving. Then, we attempt to interpret dynamic events and environment around the vehicle in such a continuous and compact temporal profile. The reduced dimension of the temporal profile allows us to browse the video intuitively and efficiently.

 

Publications:

Mehmet Kilicarslan, Jiang Yu Zheng, Visualizing Driving Video in Temporal Profile, IEEE Intelligent Vehicle Symposim, 2014, 1-7.