CVLab COMPUTER VISION LAB

National Tsing Hua University


Betpasgiris.vip restbetgiris.co betpastakip.com restbet.com betpas.com restbettakip.com nasiloynanir.co alahabibi.com hipodrombet.com malatya oto kiralama istanbul e?ya depolama istanbul-depo.net e?ya depolama sex shop ?ehirler aras? nakliyat evden eve nakliye ofis ta??ma transfernakliyat.com.tr heceder.org 30 tl deneme bonusu aviator oyunu betexper casino siteleri kamyonet.biz.tr tuzcuo?lu nakliyat malatya temizlik shell download alfa shell e?ya depolama depolama sar?yer evden eve nakliyat ?i?li evden eve nakliyat istanbul evden eve nakliyat

Pedestrian Detection

64_cover
Fri, 20 Jul 2018 - intern_james
Search Space Reduction in Pedestrian Detection for Driver Assistance System Based on Projective Geometry
K. Dimza, T.F. Su, S.H. Lai.
Vehicles are equipped with smarter and smarter driver assistance systems to improve driving safety year by year. On-board pedestrian detection system is a critical and challenging task for driving safety improvement because driving environment is very dynamic, where humans appear in wide varieties of clothing, illumination, size, speed and distance from the vehicle. Most of existing methods are based on the sliding window search methodology to localize humans in an image. The easiest and also the most popular way is to check the whole image at all possible scales. However, such methods usually produces large number of false positives and are computationally expensive because large number of inappropriate regions were checked. In this paper, we develop a method which reduce the search space in pedestrian detection by using properties of projective geometry in the case when camera parameters are unavailable. The simple user interaction with stochastic optimization is used to estimate projective parameters. We showed the efficiency of our method on public dataset with known camera parameters and self captured dataset without registered camera parameters. Experiment results show that the effectiveness of the proposed method is superior compared to the traditional uniform sliding window selection strategy.
63_cover
Fri, 20 Jul 2018 - intern_james
Single-shot person re-identification based on improved Random-Walk pedestrian segmentation
Y.-C. Chang, C.-K. Chiang and S.-H. Lai
Single-shot person re-identification is to match pedestrian images captured from different cameras at different time under the condition of large illumination variations, different viewpoints, and inadequate information of single-shot case. To deal with these challenges, we propose a four-step single-shot person re-identification algorithm that consists of pedestrian segmentation, human region partitioning, feature extraction and human feature matching. Based on an improved Random Walks algorithm, human foreground is segmented by combining the shape prior information and the color seed constraint into the Random Walk formulation. Then color features of HSV histogram and 1-D RGB signal along with texture features from human body parts are used for the person re-identification. The correct match is then determined by the similarity scores of all features with appropriate weight selection. The experimental results demonstrate the superior performance by using the proposed algorithm compared to the previous representative methods.
21_cover
Wed, 24 Nov 2010 - wowowo
Video Surveillance from Pedestrian Detection
張幼臻, 王芳瑜, 李郁慈, 蘇德峰, 江振國
Pedestrian detection is helpful to video surveillance when the crime happens. To search a pedestrian from a recorded video database, the background model is built based on a mixture Gaussian model. This model is updated by successive frames to accommodate the lighting change. Then, moving objects are detected via background subtraction. Two classifiers, pixelwise classifier and region-based classifier, are used to remove noises and shadows of objects. After obtaining complete foreground objects, different features, like color, texture, are extracted from each body segment, head, torso and legs. Then, cross matching is performed to ensure effective feature matching. In addition, color consistency is also considered for extremely dark or bright clips.

Car Detection

99_cover
Wed, 25 Jul 2018 - intern_james
Fast Vehicle Detector for Autonomous Driving
Che-Tsung Lin, Patrisia Sherryl Santoso, Shu-Ping Chen, Hung-Jin Lin and Shang-Hong Lai
This paper presents a fast vehicle detector which can be deployed on NVIDIA DrivePX2 under real-time constraints. The network predicts bounding boxes with different aspect ratio and scale priors from the specifically-designed prediction module given concatenated multi-scale feature map. A new data augmentation strategy is proposed to systematically generate a lot of vehicle training images whose appearance is randomly truncated so our detector could detect occluded vehicles better. Besides, we propose a non-region-based online hard example mining framework which performs fine-tuning by picking (1) hard examples and (2) detection results with insufficient IOU. Compared to other classical object detectors, this work achieves very competitive result in terms of average precision (AP) and computational speed. For the newly-defined vehicle class (car+bus) on VOC2007 test, our detector achieves 85.32 AP and runs at 48 FPS and 30 FPS on NVIDIA Titan X & GP106 (DrivePX2), respectively.
84_cover
Tue, 24 Jul 2018 - intern_james
Integrated Vehicle and Lane Detection with Distance Estimation
Yu-Chun Chen, Te-Feng Su, Shang-Hong Lai
In this paper, we propose an integrated system that combines vehicle detection, lane detection, and vehicle distance estimation in a collaborative manner. Adaptive search windows for vehicles provide constraints on the width between lanes. By exploiting the constraints, the search space for lane detection can be efficiently reduced. We employ local patch constraints for lane detection to improve the reliability of lane detection. Moreover, it is challenging to estimate the vehicle distance from images/videos captured form monocular camera in real time. In our approach, we utilize lane marker with the associated 3D constraint to estimate the camera pose and the distances to frontal vehicles. Experimental results on real videos show that the proposed system is robust and accurate in terms of vehicle and lane detection and vehicle distance estimation.
82_cover
Mon, 23 Jul 2018 - intern_james
Efficient vehicle detection with adaptive scan based on perspective geometry
Yu-Chun Chen, Te-Feng Su, and Shang-Hong Lai
Vehicle detection is an important research problem for Advanced Driver Assistance Systems to improve driving safety. Most existing methods are based on the sliding window search framework to locate vehicles in an image. However, such methods usually produce large numbers of false positives and are computationally intensive. In this paper, we propose an efficient vehicle detection algorithm that dramatically reduces the search space based on the perspective geometry of the road. In the training phase, we search a few images to locate all possible vehicle regions by using the standard HOG-based vehicle detector. Pairs of vehicle candidates that satisfy the projective geometry constraints are used to estimate the linear vehicle width model with respect to y coordinates in the image. Then an adaptive scan strategy based on the estimated vehicle width model is proposed to efficiently detect vehicles in an image. Experimental results show that the proposed algorithm provides improved performance in terms of both speed and accuracy compared to standard sliding-windows search strategy.
39_cover
Mon, 04 Feb 2013 - tfsu
EFFICIENT VEHICLE DETECTION WITH ADAPTIVE SCAN BASED ON PERSPECTIVE GEOMETRY
Yu-Chun Chen, Te-Feng Su, and Shang-Hong Lai
Vehicle detection is an important research problem for Advanced Driver Assistance Systems to improve driving safety. Most existing methods are based on the sliding window search framework to locate vehicles in an image. However, such methods usually produce large numbers of false positives and are computationally intensive. In this paper, we propose an efficient vehicle detection algorithm that dramatically reduces the search space based on the perspective geometry of the road. In the training phase, we search a few images to locate all possible vehicle regions by using the standard HOG-based vehicle detector. Pairs of vehicle candidates that satisfy the projective geometry constraints are used to estimate the linear vehicle width model with respect to y coordinates in the image. Then an adaptive scan strategy based on the estimated vehicle width model is proposed to efficiently detect vehicles in an image. Experimental results show that the proposed algorithm provides improved performance in terms of both speed and accuracy compared to standard sliding-windows search strategy.

Image/Video Retargeting

19_cover
Tue, 23 Nov 2010 - Savan
Image Compressibility Assessment and the Application of Structure-Preservng Image Retargeting
Shu-Fan Wang and Shang-Hong Lai
A number of algorithms have been proposed for intelligent image/video retargeting with important content retained as much as possible. In some cases, we can notice that they suffer from artifacts in the resized results, such as ridge or structure twist. In this paper, we suggest that the compressibility of an image should be estimated properly first by analyzing the image structure to determine the optimal scaling factors for the resizing algorithm. To cope with this problem, we propose a compressibility assessment scheme by combining the entropies of image gradient magnitude and orientation distributions. In order to further improve the result, we also present a structure-preserving media retargeting technique that preserves the content and image structure as best as possible. Since we focus on protecting the content structure, a block structure energy is introduced with a top-down strategy to constrain the image structure inside to scale uniformly in either x or y direction. Our experiments demonstrate that the proposed compressibility assessment scheme provides better preservation of content and structure in the resized images/videos than those by the previous methods.

Computer Vision Applications on Embedded Multi-core Systems

67_cover
Fri, 20 Jul 2018 - intern_james
Parallelized Background Substitution System on a Multi-core Embedded Platform
Yutzu Lee, Chen-Kuo Chiang, Te-Feng Su, Yu-Wei Sun, Chi-Bang Kuan, Shang-Hong Lai
We present an automatic human background substitution system based on a Random Walk (RW) algorithm on a multi-core processing architecture. Firstly, a fast algorithm is proposed to solve the large linear system in RW based on adapting the Gauss-Seidel method. Two tables, TYPE and INDEX, are introduced to fast locate the required data for the close-form solution. Then, face detection along with a human shape prior model are exploited to decide the approximated human body and background area. Pixels inside these areas are used as seed points in RW algorithm for automatic segmentation. The proposed method is designed to be highly parallelizable and suitable for running on a multi-core architecture. We demonstrate the parallelization strategies for the proposed fast RW algorithm and face detection on heterogeneous multi-core embedded platform to make the most use of the system architecture. Compared to the single processor implementation, the experimental results show significant speedup ratio of the parallelized human background substitution system on a multi-core embedded platform, which consists of an ARM processor and two DSP cores.
65_cover
Fri, 20 Jul 2018 - intern_james
Parallelized Random Walk algorithm for background substitution on a multi-core embedded platform
Y. Lee, C.-K. Chiang, Y.-W. Sun, T.-F. Su, S.-H. Lai
Random Walk (RW) is a popular algorithm and can be applied to many applications in computer vision. In this paper, a fast algorithm is proposed to solve the large linear system in RW based on adapting the Gauss-Seidel method on a multi-core embedded system. Two tables, TYPE and INDEX, are introduced to fast locate the required data for the close-form solution. The computational overhead, along with the memory requirement, to solve the linear system can be reduced greatly, thus making the RW algorithm feasible to many applications on an embedded system. In addition, the proposed fast method is parallelized for a heterogeneous multi-core embedded platform to make the most use of the benefits of the system architecture. Experimental results show that the computational overhead can be significantly reduced by the proposed algorithm.
44_cover
Thu, 19 Jul 2018 - intern_james
Support of software framework for embedded multi-core systems with Android environments
Y.-H.Chang, C.-B. Kuan, C.-Y. Lin, T.-F. Su, C.-T. Chen, J.-S. Jang
Applications on mobile devices are getting more complicated with the new wave of applications in the mobile devices. The computing power for embedded devices are increased with such trends, and embedded multi-core platform are in a position to help boost system performance. Software frameworks integrated the multi-core platforms are often needed to help boost the system performance and reduce programming complexity. In this paper, we present a software framework based on Android and multi-core embedded systems. In the framework, we integrate the compiler toolkit chain for multi-core programming environment which includes DSP C/C++ compilers, streaming RPC programming model, debugger, ESL simulator, and power management models. We also develop software framework for face detection, voice recognition, and mobile streaming management. Those frameworks are designed as multi-core programs and are used to illustrate the design flow for applications on embedded multi-core environments equipped with Android systems. We demonstrate our proposed mechanisms by implementing two applications, Face RMS and voice recognition. The proposed framework gives a case study to illustrate software framework and design flow for emerging RMS-based and voice recognition applications on embedded multi-core systems equipped with Android systems.
27_cover
Thu, 25 Nov 2010 - tfsu
Parallelized Face Based RMS System on a Multi-core Embedded Computing Platform
蘇德峰、曾聖博、段志學、王書凡、廖家德
The amount of data generated in the world continues to grow at an incredibly high speed. A new framework for the next generation, called the Recognition, Mining and Synthesis (RMS) system, was proposed to make meaningful use of the enormous amount of information on a multi-core processing architecture. Based on the same concept, we propose a face RMS system, which consists of face detection, facial expression recognition, and facial expression exaggeration components, for generating exaggerated views of different expressions on an input face image. In this paper, the parallel algorithms of the face RMS system were developed to reduce the execution time on a multi-core embedded system. The experimental results show the robustness of face detection with different scales and expressions in complex environments and an efficient non-linear method for expression exaggeration. The quantitative comparisons indicate the proposed parallelized face RMS system has a significant increase in speedup compared to the single processor implementation on the multi-core embedded platform, which consists of an ARM processor and two DSP cores.

Over-segmentation Based Background Modelling and Foreground Detection

55_cover
Fri, 20 Jul 2018 - intern_james
From co-saliency to co-segmentation: an efficient and fully unsupervised energy minimization model
K.-Y. Chang, T.-L. Liu, and S.-H. Lai
Background subtraction is commonly used to detect foreground objects in video surveillance. Traditional background subtraction methods are usually based on the assumption that the background is stationary. However, they are not applicable to dynamic background, whose background images change over time. In this paper, we propose an adaptive Local-Patch Gaussian Mixture Model (LPGMM) as the dynamic background model for detecting moving objects from video with dynamic background. Then, the SVM classification is employed to discriminate between foreground objects and shadow regions. Finally, we show some experimental results on several video sequences to demonstrate the effectiveness and robustness of the proposed method.
53_cover
Fri, 20 Jul 2018 - intern_james
Detecting Moving Objects from Dynamic Background with Shadow Removal
S.-C. Wang, T.-F. Su, and S.-H. Lai
Background subtraction is commonly used to detect foreground objects in video surveillance. Traditional background subtraction methods are usually based on the assumption that the background is stationary. However, they are not applicable to dynamic background, whose background images change over time. In this paper, we propose an adaptive Local-Patch Gaussian Mixture Model (LPGMM) as the dynamic background model for detecting moving objects from video with dynamic background. Then, the SVM classification is employed to discriminate between foreground objects and shadow regions. Finally, we show some experimental results on several video sequences to demonstrate the effectiveness and robustness of the proposed method.
28_cover
Thu, 25 Nov 2010 - tfsu
Over-Segmentation Based Background Modeling and Foreground Detection with Shadow Removal by Using Hierarchical MRFs
Te-Feng Su,Yi-Ling Chen and Shang-Hong Lai
In this paper, we propose a novel over-segmentation based method for the detection of foreground objects from a surveillance video by integrating techniques of background modeling and Markov Random Fields classification. Firstly, we introduce a fast affinity propagation clustering algorithm to produce the over-segmentation of a reference image by taking into account color difference and spatial relationship between pixels. A background model is learned by using Gaussian Mixture Models with color features of the segments to represent the time-varying background scene. Next, each segment is treated as a node in a Markov Random Field and assigned a state of foreground, shadow and background, which is determined by using hierarchical belief propagation. The relationship between neighboring regions is also considered to ensure spatial coherence of segments. Finally, we demonstrate experimental results on several image sequences to show the effectiveness and robustness of the proposed method.
In Proceedings of Asian Conference on Computer Vision (ACCV'10), Nov. 2010. (pdf)
maltepe escort ataşehir escort kadıköy escort anadolu yakası escort ataşehir escort pendik escort kurtköy escort kartal escort kadıköy escort
Hızlıcasino Zirvebet Egebet Vbet Nakitbahis Betwinner Onwin Betturkey
Trbet Canlı Bahis Canlı Bahis Canlı Bahis Canlı Bahis Canlı Bahis
redbullholdenracing.com haberanadolu.org
deneme bonusu
deneme bonusu veren siteler