自定义博客皮肤VIP专享

*博客头图:

格式为PNG、JPG,宽度*高度大于1920*100像素,不超过2MB,主视觉建议放在右侧,请参照线上博客头图

请上传大于1920*100像素的图片!

博客底图:

图片格式为PNG、JPG,不超过1MB,可上下左右平铺至整个背景

栏目图:

图片格式为PNG、JPG,图片宽度*高度为300*38像素,不超过0.5MB

主标题颜色:

RGB颜色,例如:#AFAFAF

Hover:

RGB颜色,例如:#AFAFAF

副标题颜色:

RGB颜色,例如:#AFAFAF

自定义博客皮肤

-+

AI小作坊 的博客

大道至简,天人合一

  • 博客(27)
  • 资源 (91)
  • 收藏
  • 关注

原创 CNN网络加速--Performance Guaranteed Network Acceleration via High-Order Residual Quantization

Performance Guaranteed Network Acceleration via High-Order Residual Quantization ICCV2017本文是对 XNOR-Networks 的改进,将CNN网络层的输入 进行高精度二值量化,从而实现高精度的二值网络计算,XNOR-Networks 也是对每个CNN网络层的权值和输入进行二值化,这样整个CNN计算都是二值化

2017-08-31 15:18:24 1957 1

原创 目标检测--RON: Reverse Connection with Objectness Prior Networks for Object Detection

RON: Reverse Connection with Objectness Prior Networks for Object Detection CVPR2017 https://github.com/taokong/RON本文可以看作是对 SSD 的改进, SSD 对不同尺度特征图进行独立的检测,这里我们 reverse connection block 将相邻的特征图联系起来。同时使用

2017-08-30 15:55:11 5113 2

原创 去水印--《On the Effectiveness of Visible Watermarks》

On the Effectiveness of Visible Watermarks CVPR2017大牛 William T. Freeman 去了 Google Research 本文没有使用深度学习,使用传统的优化算法来解决水印去除。首先看图有个感性认识吧 基于单张图像去除水印的难度还是很大的。这里我们在网上搜集使用同一个水印的大量图像,基于这些图像,我们估计出 Watermark (W

2017-08-30 10:57:52 10598 6

原创 CNN阴影去除--DeshadowNet: A Multi-context Embedding Deep Network for Shadow Removal

DeshadowNet: A Multi-context Embedding Deep Network for Shadow Removal CVPR2017本文使用深度学习CNN网络来进行阴影去除,最大的特色就是全自动的端对端的实现阴影去除。 automatic and end-to-end deep neural network (DeshadowNet)阴影去除也算是一个老大难问题了,目前存

2017-08-29 13:49:15 8256 11

原创 CNN边缘检测--Richer Convolutional Features for Edge Detection

Richer Convolutional Features for Edge Detection CVPR2017 Caffe:https://github.com/yun-liu/rcf本文针对边缘检测问题,基于 VGG16 网络设计了一个 richer convolutional features (RCF) 用于边缘检测,效果目前是很好的。首先来看看 VGG16不同卷积层的特征输出 3

2017-08-28 16:54:39 13518 1

原创 目标检测分割--BlitzNet: A Real-Time Deep Network for Scene Understanding

BlitzNet: A Real-Time Deep Network for Scene Understanding ICCV2017 Project: http://thoth.inrialpes.fr/research/blitznet/ Code: https://github.com/dvornikita/blitznet本文在 SSD 基础上改进提出 BlitzNet,使其

2017-08-25 16:11:50 4630

原创 语义分割--Deep Dual Learning for Semantic Image Segmentation

Deep Dual Learning for Semantic Image Segmentation ICCV2017针对语义分割问题,本文提出了一个 dual image segmentation (DIS)系统 利用一部分 per-pixel labelmaps的训练样本和 一部分 image-level tags 的样本 进行联合训练,得到较好的分割结果。本文定义了一些符号: I 输入图

2017-08-25 11:09:02 3376

原创 分割候选区域--FastMask: Segment Multi-scale Object Candidates in One Shot

FastMask: Segment Multi-scale Object Candidates in One Shot CVPR2017 https://github.com/voidrank/FastMask本文针对检测和分割问题提出 FastMask 实现 segment multi-scale objects in one shot 这里的 one shot ( original im

2017-08-24 14:30:43 1796

原创 行人检测--What Can Help Pedestrian Detection?

What Can Help Pedestrian Detection? CVPR2017本文主要分析 extra features 对于 基于CNN 的行人检测有什么帮助,设计了一个行人检测网络 HyperLearner 可以有效利用这些 extra features这里的 extra features 主要指 various channel features 行人检测中存在什么问题了?

2017-08-24 10:02:10 3937 3

原创 随机采样池化--S3Pool: Pooling with Stochastic Spatial Sampling

S3Pool: Pooling with Stochastic Spatial Sampling CVPR2017 https://github.com/Shuangfei/s3pool本文将常规池化看作两个步骤: 1)以步长为1在特征图上滑动池化窗口,尺寸大小基本保持不变, leaves the spatial resolution intact 2)以一种 uniform 和 deter

2017-08-23 15:07:14 6306

原创 有效感受野--Understanding the Effective Receptive Field in Deep Convolutional Neural Networks

Understanding the Effective Receptive Field in Deep Convolutional Neural Networks NIPS 2016本文主要分析了 CNN 网络中的 Receptive Field,发现实际有效的感受野 和 理论上的感受野 差距比较大,实际有效的感受野是一个高斯分布。We introduce the notion of an eff

2017-08-23 09:12:50 8350 6

原创 人脸检测--Recurrent Scale Approximation for Object Detection in CNN

Recurrent Scale Approximation for Object Detection in CNN ICCV2017 https://github.com/sciencefans/RSA-for-object-detection本文还是针对人脸检测 中的 尺度问题 进展展开的。主要内容有以下三点: 1)首先使用一个 scale-forecast 网络来进行图像中人脸尺度的预测,

2017-08-22 15:02:29 3589

原创 人脸检测--S3FD: Single Shot Scale-invariant Face Detector

S3FD: Single Shot Scale-invariant Face Detector CVPR2017 Caffe code will be available本文针对基于 anchor 的检测器对 小的人脸检测率低的问题进行了分析和改进。基于 anchor 的目标检测发展迅速,人脸检测也有很大进展,但是对于小的人脸检测效果仍然不是很好。 the performance of an

2017-08-22 11:07:25 5491 16

原创 人脸检测--Scale-Aware Face Detection

Scale-Aware Face Detection CVPR2017针对人脸检测中的人脸多尺度问题,本文首先用一个 Scale Proposal Network (SPN) 估计出图像中人脸的尺度分布,然后按照该尺度归一化图像,再进行人脸检测 均匀分布的多尺度人脸检测有时是一种浪费,因为图像中可能只有 1-2 个尺度的人脸 所以本文的思路是先估计出人脸的尺度,有了尺度信息就可以根据该尺

2017-08-21 11:06:29 2838 2

原创 人脸识别--SphereFace: Deep Hypersphere Embedding for Face Recognition

SphereFace: Deep Hypersphere Embedding for Face Recognition CVPR2017 https://github.com/wy1iu/sphereface针对人脸识别问题,当前的损失函数基本都基于 Euclidean margin ,这里我们提出基于 angular margin 的 angular softmax (A-Softmax) l

2017-08-18 10:51:49 7045 5

原创 语义分割--Fully Convolutional DenseNets for Semantic Segmentation

The One Hundred Layers Tiramisu: Fully Convolutional DenseNets for Semantic Segmentation CVPRW 2017 Code: https://github.com/SimJeg/FC-DenseNet本文将 DenseNets 按照 FCN 方式用于 语义分割首先看看 DenseNets 该网络主要由 d

2017-08-17 16:18:49 7633

原创 语义分割--Full-Resolution Residual Networks for Semantic Segmentation in Street Scenes

Full-Resolution Residual Networks for Semantic Segmentation in Street Scenes CVPR2017 Theano/Lasagne code:https://github.com/TobyPDE/FRRN针对语义分割问题,本文侧重于分割中的物体边界精度, precise boundary adherence 通过两个 str

2017-08-17 10:22:49 3120

原创 目标检测--Enhancement of SSD by concatenating feature maps for object detection

Enhancement of SSD by concatenating feature maps for object detection BMVC 2017 本文是对SSD 的改进,通过牺牲一点速度来提高精度,主要解决SSD 两个问题:1)同一目标多次检测,2)小目标检测率不高 改进的地方: 1)将不同尺度的 特征图 融合起来; 2)增加 feature pyramid 网络层的特征图数量

2017-08-16 16:41:38 2258

原创 目标检测--DSOD: Learning Deeply Supervised Object Detectors from Scratch

DSOD: Learning Deeply Supervised Object Detectors from Scratch ICCV2017 https://github.com/szq0214/DSOD针对目标检测问题,本文提出了不需要预训练模型的检测算法,可以看作 SSD + DenseNet 的结合以前的目标检测算法基本都是先在 ImageNet 数据库上进行预训练,然后再微调。这个微

2017-08-16 14:48:01 11068

原创 目标检测--SSD: Single Shot MultiBox Detector

SSD: Single Shot MultiBox Detector ECCV2016 https://github.com/weiliu89/caffe/tree/ssd 针对目标检测问题,本文取消了候选区域提取步骤,通过采用一系列设计在检测上得到较好的精度和速度contributions : 1)提出的 SSD 比 YOLO v1 速度快,精度好, 和 Faster R-CNN 精度差不

2017-08-15 16:06:31 3207

原创 目标检测--Focal Loss for Dense Object Detection

Focal Loss for Dense Object Detection ICCV2017 https://arxiv.org/abs/1708.02002本文算是用简单的方法解决复杂的问题了,好的思想是简单的针对目标检测,目前有两类主流算法: two-stage detectors 和 one-stage detectors, two-stage detectors 的精度好但是速度慢,on

2017-08-15 10:03:04 2312

原创 视频目标检测--Flow-Guided Feature Aggregation for Video Object Detection

Flow-Guided Feature Aggregation for Video Object Detection https://arxiv.org/abs/1703.10025 Our framework is principled, and on par with the best engineered systems winning the ImageNet VID challenge

2017-08-04 14:51:21 4981

原创 视频检测分割--Deep Feature Flow for Video Recognition

Deep Feature Flow for Video Recognition CVPR2017 Code: https://github.com/msracver/Deep-Feature-Flow基于单帧的目标检测和分割已经做的比较成熟,但是基于视频的目标检测和分割目前还是有问题的,最主要的问题就是直接将单帧的算法用于视频,计算量比较大,做不到实时。这里我们只对关键帧计算CNN特征提取,然后

2017-08-04 10:23:20 7295 2

原创 有方向的CNN--Oriented Response Networks

Oriented Response Networks CVPR2017 http://yzhou.work/ORN/ Code: https://github.com/ZhouYanzhao/ORN以前的CNN主要通过 pool 来得到小的角度变化不变性,本文设计了一个CNN 可以得到目标的角度信息,也可以做到完全的旋转不变性。 简单的来说,对以前学习到的滤波器 f 旋转 N个角度(例如,每

2017-08-03 16:58:08 3854

原创 语义分割--Efficient and Robust Deep Networks for Semantic Segmentation

Efficient and Robust Deep Networks for Semantic SegmentationCode: https://lmb.informatik.uni-freiburg.de/Publications/2017/OB17a/在 Up-Convolutional Networks 的基础上,提出了 Part-Net,Fast-Net,M-Net。 三个网络的侧重点不

2017-08-02 15:27:28 2891

原创 语义分割--Efficient Deep Models for Monocular Road Segmentation

Efficient Deep Models for Monocular Road Segmentation code: https://lmb.informatik.uni-freiburg.de/Publications/2016/OB16b/针对路面检测和分割问题,本文结合FCN 和 U-Net 提出一个网络,在速度和精度方面得到不错的效果 KITTI benchmark lane/ro

2017-08-02 10:50:19 3020 1

原创 查看 CUDA cudnn 版本 & 测试 cuda 和 cudnn 有效性

https://medium.com/@changrongko/nv-how-to-check-cuda-and-cudnn-version-e05aa21daf6ccuda 版本 cat /usr/local/cuda/version.txtcudnn 版本 cat /usr/local/cuda/include/cudnn.h | grep CUDNN_MAJOR -A 2

2017-08-01 15:27:09 244270 9

Accuracy of Laplacian Edge Detectors

The sources of error for the edge finding technique proposed by Marr and Hildreth (D. Marr and T. Poggio, Proc. R. Soc. London Ser. B204, 1979, 301–328; D. Marr and E. Hildreth, Proc. R. Soc. London Ser. B.207, 1980, 187–217) are identified, and the magnitudes of the errors are estimated, based on idealized models of the most common error producing situations. Errors are shown to be small for linear illuminations, as well as for nonlinear illuminations with a second derivative less than a critical value. Nonlinear illuminations are shown to lead to spurious contours under some conditions, and some fast techniques for discarding such contours are suggested.

2011-10-12

The Canny Edge Detector Revisited

Canny (1986) suggested that an optimal edge detector should maximize both signal-to-noise ratio and localization, and he derived mathematical expressions for these criteria. Based on these criteria, he claimed that the optimal step edge detector was similar to a derivative of a gaussian. However, Canny’s work suffers from two problems. First, his derivation of localization criterion is incorrect. Here we provide a more acurate localization criterion and derive the optimal detector from it. Second, and more seriously, the Canny criteria yield an infinitely wide optimal edge detector. The width of the optimal detector can however be limited by considering the effect of the neighbouring edges in the image. If we do so, we find that the optimal step edge detector, according to the Canny criteria, is the derivative of an ISEF filter, proposed by Shen and Castan (1992). In addition, if we also consider detecting blurred (or non-sharp) gaussian edges of different widths, we find that the optimal blurred-edge detector is the above optimal step edge detector convolved with a gaussian. This implies that edge detection must be performed at multiple scales to cover all the blur widths in the image. We derive a simple scale selection procedure for edge detection, and demonstrate it in one and two dimensions.

2011-08-11

OpenCV 2 Computer Vision Application Programming Cookbook

Overview of OpenCV 2 Computer Vision Application Programming Cookbook Teaches you how to program computer vision applications in C++ using the different features of the OpenCV library Demonstrates the important structures and functions of OpenCV in detail with complete working examples Describes fundamental concepts in computer vision and image processing Gives you advice and tips to create more effective object-oriented computer vision programs Contains examples with source code and shows results obtained on real images with detailed explanations and the required screenshots

2011-06-24

Learning based Symmetric Features Selection for Vehicle Detection

Learning based Symmetric Features Selection for Vehicle Detection This paper describes a symmetric features selection strategy based on statistical learning method for detecting vehicles with a single moving camera for autonomous driving. Symmetry is a good class of feature for vehicle detection, but the areas with high symmetry and threshold for segmentation is hard to be decided. Usually, the additional supposition is added artificially, and this will decrease the robustness of algorithms. In this paper, we focus on the problem of symmetric features selection using learning method for autonomous driving environment. Global symmetry and local symmetry are defined and used to construct a cascaded structure with a one-class classifier followed by a two-class classifier.

2011-04-11

Intensity and Edge-Based Symmetry Detection Applied to Car-Following

Intensity and Edge-Based Symmetry Detection Applied to Car-Following We present two methods for detecting symmetry in images, one based directly on the intensity values and another one based on a discrete representation of local orientation. A symmetry finder has been developed which uses the intensity-based method to search an image for compact regions which display some degree of mirror symmetry due to intensity similarities across a straight axis. In a different approach, we look at symmetry as a bilateral relationship between local orientations. A symmetryenhancing edge detector is presented which indicates edges dependent on the orientations at two different image positions. SEED, as we call it, is a detector element implemented by a feedforward network that holds the symmetry conditions. We use SEED to find the contours of symmetric objects of which we know the axis of symmetry from the intensity-based symmetry finder. The methods presented have been applied to the problem of visually guided car-following. Real-time experiments with a system for automatic headway control on motorways have been successful.

2011-04-11

Accurate Robust Symmetry Estimation

Accurate Robust Symmetry Estimation Stephen Smith and Mark Jenkinson There are various applications, both in medical and non-medical image analysis, which require the automatic detection of the line (2D images) or plane (3D) of reflective symmetry of objects. There exist relatively simple methods of finding reflective symmetry when object images are complete (i.e., completely symmetric and perfectly segmented from image “background”). A much harder problem is finding the line or plane of symmetry when the object of interest contains asymmetries, and may not have well defined edges.

2011-04-11

Approach of vehicle segmentation based on texture character

Approach of vehicle segmentation based on texture character

2011-04-01

Method of removing moving shadow based on texture

Method of removing moving shadow based on texture

2011-04-01

Environmentally Robust Motion Detection for Video Surveillance

Most video surveillance systems require to manually set a motion detection sensitivity level to generate motion alarms. The performance of motion detection algorithms, embedded in closed circuit television (CCTV) camera and digital video recorder (DVR), usually depends upon the preselected motion sensitivity level, which is expected to work in all environmental conditions. Due to the preselected sensitivity level, false alarms and detection failures usually exist in video surveillance systems. The proposed motion detection model based upon variational energy provides a robust detection method at various illumination changes and noise levels of image sequences without tuning any parameter manually. We analyze the structure mathematically and demonstrate the effectiveness of the proposed model with numerous experiments in various environmental conditions. Due to the compact structure and efficiency of the proposed model, it could be implemented in a small embedded system.

2011-03-17

Optimal multi-level thresholding using a two-stage Otsu optimization approach

Otsu’s method of image segmentation selects an optimum threshold by maximizing the between-class variance in a gray image. However, this method becomes very time-consuming when extended to a multi-level threshold problem due to the fact that a large number of iterations are required for computing the cumulative probability and the mean of a class. To greatly improve the efficiency of Otsu’s method, a new fast algorithm called the TSMO method (Two-Stage Multithreshold Otsu method) is presented. The TSMO method outperforms Otsu’s method by greatly reducing the iterations required for computing the between-class variance in an image. The experimental results show that the computational time increases exponentially for the conventional Otsu method with an average ratio of about 76. For TSMO-32, the maximum computational time is only 0.463 s when the class number M increases from two to six with relative errors of less than 1% when compared to Otsu’s method. The ratio of computational time of Otsu’s method to TSMO-32 is rather high, up to 109,708, when six classes (M = 6) in an image are used. This result indicates that the proposed method is far more efficient with an accuracy equivalent to Otsu’s method. It also has the advantage of having a small variance in runtimes for different test images.

2011-03-17

A Background Reconstruction Method Based on Double-background

In this paper, we show a new method to reconstruct and update the background. This approach is based on double-background. We use the statistical information of the pixel intensity to construct a background that represents the status during a long time, and construct another background with feedback information in motion detection that represents the recent changes at a short time. This couple of background images is fused to construct and update the background image used to motion detection. The background reconstruction algorithm can perform well on the tests that we have applied it to.

2011-03-17

Statistical Change Detection by the Pool Adjacent Violators Algorithm

In this paper we present a statistical change detection approach aimed at being robust with respect to the main disturbance factors acting in real-world applications, such as illumination changes, camera gain and exposure variations, noise. We rely on modeling the effects of disturbance factors on images as locally order-preserving transformations of pixel intensities plus additive noise. This allows us to identify within the space of all the possible image change patterns the subspace corresponding to disturbance factors effects. Hence, scene changes can be detected by a-contrario testing the hypothesis that the measured pattern is due to disturbance factors, that is by computing a distance between the pattern and the subspace. By assuming additive gaussian noise, the distance can be computed within a maximum likelihood non-parametric isotonic regression framework. In particular, the projection of the pattern onto the subspace is computed by an O(N) iterative procedure known as Pool Adjacent Violators algorithm.

2011-03-17

Cooperative Fusion of Stereo and Motion

Cooperative Fusion of Stereo and Motion This paper presents a new matching algorithm based on cooperative fusion of stereo and motion cues. In this algorithm, stereo disparity and image flow values are recovered from two successive pairs of stereo images by solving the stereo and motion corresponde

2011-03-09

A Treatise on Mathematical Theory of Elasticity (1944)(ISBN 0486601749)

Love, A Treatise on Mathematical Theory of Elasticity (1944)(ISBN 0486601749).djvu 第三部分(共三部分)

2011-02-27

A Treatise on Mathematical Theory of Elasticity (1944)(ISBN 0486601749)

Love, A Treatise on Mathematical Theory of Elasticity (1944)(ISBN 0486601749).djvu 第二部分(共三部分)

2011-02-27

Love, A Treatise on Mathematical Theory of Elasticity (1944)(ISBN 0486601749)

Love, A Treatise on Mathematical Theory of Elasticity (1944)(ISBN 0486601749) 第一部分(共三部分)

2011-02-27

Computation of Real-Time Optical Flow Based on Corner Features

This paper describes an approach to real-time optical flow computation that combines the corner features and pyramid Lucas-Kanade. Corners instead of all the points in the image are taken into optical flow computation, which could reduce the amount of calculation to a large extend. The experiment has shown that using this optical flow algorithm to track targets is effective and could meet the requirements of real-time applications.

2011-02-24

II-LK – A Real-Time Implementation for Sparse Optical Flow

In this paper we present an approach to speed up the computation of sparse optical flow fields by means of integral images and provide implementation details. Proposing a modification of the Lucas-Kanade energy functional allows us to use integral images and thus to speed up the method notably while affecting only slightly the quality of the computed optical flow. The approach is combined with an efficient scanline algorithm to reduce the computation of integral images to those areas where there are features to be tracked. The proposed method can speed up current surveillance algorithms used for scene description and crowd analysis.

2011-02-24

Medical Image Reconstruction A Conceptual Tutorial --pdf

Medical Image Reconstruction: A Conceptual Tutorial" introduces the classical and modern image reconstruction technologies, such as two-dimensional (2D) parallel-beam and fan-beam imaging, three-dimensional (3D) parallel ray, parallel plane, and cone-beam imaging. This book presents both analytical and iterative methods of these technologies and their applications in X-ray CT (computed tomography), SPECT (single photon emission computed tomography), PET (positron emission tomography), and MRI (magnetic resonance imaging). Contemporary research results in exact region-of-interest (ROI) reconstruction with truncated projections, Katsevich's cone-beam filtered backprojection algorithm, and reconstruction with highly undersampled data with l0-minimization are also included.

2011-02-24

Extraction and recognition of license plates of motorcycles and vehicles on highways

Extraction and recognition of license plates of motorcycles and vehicles on highways

2011-02-22

High Performance Implementation of License Plate Recognition in Image Sequences

High Performance Implementation of License Plate Recognition in Image Sequences

2011-02-22

Vs-star-- A visual interpretation system for visual surveillance

Vs-star-- A visual interpretation system for visual surveillance

2011-02-22

Robust fragments-based tracking with adaptive feature selection

Robust fragments-based tracking with adaptive feature selection

2011-02-22

Robust and automated unimodal histogram thresholding and potential applications

Robust and automated unimodal histogram thresholding and potential applications

2011-02-22

角点检测方法研究-- 毛雁明, 兰美辉

角点检测方法研究---根据实现方法不同可将角点检测方法分为两大类:基于边缘的角点检测方法与基于灰度变化的角点检测方法,并对现有的角点检测方法作了较为详细的分析与比较,指出角点检测技术的研究与发展方向.

2011-02-22

图像融合中角点检测技术研究

图像融合中角点检测技术研究--图像融合中角点检测技术研究

2011-02-22

Fast image region growing

Fast image region growing---Fast image region growing

2011-02-22

Simple Low Level Features for Image Analysis

Simple Low Level Features for Image Analysis

2011-02-22

Direct methods for sparse matrices

second edition 2017, Oxford University Press

2024-04-07

百面机器学习.pdf

收录了超过100道机器学习算法工程师的面试题目和解答,本书将从特征工程、模型评估、降维等经典机器学习领域出发,构建一个算法工程师必-备的知识体系。其中大部分源于Hulu算法研究岗位的真实场景。

2019-06-01

CLIP-Q CVPR2018 code

CLIP-Q: Deep Network Compression Learning by In-Parallel Pruning-Quantization,CVPR2018 code

2018-10-30

Vehicle model recognition from frontal view image measurements

This paper deals with a novel vehicle manufacturer and model recognition scheme, which is enhanced by color recognition for more robust results. A probabilistic neural network is assessed as a classifier and it is demonstrated that relatively simple image processing measurements can be used to obtain high performance vehicle authentication. The proposed system is assisted by a previously developed license plate recognition, a symmetry axis detector and an image phase congruency calculation modules. The reported results indicate a high recognition rate and a fast processing time, making the system suitable for real-time applications.

2011-10-15

Vehicle Detection and Tracking in Car Video Based on Motion Model

Vehicle Detection and Tracking in Car Video Based on Motion Model--This work aims at real-time in-car video analysis to detect and track vehicles ahead for safety, auto-driving, and target tracing. This paper describes a comprehensive approach to localize target vehicles in video under various environmental conditions. The extracted geometry features from the video are projected onto a 1D profile continuously and are tracked constantly. We rely on temporal information of features and their motion behaviors for vehicle identification, which compensates for the complexity in recognizing vehicle shapes, colors, and types. We model the motion in the field of view probabilistically according to the scene characteristic and vehicle motion model. The Hidden Markov Model is used for separating target vehicles from background, and tracking them probabilistically. We have investigated videos of day and night on different types of roads, showing that our approach is robust and effective in dealing with changes in environment and illumination, and that real time processing becomes possible for vehicle borne cameras.

2011-10-15

Projection and Least Square Fitting

Projection and Least Square Fitting with Perpendicular Offsets based Vehicle License Plate Tilt Correction

2011-10-15

An Algorithm for License Plate Recognition Applied to ITS

An algorithm for license plate recognition (LPR) applied to the intelligent transportation system is proposed on the basis of a novel shadow removal technique and character recognition algorithms. This paper has two major contributions. One contribution is a new binary method, i.e., the shadow re- moval method, which is based on the improved Bernsen algorithm combined with the Gaussian filter. Our second contribution is a character recognition algorithm known as support vector machine (SVM) integration. In SVM integration, character features are extracted from the elastic mesh, and the entire address character string is taken as the object of study, as opposed to a single character. This paper also presents improved techniques for im- age tilt correction and image gray enhancement. Our algorithm is robust to the variance of illumination, view angle, position, size, and color of the license plates when working in a complex environment. The algorithm was tested with 9026 images, such as natural-scene vehicle images using different backgrounds and ambient illumination particularly for low-resolution images. The license plates were properly located and segmented as 97.16%and 98.34%, respectively. The optical character recognition system is the SVM integration with different character features, whose performance for numerals, Kana, and address recognition reached 99.5%, 98.6%, and 97.8%, respectively. Combining the preceding tests, the overall performance of success for the license plate achieves 93.54% when the system is used for LPR in various complex conditions

2011-10-15

A Review of Computer Vision Techniques for the Analysis of Urban Traffic

Automatic video analysis from urban surveillance cameras is a fast-emerging field based on computer vision techniques. We present here a comprehensive review of the state-of-the-art computer vision for traffic video with a critical analysis and an outlook to future research directions. This field is of increasing relevance for intelligent transport systems (ITSs). The decreasing hardware cost and, therefore, the increasing de- ployment of cameras have opened a wide application field for video analytics. Several monitoring objectives such as congestion, traffic rule violation, and vehicle interaction can be targeted using cameras that were typically originally installed for human oper- ators. Systems for the detection and classification of vehicles on highways have successfully been using classical visual surveillance techniques such as background estimation and motion tracking for some time. The urban domain is more challenging with respect to traffic density, lower camera angles that lead to a high degree of occlusion, and the variety of road users. Methods from object categorization and 3-D modeling have inspired more advanced techniques to tackle these challenges. There is no commonly used data set or benchmark challenge, which makes the direct com- parison of the proposed algorithms difficult. In addition, evalu- ation under challenging weather conditions (e.g., rain, fog, and darkness) would be desirable but is rarely performed. Future work should be directed toward robust combined detectors and classifiers for all road users, with a focus on realistic conditions during evaluation.

2011-10-15

On Improving the Efficiency of Tensor Voting

This paper proposes two alternative formulations to reduce the high computational complexity of tensor voting, a robust perceptual grouping technique used to extract salient information from noisy data. The first scheme consists of numerical approximations of the votes, which have been derived from an in-depth analysis of the plate and ball voting processes. The second scheme simplifies the formulation while keeping the same perceptual meaning of the original tensor voting: The stick tensor voting and the stick component of the plate tensor voting must reinforce surfaceness, the plate components of both the plate and ball tensor voting must boost curveness, whereas junctionness must be strengthened by the ball component of the ball tensor voting. Two new parameters have been proposed for the second formulation in order to control the potentially conflictive influence of the stick component of the plate vote and the ball component of the ball vote. Results show that the proposed formulations can be used in applications where efficiency is an issue since they have a complexity of order O(1). Moreover, the second proposed formulation has been shown to be more appropriate than the original tensor voting for estimating saliencies by appropriately setting the two new parameters.

2011-10-11

Selecting Critical Patterns Based on Local Geometrical

Pattern selection methods have been traditionally developed with a dependency on a specific classifier. In contrast, this paper presents a method that selects critical patterns deemed to carry essential information applicable to train those types of classifiers which require spatial information of the training data set. Critical patterns include those edge patterns that define the boundary and those border patterns that separate classes. The proposed method selects patterns from a new perspective, primarily based on their location in input space. It determines class edge patterns with the assistance of the approximated tangent hyperplane of a class surface. It also identifies border patterns between classes using local probability. The proposed method is evaluated on benchmark problems using popular classifiers, including multilayer perceptrons, radial basis functions, support vector machines, and nearest neighbors. The proposed approach is also compared with four state-of-the-art approaches and it is shown to provide similar but more consistent accuracy from a reduced data set. Experimental results demonstrate that it selects patterns sufficient to represent class boundary and to preserve the decision surface.

2011-10-11

Fast LOG Filtering Using Recursive Filters

Marr and Hildreth's theory of LoG filtering with multiple scales has been extensively elaborated. One problem with LoG filtering is that it is very time-consuming, especially with a large size of filters. This paper presents a recursive convolution scheme for LoG filtering and a fast algorithm to extract zero-crossings. It has a constant computational complexity per pixel and is independent of the size of the filter. A line buffer is used to determine the locations of zero-crossings along with filtering hence avoiding the need for an additional convolution and extra memory units. Various images have been tested

2011-10-11

A discrete expression of Canny's criteria for step

Optimal filters for edge detection are usually developed in the continuous domain and then transposed by sampling to the discrete domain. Simpler filters are directly defined in the discrete domain. We define criteria to compare filter performances in the discrete domain. Canny has defined (1983, 1986) three criteria to derive the equation of an optimal filter for step edge detection: good detection, good localization, and low-responses multiplicity. These criteria seem to be good candidates for filter comparison. Unfortunately, they have been developed in the continuous domain, and their analytical expressions cannot be used in the discrete domain. We establish three criteria with the same meaning as Canny's.

2011-10-11

空空如也

TA创建的收藏夹 TA关注的收藏夹

TA关注的人

提示
确定要删除当前文章?
取消 删除