keras+resnet34实现车牌识别

1.使用PIL和opencv生成车牌图像数据

from PIL import ImageFont,Image,ImageDraw
import cv2
import numpy as np
import os
from math import *

#创建 生成车牌图像数据 的类
index = {"京": 0, "沪": 1, "津": 2, "渝": 3, "冀": 4, "晋": 5, "蒙": 6, "辽": 7, "吉": 8, "黑": 9, "苏": 10, "浙": 11, "皖": 12,
         "闽": 13, "赣": 14, "鲁": 15, "豫": 16, "鄂": 17, "湘": 18, "粤": 19, "桂": 20, "琼": 21, "川": 22, "贵": 23, "云": 24,
         "藏": 25, "陕": 26, "甘": 27, "青": 28, "宁": 29, "新": 30, "0": 31, "1": 32, "2": 33, "3": 34, "4": 35, "5": 36,
         "6": 37, "7": 38, "8": 39, "9": 40, "A": 41, "B": 42, "C": 43, "D": 44, "E": 45, "F": 46, "G": 47, "H": 48,
         "J": 49, "K": 50, "L": 51, "M": 52, "N": 53, "P": 54, "Q": 55, "R": 56, "S": 57, "T": 58, "U": 59, "V": 60,
         "W": 61, "X": 62, "Y": 63, "Z": 64}
 
chars = ["京", "沪", "津", "渝", "冀", "晋", "蒙", "辽", "吉", "黑", "苏", "浙", "皖", "闽", "赣", "鲁", "豫", "鄂", "湘", "粤", "桂",
             "琼", "川", "贵", "云", "藏", "陕", "甘", "青", "宁", "新", "0", "1", "2", "3", "4", "5", "6", "7", "8", "9", "A",
             "B", "C", "D", "E", "F", "G", "H", "J", "K", "L", "M", "N", "P", "Q", "R", "S", "T", "U", "V", "W", "X",
             "Y", "Z"
             ]

def r(val):
    return int(np.random.random()*val)
def GenCh(f,val):#生成中文字符,f是中文字体对象,
    img=Image.new('RGB',(45,70),(255,255,255))#创建中文字符区域的全白画布,中文一般要方正一些,所以画布设置大一点,等下再resize
    draw=ImageDraw.Draw(img)#对画布创建画画对象
    draw.text((0,3),val,(0,0,0),font=f)#画画对象画出规定字体的黑色的规定val【0】文字,在左上角位置是(0,3)的点开始
    img =  img.resize((23,70))
    A=np.array(img)#图画转换成array格式
    return A
def GenCh1(f,val):#生成英文和数字字符,f是中文字体对象,
    img=Image.new('RGB',(23,70),(255,255,255))
    draw=ImageDraw.Draw(img)
    draw.text((0,2),val,(0,0,0),font=f)#画画对象在左上角(0,2)出开始画出val,颜色全黑,字体是f
    A=np.array(img)
    return A
def rot(img,angle,shape,max_angle):#透视畸变 
    size_o=[shape[1],shape[0]]#cv读的(h,w)换成[w,h]
    size=(shape[1]+int(shape[0]*cos((float(max_angle)/180)*3.14)),shape[0])#【变化后的w,h】
    interval=abs(int(sin(float(angle)/180)*3.14)*shape[0])#h变换的绝对值
    pts1=np.float32([[0,0],[0,size_o[1]],[size_o[0],0],[size_o[0],size_o[1]]])#(00,0h,w0,wh)
    if (angle>0):
        pts2=np.float32([[interval,0],[0,size[1]],[size[0],0],[size[0]-interval,size_o[1]]])
    else:
        pts2 = np.float32([[0,0],[interval,size[1]  ],[size[0]-interval,0  ],[size[0],size_o[1]]])
    M=cv2.getPerspectiveTransform(pts1,pts2)#根据两幅图的四个坐标点计算透视矩阵
    dst=cv2.warpPerspective(img,M,size)#img再M矩阵的变化下生成size大小的变化图
    return dst
def rotRandrom(img,factor,shape):#仿射畸变
    pts1=np.float32([[0,0],[0,shape[0]],[shape[1],0],[shape[1],shape[0]]])#00,0h,w0,wh
    pts2=np.float32([[r(factor),r(factor)],[r(factor),shape[0]-r(factor)],[shape[1]-factor,r(factor)],[shape[1]-r(factor),shape[0]-r(factor)]])
    M=cv2.getPerspectiveTransform(pts1,pts2)
    dst=cv2.warpPerspective(img,M,shape)
    return dst
def tfactor(img):#添加饱和度光照的噪声
    hsv=cv2.cvtColor(img,cv2.COLOR_BGR2HSV)
    hsv[:,:,0] = hsv[:,:,0]*(0.8+ np.random.random()*0.2)
    hsv[:,:,1] = hsv[:,:,1]*(0.3+ np.random.random()*0.7)
    hsv[:,:,2] = hsv[:,:,2]*(0.2+ np.random.random()*0.8)
    img=cv2.cvtColor(hsv,cv2.COLOR_HSV2BGR)
    return img
def random_envirment(img,data_set):#添加自然环境噪声
    index=r(len(data_set))
    env=cv2.imread(data_set[index])#随机读出一张环境噪声图片
    env=cv2.resize(env,(img.shape[1],img.shape[0]))#改变环境噪声图片的大下
    bak=(img==0)
    bak=bak.astype(np.uint8)*255#图片的黑色部分变成白色部分
    inv=cv2.bitwise_and(bak,env)#图片的非黑部分和噪声求and,
    img=cv2.bitwise_or(inv,img)
    return img
def AddGauss(img,level):#添加高斯模糊
    return cv2.blur(img,(level*2+1,level*2+1))
def AddNoiseSingleChannel(single):#高斯噪声
    diff=255-single.max()
    noise=np.random.normal(0,1+r(6),single.shape)
    noise = (noise - noise.min())/(noise.max()-noise.min())
    noise= diff*noise
    noise= noise.astype(np.uint8)
    dst = single + noise
    return dst
def addNoise(img,sdev = 0.5,avg=10):#每个通道随机添加高斯噪声
    img[:,:,0] =  AddNoiseSingleChannel(img[:,:,0])
    img[:,:,1] =  AddNoiseSingleChannel(img[:,:,1])
    img[:,:,2] =  AddNoiseSingleChannel(img[:,:,2])
    return img

class GenPlate:#生成车牌图像数据 的类
    def __init__(self,fontCh,fontEng,NoPlates):
        self.fontC=ImageFont.truetype(fontCh,43,0)#创建中文字体对象,fontCh是字体文件地址,43是字体大小,规定文字字体
        self.fontE=ImageFont.truetype(fontEng,60,0)
        self.img=np.array(Image.new('RGB',(226,70),(255,255,255)))#创建(226,70)大小的全白图像,并转换成array格式
        self.bg=cv2.resize(cv2.imread('./input_data/images/template.bmp'),(226,70))#读取出背景模板图
        self.noplates_path=[]
        for parent,parent_folder,filenames in os.walk(NoPlates):#NoPlates的上级目录,NoPlates的子目录(没有),NoPlates的子文件
            for filename in filenames:
                path=parent+'/'+filename
                self.noplates_path.append(path)#环境噪声图片
    def draw(self,val):
        offset=2
        self.img[0:70,offset+8:offset+8+23]=GenCh(self.fontC,val[0])#再self.img画布上画出中文字
        self.img[0:79,offset+8+23+6:offset+8+23+6+23]=GenCh1(self.fontE,val[1])#英文字
        for i in range(5):#画出5个数字
            base=offset+8+23+6+23+17+ i*23+ i*6
            self.img[0:70,base:base+23]=GenCh1(self.fontE,val[i+2])
        return self.img#画出背景白色,文字黑色的文字内容
    def generate(self,text):
        if len(text)==7:
            fg=self.draw(text)#调用draw函数,画出这7个字符)
            fg=cv2.bitwise_not(fg)#图像的array按位取反:黑色背景,白色文字
            com=cv2.bitwise_or(fg,self.bg)#再与背景图片取或,则生成背景是背景图片,前景是白色文字的图片array格式
            com=rot(com,r(60)-30,com.shape,30)#透视畸变效果(img,angle,shape,max_angle)
            com = rotRandrom(com,10,(com.shape[1],com.shape[0]))#仿射畸变
            com = tfactor(com)#添加饱和度光照的噪声
            com = random_envirment(com,self.noplates_path)#环境图片的噪声
            com = AddGauss(com, 1+r(4))#高斯模糊
            com = addNoise(com)#高斯噪声
            return com
    def genPlateString(self,pos,val):#随机生成(中文 英文 数字*5)的字符
        #pos!=-1时,读取出val值就是车牌值************************pos=-1时,val随机,主要是生成车牌
        plateStr=''
        if (pos!=-1):#
            plateStr+=val#读出你想要生成的车牌号
        else:#随机生成车牌号
            for cpos in range(7):
                if cpos==0:
                    plateStr+=chars[r(31)]
                elif cpos==1:
                    plateStr+=chars[41+r(24)]
                else:
                    plateStr+=chars[31+r(10)]
        return plateStr
    def genBatch(self,batchSize,size):#生成batch——size个车牌数据
        for i in range(batchSize):
            plateStr=self.genPlateString(-1,-1)
            print(plateStr)
            img=self.generate(plateStr)
            
            img=cv2.resize(img,size)
            filename=str(plateStr)+'.jpg'
            cv2.imencode('.jpg',img)[1].tofile(filename)#此处解决中文名乱码
            

if __name__=='__main__':
    G=GenPlate("./input_data/font/platech.ttf",'./input_data/font/platechar.ttf',"./input_data/NoPlates")
    G.genBatch(10,(224,224))

2.使用keras生成resnet34模型

#2.构建网络模型
from keras.models import Model
from keras.layers import Input,Dropout,BatchNormalization,Conv2D,MaxPooling2D,AveragePooling2D,concatenate,Activation,ZeroPadding2D,add,Flatten,Dense
import numpy as np
seed=1
np.random.seed(seed)
import matplotlib.pyplot as plt
%matplotlib inline
characters="京沪津渝冀晋蒙辽吉黑苏浙皖闽赣鲁豫鄂湘粤桂琼川贵云藏陕甘青宁新0123456789ABCDEFGHJKLMNPQRSTUVWXYZ"
width,height,n_len,n_classes=224,224,7,len(characters)
def gen(batch_size=32):
    X = np.zeros((batch_size, height, width, 3), dtype=np.uint8)
    y = [np.zeros((batch_size, n_classes), dtype=np.uint8) for i in range(n_len)]
   # generator = ImageCaptcha(width=width, height=height)
    while True:
        for i in range(batch_size):
            a=G.genPlateString(-1,-1)
            b=G.generate(a)
            img = cv2.resize(b,(224,224))
            X[i] = img.astype('float32')
            for j, ch in enumerate(a):
                y[j][i, :] = 0
                y[j][i, characters.find(ch)] = 1
        yield X, y
 
def decode(y):
    y = np.argmax(np.array(y), axis=2)[:,0]
    return ''.join([characters[x] for x in y])
  
"""
使用者三段代码测试生成结果,以及解码结果
X, y = next(gen(1))
plt.imshow(X[0])
print(decode(y))
plt.title(decode(y))
 
"""
def gen(batch_size=32):#生成数据
    x=np.zeros((batch_size,height,width,3),dtype=np.uint8)
    #y=[np.zeros((batch_size,n_classes),dtype=np.uint8) for i in range (n_len)]
    y=np.zeros((batch_size,n_len,n_classes),dtype=np.uint8)
    for i in range(batch_size):
        string=G.genPlateString(-1,1)
        img=G.generate(string)
        img=cv2.resize(img,(224,224))
        x[i]=img.astype('float32')
        for j,ch in enumerate(string):
            y[i][j,:]=0
            y[i][j,characters.find(ch)]=1
    yield x,y
def decode(y):
 
    y=np.argmax(np.array(y),axis=2)[0]
    return ''.join([characters[x] for x in y])
#x,y=next(gen(1))
#plt.imshow(x[0])
#print(decode(y))
#plt.title(decode(y))
#resnet34网络
def Conv2d_BN(x,nb_filter,kernel_size,strides=(1,1),padding='same',name=None):
    if name is not None:
        bn_name=name+'_bn'
        conv_name=name+'_conv'
    else:
        bn_name=None
        conv_name=None
    x=Conv2D(nb_filter,kernel_size,padding=padding,strides=strides,activation='relu',name=conv_name)(x)
    x=BatchNormalization(name=bn_name)(x)
    return x
def Conv_Block(input_layer,nb_filter,kernel_size,strides=(1,1),with_conv_shortcut=False):
    x=Conv2d_BN(input_layer,nb_filter=nb_filter,kernel_size=kernel_size,strides=strides,padding='same')
    x=Conv2d_BN(x,nb_filter=nb_filter,kernel_size=kernel_size,padding='same')
    if with_conv_shortcut:#残差边
        shortcut=Conv2d_BN(input_layer,nb_filter=nb_filter,strides=strides,kernel_size=kernel_size)
        x=add([x,shortcut])#残差边卷积后相加,缩小网络输出图像尺寸
        return x
    else:
        x=add([x,input_layer])#残差边直接相加,加深网络卷积深度
        return x
input_x=Input(shape=(224,224,3))
x=ZeroPadding2D((3,3))(input_x)
x=Conv2d_BN(x,nb_filter=64,kernel_size=(7,7),strides=(2,2),padding='valid')
x=MaxPooling2D(pool_size=(3,3),strides=(2,2),padding='same')(x)#56,56,64
x = Conv_Block(x,nb_filter=64,kernel_size=(3,3)) 
x = Conv_Block(x,nb_filter=64,kernel_size=(3,3)) 
x = Conv_Block(x,nb_filter=64,kernel_size=(3,3))  #56,56,34
x = Conv_Block(x,nb_filter=128,kernel_size=(3,3),strides=(2,2),with_conv_shortcut=True) 
x = Conv_Block(x,nb_filter=128,kernel_size=(3,3)) 
x = Conv_Block(x,nb_filter=128,kernel_size=(3,3)) 
x = Conv_Block(x,nb_filter=128,kernel_size=(3,3)) #28,28,128
x = Conv_Block(x,nb_filter=256,kernel_size=(3,3),strides=(2,2),with_conv_shortcut=True) 
x = Conv_Block(x,nb_filter=256,kernel_size=(3,3)) 
x = Conv_Block(x,nb_filter=256,kernel_size=(3,3)) 
x = Conv_Block(x,nb_filter=256,kernel_size=(3,3)) 
x = Conv_Block(x,nb_filter=256,kernel_size=(3,3)) 
x = Conv_Block(x,nb_filter=256,kernel_size=(3,3))  #14,14,256
x = Conv_Block(x,nb_filter=512,kernel_size=(3,3),strides=(2,2),with_conv_shortcut=True) 
x = Conv_Block(x,nb_filter=512,kernel_size=(3,3)) 
x = Conv_Block(x,nb_filter=512,kernel_size=(3,3))#7,7,512 
x = AveragePooling2D(pool_size=(7,7))(x) #1,1,512
x=Flatten()(x)#512
#x = Dense(1000,activation='softmax')(x) 
x = [Dense(n_classes, activation='softmax', name='P%d'%(i+1))(x) for i in range(7)]#有7个输出结果,所以7个Dense层和x分别连接
model = Model(inputs=input_x,outputs=x) 
model.compile(loss='categorical_crossentropy',optimizer='sgd',metrics=['accuracy']) 
model.summary()
model.fit_generator(gen(), epochs=1000, steps_per_epoch=20,validation_data=gen(),validation_steps=1)       
#保存模型图
from keras.utils import plot_model
plot_model(model, to_file='model.png',show_shapes='True')
#测试下结果
X, y = next(gen(1))
y_pred = model.predict(X)
print(X)
print("ddddd")
print(X[0])
print(decode(y))
print(decode(y_pred))
plt.title('real: %s\npred:%s'%(decode(y), decode(y_pred)))
plt.imshow(X[0], cmap='gray')
#这里显示是和正常的颜色不一样,这是因为,plt读取的通道顺序和cv2的通道顺序是不同的
#保存模型和参数
model.save('resnet34_model.h5')
score=model.evaluate(X,y,verbose=0)
print("test score=",score[0])
print("test accuracy=",score[1])

3.利用模型进行实际预测输出

#进行实际预测
from keras.models import load_model
model=load_model("resnet34_model.h5")
print("导入模型完成")
print("读取图片")
#pic = Image.open("./宁R46974.jpg")
#pic.show()
img = cv2.imread('./x.jpg')#地址不能有中文
img=img[np.newaxis,:,:,:]#图片是三维的但是训练时是转换成4维了所以需要增加一个维度
predict = model.predict(img)
print("车牌号为:",decode(predict))

参考:https://github.com/Haveoneriver/License-Plate-Recognition-Items

https://github.com/szad670401/end-to-end-for-chinese-plate-recognition

  • 1
    点赞
  • 19
    收藏
    觉得还不错? 一键收藏
  • 打赏
    打赏
  • 0
    评论
目标识别是计算机视觉一个重要的研究领域,由此延伸出的车辆型号识别具有重 要的实际应用价值,特别是在当今交通状况复杂的大城市,智能交通系统成为发展趋 势,这离不开对车辆型号进行识别和分类的工作,本文围绕如何利用计算机视觉的方 法进行车辆型号的识别和分类展开了一系列研究: 本文对当前的目标识别和分类的特征和算法做了总结和归纳。分析比较了作为图 像特征描述常见的特征算子,总结归纳了他们的提取方法、特征性能以及相互之间的 关联。另外,介绍了在目标识别工作中常用的分类方法,阐述了他们各自的原理和工作 方法。研究了深度神经网络的理论依据,分析比较了深度神经网络不同的特征学习方 法,以及卷积神经网络的训练方法。分析比较不同特征学习方法的特点选取 k-means 作为本文使用的特征学习方法,利用卷积神经网络结构搭建深度学习模型,进行车辆 车型识别工作。 本文为了测试基于深度学习的车辆型号分类算法的性能在 30 个不同型号共 7158 张图片上进行实验;并在相同数据上利用改进了的 SIFT 特征匹配的算法进行对比实验; 进过实验测试,深度学习方法在进行车型分类的实验中取得 94%的正确率,并在与 SIFT 匹配实验结果对比后进一步证实:深度学习的方法能够应用在车辆型号识别领域

“相关推荐”对你有帮助么?

  • 非常没帮助
  • 没帮助
  • 一般
  • 有帮助
  • 非常有帮助
提交
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包

打赏作者

我是小z呀

你的鼓励将是我创作的最大动力

¥1 ¥2 ¥4 ¥6 ¥10 ¥20
扫码支付:¥1
获取中
扫码支付

您的余额不足,请更换扫码支付或充值

打赏作者

实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值