基于机器学习的数字电表数字识别二(源码+数据集)

基于机器学习的数字电表数字识别二(源码+数据集),第1张

上一期用了传统的模板匹配的算法对单字符进行识别,本期就来写简单的机器学习的方式,主要是SVM和CNN,入门级别,可供初学者参考,文章最后给出整个开源的代码和数据集,参考学习。
数据标签为0-9十个文件夹,里面分别包含切割出来的字符图片。

训练集和测试集按如此划分好,接下来就构建机器学习的网络模型了,首先给出SVM。

import numpy as np
from tqdm import tqdm
import glob
import os
from PIL import Image
from sklearn.svm import SVC
from sklearn.preprocessing import StandardScaler
from sklearn.model_selection import GridSearchCV, train_test_split,learning_curve
from sklearn.externals import joblib



def img2vec(filename):
    dataVec = np.zeros((1,800))
    img = Image.open(filename)
    img = img.crop((0,0,20,40))
    img = np.array(img)
    for i in range(40):
        for j in range(20):
            dataVec[0,20*i+j] = img[i][j]
    return dataVec

def make_dataset(parent_dir,sub_dirs,file_ext="*.jpg"):
    label = []
    feature = []
    for sub_dir in sub_dirs:
        for fn in tqdm(glob.glob(os.path.join(parent_dir,sub_dir,file_ext))):
            label_name =  fn.split("\")[-2]
            label_name = label_name.split("/")[-1]
            imgdata = img2vec(fn)
            label.extend([label_name])
            feature.extend([imgdata])
    return [feature,label]


parent_dir = "./data/train"
sub_dirs = np.array(['0','1','2','3','4','5','6','7','8','9'])
temp = make_dataset(parent_dir,sub_dirs)

temp = np.array(temp)
data = temp.transpose()
# 获取特征
X = np.vstack(data[:, 0])
# 获取标签
Y = np.array(data[:, 1])
print('X的特征尺寸是:',X.shape)
print('Y的特征尺寸是:',Y.shape)
scaler = StandardScaler()
x_std = scaler.fit_transform(X)  # 标准化
x_train, x_test, y_train, y_test = train_test_split(x_std, Y, test_size=.2)
# rbf核函数,设置数据权重
svc = SVC(kernel='rbf', class_weight='balanced',)
c_range = np.logspace(-5, 4, 10, base=2)
gamma_range = np.logspace(-9, 0, 10, base=2)
# 网格搜索交叉验证的参数范围,cv=3,3折交叉,n_jobs=-1,多核计算
param_grid = [{'kernel': ['rbf'], 'C': c_range, 'gamma': gamma_range}]
grid = GridSearchCV(svc, param_grid, cv=3, n_jobs=-1)
# 训练模型
clf = grid.fit(x_train, y_train)
joblib.dump(clf,"./model_save/svm.pkl")


模型保存再文件夹里,方便后续推理。
下面再给出CNN的baseline代码:

from tensorflow.keras.models import Sequential
from tensorflow.keras.layers import Conv2D, Flatten, Dense, MaxPool2D, Dropout
from tensorflow.keras.utils import to_categorical
import numpy as np
from PIL import Image
from tqdm import tqdm
import glob
import os
from sklearn.preprocessing import StandardScaler
from sklearn.model_selection import train_test_split
from sklearn.externals import joblib
def img2vec(filename):
    dataVec = np.zeros((1,800))
    img = Image.open(filename)
    img = img.crop((0,0,20,40))
    img = np.array(img)
    for i in range(40):
        for j in range(20):
            dataVec[0,20*i+j] = img[i][j]
    return dataVec

def make_dataset(parent_dir,sub_dirs,file_ext="*.jpg"):
    label = []
    feature = []
    for sub_dir in sub_dirs:
        for fn in tqdm(glob.glob(os.path.join(parent_dir,sub_dir,file_ext))):
            label_name =  fn.split("\")[-2]
            label_name = label_name.split("/")[-1]
            imgdata = img2vec(fn)
            label.extend([label_name])
            feature.extend([imgdata])
    return [feature,label]

parent_dir = "./data/train"
sub_dirs = np.array(['0','1','2','3','4','5','6','7','8','9'])
temp = make_dataset(parent_dir,sub_dirs)

temp = np.array(temp)
data = temp.transpose()
# 获取特征
X = np.vstack(data[:, 0])

# 获取标签
Y = np.array(data[:, 1])
Y = to_categorical(Y)

print('X的特征尺寸是:',X.shape)
print('Y的特征尺寸是:',Y.shape)

x_train, x_test, y_train, y_test = train_test_split(X, Y, test_size=0.2)
model = Sequential()
x_train = x_train.reshape(-1,40,20,1)
x_test = x_test.reshape(-1,40,20,1)

# 输入的大小
input_dim = (40, 20, 1)

model.add(Conv2D(64, (3, 3), padding = "same", activation = "relu", input_shape = input_dim))# 卷积层
model.add(MaxPool2D(pool_size=(2, 2)))# 最大池化
model.add(Conv2D(256, (3, 3), padding = "same", activation = "relu")) #卷积层
model.add(MaxPool2D(pool_size=(2, 2))) # 最大池化层
model.add(Dropout(0.3))
model.add(Flatten()) # 展开
model.add(Dense(512, activation = "relu"))
model.add(Dense(10, activation = "softmax"))
# 编译模型,设置损失函数,优化方法以及评价标准
model.compile(optimizer = 'adam', loss = 'categorical_crossentropy', metrics = ['accuracy'])
model.summary()
# 训练模型
model.fit(x_train, y_train, epochs = 16, batch_size = 16, validation_data = (x_test, y_test))
model.save("./model_save/base_cnn_model")

同理,模型保存好,可以利用测试集测试准确率,代码如下:

from sklearn.externals import joblib
from sklearn.preprocessing import StandardScaler
from tensorflow.keras.models import load_model
import numpy as np
from tqdm import tqdm
import os
import glob
from PIL import Image

def img2vec(filename):
    dataVec = np.zeros((1,800))
    img = Image.open(filename)
    img = img.crop((0,0,20,40))
    img = np.array(img)
    for i in range(40):
        for j in range(20):
            dataVec[0,20*i+j] = img[i][j]
    return dataVec

def make_dataset(parent_dir,sub_dirs,file_ext="*.jpg"):
    feature = []
    for sub_dir in sub_dirs:
        for fn in tqdm(glob.glob(os.path.join(parent_dir,sub_dir,file_ext))):
            imgdata = img2vec(fn)
            feature.extend([imgdata])
    return feature

def svm_pred(root_dir,sub_dirs,data="*.jpg"):
    img_total = 0
    acc_total = 0
    clf = joblib.load("./model_save/svm.pkl")
    data = make_dataset(root_dir,sub_dirs)
    data = np.array(data)
    X = np.vstack(data[:, 0])
    scaler = StandardScaler()
    x_std = scaler.fit_transform(X)  # 标准化
    pred = clf.predict(x_std)
    k = 0
    for sub_dir in sub_dirs:
        acc = 0
        sub_path = os.path.join(root_dir,sub_dir)
        num = len(os.listdir(sub_path))
        img_total+=num
        for i in range(num):
            if pred[k] == sub_dir:
                acc+=1
                k+=1
            else:
                k+=1
        print("当前识别数字为:",sub_dir, "    识别正确个数为:",acc, "总数为:",num, "识别准确率为:",float(acc/num))
        acc_total+=acc
    print("训练集总数:",img_total, "正确识别总数:",acc_total, "准确率:",float(acc_total/img_total))

def cnn_pred(root_dir,sub_dirs):
    model = load_model("./model_save/base_cnn_model")
    img_total = 0
    acc_total = 0
    for sub_dir in sub_dirs:
        acc = 0
        sub_path = os.path.join(root_dir,sub_dir)
        num = len(os.listdir(sub_path))
        img_total+=num
        imgfile = [os.path.join(sub_path,x) for x in os.listdir(sub_path)]
        for img in imgfile:
            pred_img = img2vec(img)
            pred_img = pred_img.reshape(-1,40,20,1)
            predlabel = np.argmax(model.predict(pred_img))
            if str(predlabel) == sub_dir:
                acc+=1
        print("当前识别数字为:",sub_dir, "    识别正确个数为:",acc, "总数为:",num, "识别准确率为:",float(acc/num))
        acc_total+=acc
    print("训练集总数:", img_total, "正确识别总数:", acc_total, "准确率:", float(acc_total / img_total))

def main():
    i = input("请输入识别类别1.train   2.test")
    if i == '1':
        root_dir = "./data/train"
        sub_dirs = np.array(['0', '1', '2', '3', '4', '5', '6', '7', '8', '9'])
        print("svm+++++++Train++++++++")
        svm_pred(root_dir,sub_dirs)
        print("cnn+++++++train+++++++")
        cnn_pred(root_dir,sub_dirs)
    if i == '2':
        root_dir = "./data/test"
        sub_dirs = np.array(['0', '1', '2', '3', '4', '5', '6', '7', '8', '9'])
        print("svm+++++++TEST++++++++")
        svm_pred(root_dir,sub_dirs)
        print("cnn+++++++TEST+++++++")
        cnn_pred(root_dir,sub_dirs)

if __name__ == "__main__":
    main()

欢迎入门者参考学习!
源码资源:源码+数据集+预训练模型

欢迎分享,转载请注明来源:内存溢出

原文地址: http://outofmemory.cn/langs/921627.html

(0)
打赏 微信扫一扫 微信扫一扫 支付宝扫一扫 支付宝扫一扫
上一篇 2022-05-16
下一篇 2022-05-16

发表评论

登录后才能评论

评论列表(0条)

保存