利用CNN卷积神经网络实现人脸识别(python+TensorFlow)
使用的人脸数据是耶鲁大学的一个人脸数据集Yale_64x64.mat,数据集已经上传Yale 64x64.mat
程序:
'''''''''
使用Yale_64x64.mat人脸数据,利用CNN卷积神经网络实现人脸识别
Yale_64x64.mat数据构成:分为fea(人脸数据165x4096) gnd(标签165x1)图像大小为64x64(64x64=4096)一共15个人的人脸,每个人11条人脸数据
'''
import tensorflow as tf
import numpy as np
import scipy.io as siof = open('Yale_64x64.mat','rb')
mdict = sio.loadmat(f)
# fea:数据 gnd:标签
train_data = mdict['fea']
train_label = mdict['gnd']# 将数据分为训练数据与测试数据
train_data = np.random.permutation(train_data)
train_label = np.random.permutation(train_label)
test_data = train_data[0:64]
test_label = train_label[0:64]
np.random.seed(100)
test_data = np.random.permutation(test_data)
np.random.seed(100)
test_label = np.random.permutation(test_label)
train_data = train_data.reshape(train_data.shape[0], 64, 64, 1).astype(np.float32)/255# 将标签数据改为one_hot编码格式的数据
train_labels_new = np.zeros((165, 15))
for i in range(0, 165):j = int(train_label[i, 0])-1train_labels_new[i, j] = 1test_data_input = test_data.reshape(test_data.shape[0], 64, 64, 1).astype(np.float32)/255
test_labels_input = np.zeros((64,15))
for i in range(0,64):j = int(test_label[i, 0])-1test_labels_input[i, j] = 1# CNN
data_input = tf.placeholder(tf.float32,[None, 64, 64, 1])
label_input = tf.placeholder(tf.float32,[None, 15])layer1 = tf.layers.conv2d(inputs=data_input, filters=32, kernel_size=2, strides=1, padding='SAME', activation=tf.nn.relu)
layer1_pool = tf.layers.max_pooling2d(layer1, pool_size=2, strides=2)
layer2 = tf.layers.conv2d(inputs=layer1_pool, filters=64, kernel_size=2, strides=1, padding='SAME', activation=tf.nn.relu)
layer2_pool = tf.layers.max_pooling2d(layer2, pool_size=2, strides=2)
layer3 = tf.reshape(layer2_pool, [-1,16*16*64])
layer3_relu = tf.layers.dense(layer3,1024, tf.nn.relu)
output = tf.layers.dense(layer3_relu, 15)# 计算损失函数 最小化损失函数 计算测试精确度
loss = tf.losses.softmax_cross_entropy(onehot_labels=label_input, logits=output)
train = tf.train.GradientDescentOptimizer(0.01).minimize(loss)
accuracy = tf.metrics.accuracy(labels=tf.argmax(label_input,axis=1), predictions=tf.argmax(output, axis=1))[1]# 初始化 运行计算图
init = tf.group(tf.global_variables_initializer(),tf.local_variables_initializer())
with tf.Session() as sess:sess.run(init)tf.summary.FileWriter('D:/face_log', sess.graph)for i in range(0,1500):train_data_input = np.array(train_data)train_label_input = np.array(train_labels_new)sess.run([train, loss], feed_dict={data_input: train_data_input, label_input: train_label_input})acc = sess.run(accuracy, feed_dict={data_input: test_data_input, label_input: test_labels_input})print('step:%d accuracy:%.2f%%' % (i+1, acc*100))
运行结果: