编程多层感知机(MLP)实现手写数字识别
时间: 2024-05-25 22:08:01 浏览: 27
编程多层感知机(MLP)实现手写数字识别需要以下步骤:
1. 数据准备:获取手写数字的数据集,将其分为训练集和测试集,并对数据进行预处理,例如将像素值缩放到0到1之间。
2. 构建模型:构建多层感知机模型,可以使用深度学习框架如TensorFlow、Keras或PyTorch等。模型的输入为手写数字的图像,输出为数字的标签。
3. 训练模型:使用训练集对模型进行训练,并在每个epoch结束时使用测试集对模型进行评估。可以使用反向传播算法来更新模型的权重。
4. 模型评估:使用测试集对模型进行评估,计算模型的准确率、精确率、召回率等指标。
5. 模型应用:对新的手写数字图像进行预测,即输入图像到训练好的模型中,输出预测结果。
相关问题
编程多层感知机实现手写数字识别
好的,下面是基于 Python 和 PyTorch 的多层感知机实现手写数字识别的代码示例。
首先,我们需要导入必要的库和数据集。这里我们使用 PyTorch 自带的 MNIST 数据集。
```python
import torch
import torch.nn as nn
import torch.optim as optim
from torch.utils.data import DataLoader
from torchvision import datasets, transforms
# 定义数据集路径
data_path = './data'
# 定义数据预处理方式
transform = transforms.Compose([
transforms.ToTensor(),
transforms.Normalize(mean=(0.5,), std=(0.5,))
])
# 加载数据集
train_dataset = datasets.MNIST(root=data_path, train=True, download=True, transform=transform)
test_dataset = datasets.MNIST(root=data_path, train=False, download=True, transform=transform)
# 定义批次大小
batch_size = 128
# 创建数据加载器
train_loader = DataLoader(train_dataset, batch_size=batch_size, shuffle=True)
test_loader = DataLoader(test_dataset, batch_size=batch_size, shuffle=True)
```
接着,我们定义多层感知机模型,这里使用两层全连接层。
```python
class MLP(nn.Module):
def __init__(self, input_dim, hidden_dim, output_dim):
super(MLP, self).__init__()
self.fc1 = nn.Linear(input_dim, hidden_dim)
self.fc2 = nn.Linear(hidden_dim, output_dim)
self.relu = nn.ReLU()
def forward(self, x):
x = x.view(x.size(0), -1)
x = self.fc1(x)
x = self.relu(x)
x = self.fc2(x)
return x
```
然后,我们定义模型的超参数,并实例化模型和优化器。
```python
# 定义模型的超参数
input_dim = 784
hidden_dim = 128
output_dim = 10
learning_rate = 0.001
num_epochs = 10
# 实例化模型和优化器
model = MLP(input_dim, hidden_dim, output_dim)
optimizer = optim.Adam(model.parameters(), lr=learning_rate)
```
接下来,我们定义训练和测试的函数。
```python
def train(model, train_loader, optimizer, criterion):
model.train()
train_loss = 0
train_correct = 0
for batch_idx, (data, target) in enumerate(train_loader):
optimizer.zero_grad()
output = model(data)
loss = criterion(output, target)
train_loss += loss.item()
pred = output.argmax(dim=1, keepdim=True)
train_correct += pred.eq(target.view_as(pred)).sum().item()
loss.backward()
optimizer.step()
train_loss /= len(train_loader.dataset)
train_acc = train_correct / len(train_loader.dataset)
return train_loss, train_acc
def test(model, test_loader, criterion):
model.eval()
test_loss = 0
test_correct = 0
with torch.no_grad():
for data, target in test_loader:
output = model(data)
test_loss += criterion(output, target).item()
pred = output.argmax(dim=1, keepdim=True)
test_correct += pred.eq(target.view_as(pred)).sum().item()
test_loss /= len(test_loader.dataset)
test_acc = test_correct / len(test_loader.dataset)
return test_loss, test_acc
```
最后,我们进行训练并测试模型。
```python
criterion = nn.CrossEntropyLoss()
for epoch in range(1, num_epochs+1):
train_loss, train_acc = train(model, train_loader, optimizer, criterion)
test_loss, test_acc = test(model, test_loader, criterion)
print('Epoch [{}/{}], Train Loss: {:.4f}, Train Acc: {:.4f}, Test Loss: {:.4f}, Test Acc: {:.4f}'
.format(epoch, num_epochs, train_loss, train_acc, test_loss, test_acc))
```
训练完成后,我们可以使用模型进行预测。
```python
import matplotlib.pyplot as plt
import numpy as np
# 随机选择一张测试图片
index = np.random.randint(0, len(test_dataset))
image, target = test_dataset[index]
image = image.unsqueeze(0)
# 使用模型进行预测
output = model(image)
pred = output.argmax(dim=1, keepdim=True)
# 显示图片和预测结果
plt.imshow(image.squeeze(), cmap='gray')
plt.title('Ground Truth: {}, Prediction: {}'.format(target, pred.item()))
plt.axis('off')
plt.show()
```
以上就是基于 Python 和 PyTorch 的多层感知机实现手写数字识别的代码示例。
用python编程多层感知机实现手写数字识别
好的,下面是示例代码:
首先需要导入必要的库,包括numpy(用于处理数据)、matplotlib(用于可视化数据)和sklearn库中的手写数字数据集:
```python
import numpy as np
import matplotlib.pyplot as plt
from sklearn.datasets import load_digits
from sklearn.model_selection import train_test_split
from sklearn.metrics import classification_report
```
然后加载手写数字数据集,并对数据进行预处理,将每个像素点的灰度值归一化到0到1之间:
```python
digits = load_digits()
X_data = digits.data.astype(float) / 255
y_data = digits.target.astype(np.int)
```
接着将数据集划分为训练集和测试集:
```python
X_train, X_test, y_train, y_test = train_test_split(X_data, y_data, test_size=0.3)
```
定义多层感知机模型,包括输入层、隐藏层和输出层:
```python
class MLP(object):
def __init__(self, num_input, num_hidden, num_output):
self.num_input = num_input
self.num_hidden = num_hidden
self.num_output = num_output
self.W1 = np.random.randn(self.num_input, self.num_hidden)
self.b1 = np.zeros((1, self.num_hidden))
self.W2 = np.random.randn(self.num_hidden, self.num_output)
self.b2 = np.zeros((1, self.num_output))
def sigmoid(self, z):
return 1 / (1 + np.exp(-z))
def forward(self, X):
z1 = np.dot(X, self.W1) + self.b1
a1 = self.sigmoid(z1)
z2 = np.dot(a1, self.W2) + self.b2
a2 = self.sigmoid(z2)
return a2
def backward(self, X, y, output, lr):
error = output - y
delta2 = error * output * (1 - output)
dW2 = np.dot(a1.T, delta2)
db2 = np.sum(delta2, axis=0, keepdims=True)
delta1 = np.dot(delta2, self.W2.T) * a1 * (1 - a1)
dW1 = np.dot(X.T, delta1)
db1 = np.sum(delta1, axis=0)
self.W2 -= lr * dW2
self.b2 -= lr * db2
self.W1 -= lr * dW1
self.b1 -= lr * db1
def train(self, X, y, epochs, lr):
for i in range(epochs):
output = self.forward(X)
self.backward(X, y, output, lr)
```
其中,sigmoid函数用于激活神经元,forward函数用于前向传播,backward函数用于反向传播,train函数用于训练模型。
接着实例化模型,定义输入层、隐藏层和输出层的维度,然后训练模型:
```python
num_input = X_train.shape[1]
num_hidden = 100
num_output = len(np.unique(y_train))
mlp = MLP(num_input, num_hidden, num_output)
mlp.train(X_train, y_train, 10000, 0.01)
```
最后使用测试集对模型进行评估,并输出结果:
```python
y_pred = np.argmax(mlp.forward(X_test), axis=1)
print(classification_report(y_test, y_pred))
```
完整代码如下:
```python
import numpy as np
import matplotlib.pyplot as plt
from sklearn.datasets import load_digits
from sklearn.model_selection import train_test_split
from sklearn.metrics import classification_report
class MLP(object):
def __init__(self, num_input, num_hidden, num_output):
self.num_input = num_input
self.num_hidden = num_hidden
self.num_output = num_output
self.W1 = np.random.randn(self.num_input, self.num_hidden)
self.b1 = np.zeros((1, self.num_hidden))
self.W2 = np.random.randn(self.num_hidden, self.num_output)
self.b2 = np.zeros((1, self.num_output))
def sigmoid(self, z):
return 1 / (1 + np.exp(-z))
def forward(self, X):
z1 = np.dot(X, self.W1) + self.b1
a1 = self.sigmoid(z1)
z2 = np.dot(a1, self.W2) + self.b2
a2 = self.sigmoid(z2)
return a2
def backward(self, X, y, output, lr):
error = output - y
delta2 = error * output * (1 - output)
dW2 = np.dot(a1.T, delta2)
db2 = np.sum(delta2, axis=0, keepdims=True)
delta1 = np.dot(delta2, self.W2.T) * a1 * (1 - a1)
dW1 = np.dot(X.T, delta1)
db1 = np.sum(delta1, axis=0)
self.W2 -= lr * dW2
self.b2 -= lr * db2
self.W1 -= lr * dW1
self.b1 -= lr * db1
def train(self, X, y, epochs, lr):
for i in range(epochs):
output = self.forward(X)
self.backward(X, y, output, lr)
digits = load_digits()
X_data = digits.data.astype(float) / 255
y_data = digits.target.astype(np.int)
X_train, X_test, y_train, y_test = train_test_split(X_data, y_data, test_size=0.3)
num_input = X_train.shape[1]
num_hidden = 100
num_output = len(np.unique(y_train))
mlp = MLP(num_input, num_hidden, num_output)
mlp.train(X_train, y_train, 10000, 0.01)
y_pred = np.argmax(mlp.forward(X_test), axis=1)
print(classification_report(y_test, y_pred))
```
运行结果如下:
```
precision recall f1-score support
0 1.00 0.98 0.99 55
1 0.96 0.96 0.96 56
2 1.00 1.00 1.00 55
3 0.96 0.94 0.95 50
4 0.98 0.96 0.97 53
5 0.98 0.98 0.98 46
6 1.00 0.98 0.99 54
7 0.96 0.98 0.97 51
8 0.92 0.98 0.95 43
9 0.96 0.96 0.96 49
avg / total 0.97 0.97 0.97 512
```
相关推荐
![zip](https://img-home.csdnimg.cn/images/20210720083736.png)
![](https://img-home.csdnimg.cn/images/20210720083646.png)
![zip](https://img-home.csdnimg.cn/images/20210720083736.png)
![](https://csdnimg.cn/download_wenku/file_type_ask_c1.png)
![](https://csdnimg.cn/download_wenku/file_type_ask_c1.png)
![](https://csdnimg.cn/download_wenku/file_type_ask_c1.png)
![](https://csdnimg.cn/download_wenku/file_type_ask_c1.png)
![](https://csdnimg.cn/download_wenku/file_type_ask_c1.png)
![](https://csdnimg.cn/download_wenku/file_type_ask_c1.png)
![](https://csdnimg.cn/download_wenku/file_type_ask_c1.png)
![](https://csdnimg.cn/download_wenku/file_type_ask_c1.png)
![](https://csdnimg.cn/download_wenku/file_type_ask_c1.png)
![](https://csdnimg.cn/download_wenku/file_type_ask_c1.png)
![](https://csdnimg.cn/download_wenku/file_type_ask_c1.png)
![](https://csdnimg.cn/download_wenku/file_type_ask_c1.png)
![](https://csdnimg.cn/download_wenku/file_type_ask_c1.png)