其他分享
首页 > 其他分享> > 深度学习识别CIFAR10:pytorch训练LeNet、AlexNet、VGG19实现及比较(一)

深度学习识别CIFAR10:pytorch训练LeNet、AlexNet、VGG19实现及比较(一)

作者:互联网

版权声明:本文为博主原创文章,欢迎转载,并请注明出处。联系方式:460356155@qq.com

前面几篇文章介绍了MINIST,对这种简单图片的识别,LeNet-5可以达到99%的识别率。

CIFAR10是另一个著名的深度学习图像分类识别数据集,比MINIST更复杂,而且是RGB彩色图片。

看看较简单的LeNet-5可以达到多少准确率。网络结构基本和前面MINIST代码中的差不多,主要是输入图片的通道数不同,代码如下:

  1 # -*- coding:utf-8 -*-
  2 
  3 u"""LeNet卷积神经网络训练学习CIFAR10"""
  4 
  5 __author__ = 'zhengbiqing 460356155@qq.com'
  6 
  7 
  8 import torch as t
  9 import torchvision as tv
 10 import torch.nn as nn
 11 import torch.optim as optim
 12 import torchvision.transforms as transforms
 13 from torchvision.transforms import ToPILImage
 14 import torch.backends.cudnn as cudnn
 15 
 16 import datetime
 17 import argparse
 18 
 19 
 20 # 样本读取线程数
 21 WORKERS = 4
 22 
 23 # 网络参赛保存文件名
 24 PARAS_FN = 'cifar_lenet_params.pkl'
 25 
 26 # minist数据存放位置
 27 ROOT = '/home/zbq/PycharmProjects/cifar'
 28 
 29 # 目标函数
 30 loss_func = nn.CrossEntropyLoss()
 31 
 32 # 最优结果
 33 best_acc = 0
 34 
 35 
 36 # 定义网络模型
 37 class LeNet(nn.Module):
 38     def __init__(self):
 39         super(LeNet, self).__init__()
 40 
 41         # 卷积层
 42         self.cnn = nn.Sequential(
 43             # 卷积层1,3通道输入,6个卷积核,核大小5*5
 44             # 经过该层图像大小变为32-5+1,28*28
 45             # 经2*2最大池化,图像变为14*14
 46             nn.Conv2d(3, 6, 5),
 47             nn.ReLU(),
 48             nn.MaxPool2d(2),
 49 
 50             # 卷积层2,6输入通道,16个卷积核,核大小5*5
 51             # 经过该层图像变为14-5+1,10*10
 52             # 经2*2最大池化,图像变为5*5
 53             nn.Conv2d(6, 16, 5),
 54             nn.ReLU(),
 55             nn.MaxPool2d(2)
 56         )
 57 
 58         # 全连接层
 59         self.fc = nn.Sequential(
 60             # 16个feature,每个feature 5*5
 61             nn.Linear(16 * 5 * 5, 120),
 62             nn.ReLU(),
 63             nn.Linear(120, 84),
 64             nn.ReLU(),
 65             nn.Linear(84, 10)
 66         )
 67 
 68     def forward(self, x):
 69         x = self.cnn(x)
 70 
 71         # x.size()[0]: batch size
 72         x = x.view(x.size()[0], -1)
 73         x = self.fc(x)
 74 
 75         return x
 76 
 77 
 78 '''
 79 训练并测试网络
 80 net:网络模型
 81 train_data_load:训练数据集
 82 optimizer:优化器
 83 epoch:第几次训练迭代
 84 log_interval:训练过程中损失函数值和准确率的打印频率
 85 '''
 86 def net_train(net, train_data_load, optimizer, epoch, log_interval):
 87     net.train()
 88 
 89     begin = datetime.datetime.now()
 90 
 91     # 样本总数
 92     total = len(train_data_load.dataset)
 93 
 94     # 样本批次训练的损失函数值的和
 95     train_loss = 0
 96 
 97     # 识别正确的样本数
 98     ok = 0
 99 
100     for i, data in enumerate(train_data_load, 0):
101         img, label = data
102         img, label = img.cuda(), label.cuda()
103 
104         optimizer.zero_grad()
105 
106         outs = net(img)
107         loss = loss_func(outs, label)
108         loss.backward()
109         optimizer.step()
110 
111         # 累加损失值和训练样本数
112         train_loss += loss.item()
113         # total += label.size(0)
114 
115         _, predicted = t.max(outs.data, 1)
116         # 累加识别正确的样本数
117         ok += (predicted == label).sum()
118 
119         if (i + 1) % log_interval == 0:
120             # 训练结果输出
121 
122             # 损失函数均值
123             loss_mean = train_loss / (i + 1)
124 
125             # 已训练的样本数
126             traind_total = (i + 1) * len(label)
127 
128             # 准确度
129             acc = 100. * ok / traind_total
130 
131             # 一个迭代的进度百分比
132             progress = 100. * traind_total / total
133 
134             print('Train Epoch: {} [{}/{} ({:.0f}%)]\tLoss: {:.6f}  Acc: {:.6f}'.format(
135                 epoch, traind_total, total, progress, loss_mean, acc))
136 
137     end = datetime.datetime.now()
138     print('one epoch spend: ', end - begin)
139 
140 
141 '''
142 用测试集检查准确率
143 '''
144 def net_test(net, test_data_load, epoch):
145     net.eval()
146 
147     ok = 0
148 
149     for i, data in enumerate(test_data_load):
150         img, label = data
151         img, label = img.cuda(), label.cuda()
152 
153         outs = net(img)
154         _, pre = t.max(outs.data, 1)
155         ok += (pre == label).sum()
156 
157     acc = ok.item() * 100. / (len(test_data_load.dataset))
158     print('EPOCH:{}, ACC:{}\n'.format(epoch, acc))
159 
160     global best_acc
161     if acc > best_acc:
162         best_acc = acc
163 
164 
165 '''
166 显示数据集中一个图片
167 '''
168 def img_show(dataset, index):
169     classes = ('plane', 'car', 'bird', 'cat',
170                'deer', 'dog', 'frog', 'horse', 'ship', 'truck')
171 
172     show = ToPILImage()
173 
174     data, label = dataset[index]
175     print('img is a ', classes[label])
176     show((data + 1) / 2).resize((100, 100)).show()
177 
178 
179 def main():
180     # 训练超参数设置,可通过命令行设置
181     parser = argparse.ArgumentParser(description='PyTorch CIFA10 LeNet Example')
182     parser.add_argument('--batch-size', type=int, default=64, metavar='N',
183                         help='input batch size for training (default: 64)')
184     parser.add_argument('--test-batch-size', type=int, default=1000, metavar='N',
185                         help='input batch size for testing (default: 1000)')
186     parser.add_argument('--epochs', type=int, default=20, metavar='N',
187                         help='number of epochs to train (default: 20)')
188     parser.add_argument('--lr', type=float, default=0.01, metavar='LR',
189                         help='learning rate (default: 0.01)')
190     parser.add_argument('--momentum', type=float, default=0.9, metavar='M',
191                         help='SGD momentum (default: 0.9)')
192     parser.add_argument('--log-interval', type=int, default=100, metavar='N',
193                         help='how many batches to wait before logging training status (default: 100)')
194     parser.add_argument('--no-train', action='store_true', default=False,
195                         help='If train the Model')
196     parser.add_argument('--save-model', action='store_true', default=False,
197                         help='For Saving the current Model')
198     args = parser.parse_args()
199 
200     # 图像数值转换,ToTensor源码注释
201     """Convert a ``PIL Image`` or ``numpy.ndarray`` to tensor.
202         Converts a PIL Image or numpy.ndarray (H x W x C) in the range
203         [0, 255] to a torch.FloatTensor of shape (C x H x W) in the range [0.0, 1.0].
204         """
205     # 归一化把[0.0, 1.0]变换为[-1,1], ([0, 1] - 0.5) / 0.5 = [-1, 1]
206     transform = tv.transforms.Compose([
207         transforms.ToTensor(),
208         transforms.Normalize([0.5, 0.5, 0.5], [0.5, 0.5, 0.5])])
209 
210     # 定义数据集
211     train_data = tv.datasets.CIFAR10(root=ROOT, train=True, download=True, transform=transform)
212     test_data = tv.datasets.CIFAR10(root=ROOT, train=False, download=False, transform=transform)
213 
214     train_load = t.utils.data.DataLoader(train_data, batch_size=args.batch_size, shuffle=True, num_workers=WORKERS)
215     test_load = t.utils.data.DataLoader(test_data, batch_size=args.test_batch_size, shuffle=False, num_workers=WORKERS)
216 
217     net = LeNet().cuda()
218     print(net)
219 
220     # 如果不训练,直接加载保存的网络参数进行测试集验证
221     if args.no_train:
222         net.load_state_dict(t.load(PARAS_FN))
223         net_test(net, test_load, 0)
224         return
225 
226     optimizer = optim.SGD(net.parameters(), lr=args.lr, momentum=args.momentum)
227 
228     start_time = datetime.datetime.now()
229 
230     for epoch in range(1, args.epochs + 1):
231         net_train(net, train_load, optimizer, epoch, args.log_interval)
232 
233         # 每个epoch结束后用测试集检查识别准确度
234         net_test(net, test_load, epoch)
235 
236     end_time = datetime.datetime.now()
237 
238     global best_acc
239     print('CIFAR10 pytorch LeNet Train: EPOCH:{}, BATCH_SZ:{}, LR:{}, ACC:{}'.format(args.epochs, args.batch_size, args.lr, best_acc))
240     print('train spend time: ', end_time - start_time)
241 
242     if args.save_model:
243         t.save(net.state_dict(), PARAS_FN)
244 
245 
246 if __name__ == '__main__':
247     main()

 

运行结果如下:

Files already downloaded and verified
LeNet(
  (cnn): Sequential(
    (0): Conv2d(3, 6, kernel_size=(5, 5), stride=(1, 1))
    (1): ReLU()
    (2): MaxPool2d(kernel_size=2, stride=2, padding=0, dilation=1, ceil_mode=False)
    (3): Conv2d(6, 16, kernel_size=(5, 5), stride=(1, 1))
    (4): ReLU()
    (5): MaxPool2d(kernel_size=2, stride=2, padding=0, dilation=1, ceil_mode=False)
  )
  (fc): Sequential(
    (0): Linear(in_features=400, out_features=120, bias=True)
    (1): ReLU()
    (2): Linear(in_features=120, out_features=84, bias=True)
    (3): ReLU()
    (4): Linear(in_features=84, out_features=10, bias=True)
  )
)
Train Epoch: 1 [6400/50000 (13%)]    Loss: 2.297558  Acc: 10.000000
Train Epoch: 1 [12800/50000 (26%)]    Loss: 2.219855  Acc: 16.000000
Train Epoch: 1 [19200/50000 (38%)]    Loss: 2.117518  Acc: 20.000000
Train Epoch: 1 [25600/50000 (51%)]    Loss: 2.030452  Acc: 23.000000
Train Epoch: 1 [32000/50000 (64%)]    Loss: 1.956154  Acc: 26.000000
Train Epoch: 1 [38400/50000 (77%)]    Loss: 1.894052  Acc: 29.000000
Train Epoch: 1 [44800/50000 (90%)]    Loss: 1.845520  Acc: 31.000000
one epoch spend:  0:00:02.007186
EPOCH:1, ACC:43.86

Train Epoch: 2 [6400/50000 (13%)]    Loss: 1.497962  Acc: 44.000000
Train Epoch: 2 [12800/50000 (26%)]    Loss: 1.471271  Acc: 45.000000
Train Epoch: 2 [19200/50000 (38%)]    Loss: 1.458853  Acc: 46.000000
Train Epoch: 2 [25600/50000 (51%)]    Loss: 1.445787  Acc: 47.000000
Train Epoch: 2 [32000/50000 (64%)]    Loss: 1.436431  Acc: 47.000000
Train Epoch: 2 [38400/50000 (77%)]    Loss: 1.425798  Acc: 47.000000
Train Epoch: 2 [44800/50000 (90%)]    Loss: 1.415501  Acc: 48.000000
one epoch spend:  0:00:01.879316
EPOCH:2, ACC:53.16

Train Epoch: 3 [6400/50000 (13%)]    Loss: 1.288907  Acc: 52.000000
Train Epoch: 3 [12800/50000 (26%)]    Loss: 1.293646  Acc: 53.000000
Train Epoch: 3 [19200/50000 (38%)]    Loss: 1.284784  Acc: 53.000000
Train Epoch: 3 [25600/50000 (51%)]    Loss: 1.281050  Acc: 53.000000
Train Epoch: 3 [32000/50000 (64%)]    Loss: 1.281222  Acc: 53.000000
Train Epoch: 3 [38400/50000 (77%)]    Loss: 1.269620  Acc: 54.000000
Train Epoch: 3 [44800/50000 (90%)]    Loss: 1.262982  Acc: 54.000000
one epoch spend:  0:00:01.928787
EPOCH:3, ACC:54.31

Train Epoch: 4 [6400/50000 (13%)]    Loss: 1.157912  Acc: 58.000000
Train Epoch: 4 [12800/50000 (26%)]    Loss: 1.157038  Acc: 58.000000
Train Epoch: 4 [19200/50000 (38%)]    Loss: 1.164880  Acc: 58.000000
Train Epoch: 4 [25600/50000 (51%)]    Loss: 1.169460  Acc: 58.000000
Train Epoch: 4 [32000/50000 (64%)]    Loss: 1.169655  Acc: 58.000000
Train Epoch: 4 [38400/50000 (77%)]    Loss: 1.169239  Acc: 58.000000
Train Epoch: 4 [44800/50000 (90%)]    Loss: 1.159252  Acc: 58.000000
one epoch spend:  0:00:01.928551
EPOCH:4, ACC:60.15

Train Epoch: 5 [6400/50000 (13%)]    Loss: 1.081296  Acc: 61.000000
Train Epoch: 5 [12800/50000 (26%)]    Loss: 1.073868  Acc: 61.000000
Train Epoch: 5 [19200/50000 (38%)]    Loss: 1.086076  Acc: 61.000000
Train Epoch: 5 [25600/50000 (51%)]    Loss: 1.088019  Acc: 61.000000
Train Epoch: 5 [32000/50000 (64%)]    Loss: 1.083983  Acc: 61.000000
Train Epoch: 5 [38400/50000 (77%)]    Loss: 1.088050  Acc: 61.000000
Train Epoch: 5 [44800/50000 (90%)]    Loss: 1.087298  Acc: 61.000000
one epoch spend:  0:00:01.898825
EPOCH:5, ACC:59.84

Train Epoch: 6 [6400/50000 (13%)]    Loss: 0.979352  Acc: 65.000000
Train Epoch: 6 [12800/50000 (26%)]    Loss: 1.005338  Acc: 64.000000
Train Epoch: 6 [19200/50000 (38%)]    Loss: 1.019300  Acc: 63.000000
Train Epoch: 6 [25600/50000 (51%)]    Loss: 1.022704  Acc: 63.000000
Train Epoch: 6 [32000/50000 (64%)]    Loss: 1.021217  Acc: 63.000000
Train Epoch: 6 [38400/50000 (77%)]    Loss: 1.022035  Acc: 63.000000
Train Epoch: 6 [44800/50000 (90%)]    Loss: 1.024987  Acc: 63.000000
one epoch spend:  0:00:01.926922
EPOCH:6, ACC:60.04

Train Epoch: 7 [6400/50000 (13%)]    Loss: 0.952975  Acc: 66.000000
Train Epoch: 7 [12800/50000 (26%)]    Loss: 0.965437  Acc: 65.000000
Train Epoch: 7 [19200/50000 (38%)]    Loss: 0.964711  Acc: 65.000000
Train Epoch: 7 [25600/50000 (51%)]    Loss: 0.962520  Acc: 65.000000
Train Epoch: 7 [32000/50000 (64%)]    Loss: 0.964768  Acc: 65.000000
Train Epoch: 7 [38400/50000 (77%)]    Loss: 0.966530  Acc: 65.000000
Train Epoch: 7 [44800/50000 (90%)]    Loss: 0.971995  Acc: 65.000000
one epoch spend:  0:00:01.858537
EPOCH:7, ACC:62.63

Train Epoch: 8 [6400/50000 (13%)]    Loss: 0.901441  Acc: 67.000000
Train Epoch: 8 [12800/50000 (26%)]    Loss: 0.896776  Acc: 68.000000
Train Epoch: 8 [19200/50000 (38%)]    Loss: 0.898365  Acc: 68.000000
Train Epoch: 8 [25600/50000 (51%)]    Loss: 0.898383  Acc: 68.000000
Train Epoch: 8 [32000/50000 (64%)]    Loss: 0.909455  Acc: 67.000000
Train Epoch: 8 [38400/50000 (77%)]    Loss: 0.910068  Acc: 67.000000
Train Epoch: 8 [44800/50000 (90%)]    Loss: 0.914733  Acc: 67.000000
one epoch spend:  0:00:01.849259
EPOCH:8, ACC:62.99

Train Epoch: 9 [6400/50000 (13%)]    Loss: 0.842184  Acc: 69.000000
Train Epoch: 9 [12800/50000 (26%)]    Loss: 0.853178  Acc: 69.000000
Train Epoch: 9 [19200/50000 (38%)]    Loss: 0.863828  Acc: 69.000000
Train Epoch: 9 [25600/50000 (51%)]    Loss: 0.868452  Acc: 69.000000
Train Epoch: 9 [32000/50000 (64%)]    Loss: 0.870991  Acc: 69.000000
Train Epoch: 9 [38400/50000 (77%)]    Loss: 0.874963  Acc: 69.000000
Train Epoch: 9 [44800/50000 (90%)]    Loss: 0.878533  Acc: 68.000000
one epoch spend:  0:00:01.954615
EPOCH:9, ACC:62.5

Train Epoch: 10 [6400/50000 (13%)]    Loss: 0.837819  Acc: 70.000000
Train Epoch: 10 [12800/50000 (26%)]    Loss: 0.823905  Acc: 70.000000
Train Epoch: 10 [19200/50000 (38%)]    Loss: 0.833733  Acc: 70.000000
Train Epoch: 10 [25600/50000 (51%)]    Loss: 0.838861  Acc: 70.000000
Train Epoch: 10 [32000/50000 (64%)]    Loss: 0.841117  Acc: 70.000000
Train Epoch: 10 [38400/50000 (77%)]    Loss: 0.849762  Acc: 69.000000
Train Epoch: 10 [44800/50000 (90%)]    Loss: 0.850071  Acc: 69.000000
one epoch spend:  0:00:01.812348
EPOCH:10, ACC:63.48

Train Epoch: 11 [6400/50000 (13%)]    Loss: 0.781857  Acc: 72.000000
Train Epoch: 11 [12800/50000 (26%)]    Loss: 0.773329  Acc: 72.000000
Train Epoch: 11 [19200/50000 (38%)]    Loss: 0.785191  Acc: 72.000000
Train Epoch: 11 [25600/50000 (51%)]    Loss: 0.797921  Acc: 71.000000
Train Epoch: 11 [32000/50000 (64%)]    Loss: 0.802146  Acc: 71.000000
Train Epoch: 11 [38400/50000 (77%)]    Loss: 0.804404  Acc: 71.000000
Train Epoch: 11 [44800/50000 (90%)]    Loss: 0.805919  Acc: 71.000000
one epoch spend:  0:00:01.881838
EPOCH:11, ACC:63.72

Train Epoch: 12 [6400/50000 (13%)]    Loss: 0.734165  Acc: 74.000000
Train Epoch: 12 [12800/50000 (26%)]    Loss: 0.739923  Acc: 74.000000
Train Epoch: 12 [19200/50000 (38%)]    Loss: 0.753080  Acc: 73.000000
Train Epoch: 12 [25600/50000 (51%)]    Loss: 0.755026  Acc: 73.000000
Train Epoch: 12 [32000/50000 (64%)]    Loss: 0.758760  Acc: 73.000000
Train Epoch: 12 [38400/50000 (77%)]    Loss: 0.765208  Acc: 72.000000
Train Epoch: 12 [44800/50000 (90%)]    Loss: 0.774539  Acc: 72.000000
one epoch spend:  0:00:01.856290
EPOCH:12, ACC:63.71

Train Epoch: 13 [6400/50000 (13%)]    Loss: 0.709528  Acc: 75.000000
Train Epoch: 13 [12800/50000 (26%)]    Loss: 0.713831  Acc: 74.000000
Train Epoch: 13 [19200/50000 (38%)]    Loss: 0.720146  Acc: 74.000000
Train Epoch: 13 [25600/50000 (51%)]    Loss: 0.723680  Acc: 74.000000
Train Epoch: 13 [32000/50000 (64%)]    Loss: 0.730473  Acc: 73.000000
Train Epoch: 13 [38400/50000 (77%)]    Loss: 0.742575  Acc: 73.000000
Train Epoch: 13 [44800/50000 (90%)]    Loss: 0.744857  Acc: 73.000000
one epoch spend:  0:00:01.808256
EPOCH:13, ACC:61.71

Train Epoch: 14 [6400/50000 (13%)]    Loss: 0.700821  Acc: 74.000000
Train Epoch: 14 [12800/50000 (26%)]    Loss: 0.691082  Acc: 75.000000
Train Epoch: 14 [19200/50000 (38%)]    Loss: 0.693119  Acc: 75.000000
Train Epoch: 14 [25600/50000 (51%)]    Loss: 0.706147  Acc: 74.000000
Train Epoch: 14 [32000/50000 (64%)]    Loss: 0.710033  Acc: 74.000000
Train Epoch: 14 [38400/50000 (77%)]    Loss: 0.717097  Acc: 74.000000
Train Epoch: 14 [44800/50000 (90%)]    Loss: 0.724987  Acc: 74.000000
one epoch spend:  0:00:01.797417
EPOCH:14, ACC:63.15

Train Epoch: 15 [6400/50000 (13%)]    Loss: 0.624073  Acc: 77.000000
Train Epoch: 15 [12800/50000 (26%)]    Loss: 0.637354  Acc: 77.000000
Train Epoch: 15 [19200/50000 (38%)]    Loss: 0.646385  Acc: 76.000000
Train Epoch: 15 [25600/50000 (51%)]    Loss: 0.662080  Acc: 76.000000
Train Epoch: 15 [32000/50000 (64%)]    Loss: 0.668658  Acc: 76.000000
Train Epoch: 15 [38400/50000 (77%)]    Loss: 0.679682  Acc: 75.000000
Train Epoch: 15 [44800/50000 (90%)]    Loss: 0.688876  Acc: 75.000000
one epoch spend:  0:00:01.916400
EPOCH:15, ACC:62.81

Train Epoch: 16 [6400/50000 (13%)]    Loss: 0.611007  Acc: 78.000000
Train Epoch: 16 [12800/50000 (26%)]    Loss: 0.612629  Acc: 78.000000
Train Epoch: 16 [19200/50000 (38%)]    Loss: 0.622980  Acc: 77.000000
Train Epoch: 16 [25600/50000 (51%)]    Loss: 0.638267  Acc: 77.000000
Train Epoch: 16 [32000/50000 (64%)]    Loss: 0.650756  Acc: 76.000000
Train Epoch: 16 [38400/50000 (77%)]    Loss: 0.656675  Acc: 76.000000
Train Epoch: 16 [44800/50000 (90%)]    Loss: 0.665181  Acc: 75.000000
one epoch spend:  0:00:01.878367
EPOCH:16, ACC:61.64

Train Epoch: 17 [6400/50000 (13%)]    Loss: 0.591583  Acc: 78.000000
Train Epoch: 17 [12800/50000 (26%)]    Loss: 0.601943  Acc: 78.000000
Train Epoch: 17 [19200/50000 (38%)]    Loss: 0.612084  Acc: 78.000000
Train Epoch: 17 [25600/50000 (51%)]    Loss: 0.619225  Acc: 77.000000
Train Epoch: 17 [32000/50000 (64%)]    Loss: 0.633562  Acc: 77.000000
Train Epoch: 17 [38400/50000 (77%)]    Loss: 0.641217  Acc: 77.000000
Train Epoch: 17 [44800/50000 (90%)]    Loss: 0.648393  Acc: 76.000000
one epoch spend:  0:00:01.894760
EPOCH:17, ACC:61.44

Train Epoch: 18 [6400/50000 (13%)]    Loss: 0.553651  Acc: 80.000000
Train Epoch: 18 [12800/50000 (26%)]    Loss: 0.569668  Acc: 79.000000
Train Epoch: 18 [19200/50000 (38%)]    Loss: 0.584057  Acc: 78.000000
Train Epoch: 18 [25600/50000 (51%)]    Loss: 0.598776  Acc: 78.000000
Train Epoch: 18 [32000/50000 (64%)]    Loss: 0.610767  Acc: 78.000000
Train Epoch: 18 [38400/50000 (77%)]    Loss: 0.617563  Acc: 77.000000
Train Epoch: 18 [44800/50000 (90%)]    Loss: 0.628669  Acc: 77.000000
one epoch spend:  0:00:01.925175
EPOCH:18, ACC:62.46

Train Epoch: 19 [6400/50000 (13%)]    Loss: 0.554530  Acc: 79.000000
Train Epoch: 19 [12800/50000 (26%)]    Loss: 0.574952  Acc: 78.000000
Train Epoch: 19 [19200/50000 (38%)]    Loss: 0.576819  Acc: 79.000000
Train Epoch: 19 [25600/50000 (51%)]    Loss: 0.584052  Acc: 78.000000
Train Epoch: 19 [32000/50000 (64%)]    Loss: 0.590673  Acc: 78.000000
Train Epoch: 19 [38400/50000 (77%)]    Loss: 0.599807  Acc: 78.000000
Train Epoch: 19 [44800/50000 (90%)]    Loss: 0.607849  Acc: 78.000000
one epoch spend:  0:00:01.827582
EPOCH:19, ACC:62.16

Train Epoch: 20 [6400/50000 (13%)]    Loss: 0.534505  Acc: 80.000000
Train Epoch: 20 [12800/50000 (26%)]    Loss: 0.547133  Acc: 80.000000
Train Epoch: 20 [19200/50000 (38%)]    Loss: 0.557482  Acc: 79.000000
Train Epoch: 20 [25600/50000 (51%)]    Loss: 0.567949  Acc: 79.000000
Train Epoch: 20 [32000/50000 (64%)]    Loss: 0.579047  Acc: 79.000000
Train Epoch: 20 [38400/50000 (77%)]    Loss: 0.591825  Acc: 78.000000
Train Epoch: 20 [44800/50000 (90%)]    Loss: 0.598099  Acc: 78.000000
one epoch spend:  0:00:01.846124
EPOCH:20, ACC:62.47

CIFAR10 pytorch LeNet Train: EPOCH:20, BATCH_SZ:64, LR:0.01, ACC:63.72
train spend time:  0:00:46.669295

Process finished with exit code 0

训练的lenet准确度在63%左右,远低于MINIST的99%,简单的LeNet对较复杂的图片准确度不高。

 

标签:Acc,50000,Loss,CIFAR10,VGG19,epoch,Epoch,Train,LeNet
来源: https://www.cnblogs.com/zhengbiqing/p/10424693.html