赞
踩
python 实现单隐层神经网络,处理cifar10数据集
forward_neural_network.py
#!/usr/bin/python # -*- coding: utf-8 -*- ##################################### # File name : forward_neural_network.py # Create date : 2018-12-25 20:04 # Modified date : 2018-12-31 16:02 # Author : DARREN # Describe : not set # Email : lzygzh@126.com ##################################### from __future__ import division from __future__ import print_function import time import numpy as np #import matplotlib.pyplot as plt import cifar10_dataset import func import model_base import graph class DeepModel(model_base.ModelBase): def __init__(self): super(DeepModel, self).__init__() self.hyper = self._get_hyperparameters() self.graph = graph.NeuralGraph() def _get_hyperparameters(self): dic = {} dic["batch_size"] = 128 dic["epsilon"] = 0.0000001 dic["reg_lambda"] = 0.05 dic["learn_rate"] = 0.005 dic["max_steps"] = 50000 dic["train_steps"] = 50 dic["max_epochs"] = 50 dic["input_dim"] = 32*32*3#img size dic["hidden_dim"] = 2048 dic["output_dim"] = 10 dic["file_path"] = "./data/" dic["activation"] = "sigmoid" dic["architecture"] = [32*32*3, 2048, 10] return dic def _train_model_with_epochs(self, model=None): data_generator = self._get_train_generator() while 1: X, Y, status = self._get_a_batch_data(data_generator) if status == False: model['epochs'] += 1 break model = self.graph.core_graph(model, X, Y, self.hyper) model["steps"] += 1 if model["steps"] % self.hyper["train_steps"] == 0: self._print_train_status(model) return model def _train_model_with_steps(self, model=None, data_generator=None): if data_generator == None: data_generator = self._get_train_generator() while 1: X, Y, status = self._get_a_batch_data(data_generator) if status == False: data_generator = self._get_train_generator() model['epochs'] += 1 model = self.graph.core_graph(model, X, Y, self.hyper) model["steps"] += 1 if model["steps"] % self.hyper["train_steps"] == 0: break return model, data_generator def _test_model(self, model): data_generator = self._get_test_generator() count = 1 all_correct_numbers = 0 all_loss = 0.0 while count: X, Y, status = self._get_a_batch_data(data_generator) if status == False: break start_time = time.time() model, prob, a1, Z1, loss, accuracy, comp = self.graph.forward_propagation(model, X, Y, self.hyper) end_time = time.time() model = self._record_test_speed(model, start_time, end_time) all_loss += loss all_correct_numbers += len(np.flatnonzero(comp)) count += 1 avg_loss = all_loss / count accuracy = all_correct_numbers / (count * self.hyper["batch_size"]) self._test_update_model(model, avg_loss, accuracy) self._print_test_status(model) self._record_model_status(model) return model
#!/usr/bin/python # -*- coding: utf-8 -*- ##################################### # File name : graph.py # Create date : 2018-12-30 23:06 # Modified date : 2018-12-31 16:01 # Author : DARREN # Describe : not set # Email : lzygzh@126.com ##################################### from __future__ import division from __future__ import print_function import time import numpy as np import func def _record_speed(model, start_time, end_time, hyper_dic): dual_time = end_time - start_time speed = hyper_dic["batch_size"]/dual_time model["train_speed"] = speed return model class NeuralGraph(object): def __init__(self): super(NeuralGraph, self).__init__() def _init_model(self, hyper_dic, model=None): if not model: #np.random.seed(0) architecture = hyper_dic["architecture"] W1 = np.random.randn(architecture[0], architecture[1]) b1 = np.ones((1, architecture[1])) W2 = np.random.randn(architecture[1], architecture[2]) b2 = np.ones((1, architecture[2])) model = {} model["W1"] = W1 model["b1"] = b1 model["W2"] = W2 model["b2"] = b2 model["steps"] = 1 model["epochs"] = 0 model["record"] = {} model["best_test_accuracy"] = 0.0 model["best_epoch"] = 0 return model def _normalization(self, batch_img): return batch_img / 255. def forward_propagation(self, model, X, Y, hyper_dic): activation_str = hyper_dic["activation"] model = self._init_model(hyper_dic, model) W1 = model["W1"] b1 = model["b1"] W2 = model["W2"] b2 = model["b2"] X = self._normalization(X) Z1 = np.dot(X, W1)+b1 a1 = func.activation(Z1, activation_str) logits = np.dot(a1, W2)+b2 prob = func.softmax(logits) correct_probs = prob[range(X.shape[0]), np.argmax(Y, axis=1)] correct_logprobs = - func.log(correct_probs) data_loss = np.sum(correct_logprobs) loss = 1./X.shape[0] * data_loss pre_Y = np.argmax(prob, axis=1) comp = pre_Y == np.argmax(Y, axis=1) accuracy = len(np.flatnonzero(comp))/Y.shape[0] return model, prob, a1, Z1, loss, accuracy, comp def backward_propagation(self, model, prob, X, Y, a1, Z1, hyper_dic): activation_str = hyper_dic["activation"] learn_rate = hyper_dic["learn_rate"] W2 = model["W2"] dY_pred = prob - Y dW2 = np.dot(a1.T, dY_pred) da1 = np.dot(dY_pred, W2.T) dadZ = func.activation_derivative(Z1, activation_str) dZ1 = da1 * dadZ dW1 = np.dot(X.T, dZ1) model["W2"] += -learn_rate * dW2 model["W1"] += -learn_rate * dW1 return model def core_graph(self, model, X, Y, hyper_dic): start_time = time.time() model, prob, a1, Z1, loss, accuracy, comp = self.forward_propagation(model, X, Y, hyper_dic) model["train_loss"] = loss model["train_accuracy"] = accuracy model = self.backward_propagation(model, prob, X, Y, a1, Z1, hyper_dic) end_time = time.time() model = _record_speed(model, start_time, end_time, hyper_dic) return model
#!/usr/bin/python # -*- coding: utf-8 -*- ##################################### # File name : func.py # Create date : 2018-12-30 21:33 # Modified date : 2018-12-31 14:23 # Author : DARREN # Describe : not set # Email : lzygzh@126.com ##################################### from __future__ import division from __future__ import print_function import numpy as np def sig(x): pos = np.where(x >= 0) p = 1.0 / (1 + np.exp(-x[pos])) neg = np.where(x < 0) n = np.exp(x[neg]) / (1 + np.exp(x[neg])) x[pos] = p x[neg] = n return x def relu(x): return (np.abs(x) + x) / 2.0 def relu_derivative(x): x[x <= 0] = 0 x[x > 0] = 1 return x def sig_deirvative(x): return sig(x) * (1 - sig(x)) def log(x, epsilon=0.0000001): x = x + epsilon return np.log(x) def softmax(logits, epsilon=0.0000001): logits_max = np.max(logits, axis=1) for i in range(len(logits)): logits[i] = logits[i] - logits_max[i] logits = logits + epsilon exp_score = np.exp(logits) prob = exp_score/np.sum(exp_score, axis=1, keepdims=1) return prob def activation(x, activation_str="relu"): activation_dic = {} activation_dic["relu"] = relu activation_dic["sigmoid"] = sig f = activation_dic[activation_str] return f(x) def activation_derivative(x, activation_str="relu"): activation_derivative_dic = {} activation_derivative_dic["relu"] = relu_derivative activation_derivative_dic["sigmoid"] = sig_deirvative f = activation_derivative_dic[activation_str] return f(x)
model_base.py
#!/usr/bin/python # -*- coding: utf-8 -*- ##################################### # File name : model_base.py # Create date : 2018-12-25 20:04 # Modified date : 2018-12-31 16:02 # Author : DARREN # Describe : not set # Email : lzygzh@126.com ##################################### from __future__ import division from __future__ import print_function import numpy as np import matplotlib.pyplot as plt import cifar10_dataset import func class ModelBase(object): def __init__(self): super(ModelBase, self).__init__() self.hyper = self._get_hyperparameters() def _get_dataset(self): dataset = cifar10_dataset.Cifar10Set(self.hyper["file_path"]) return dataset def _get_train_generator_with(self, dataset): data_generator = dataset.get_train_data_generator(self.hyper["batch_size"]) return data_generator def _get_test_generator_with(self, dataset): data_generator = dataset.get_test_data_generator(self.hyper["batch_size"]) return data_generator def _get_train_generator(self): dataset = self._get_dataset() data_generator = self._get_train_generator_with(dataset) return data_generator def _get_test_generator(self): dataset = self._get_dataset() data_generator = self._get_test_generator_with(dataset) return data_generator def _get_a_batch_data(self, data_generator): dataset = self._get_dataset() batch_img, batch_labels, status = dataset.get_a_batch_data(data_generator) return batch_img, batch_labels, status def _record_test_speed(self, model, start_time, end_time): dual_time = end_time - start_time speed = self.hyper["batch_size"]/dual_time model["test_speed"] = speed return model def show_hyperparameters(self): print("pyperparameters:") for key in self.hyper: print("%s:%s" % (key, self.hyper[key])) def _normalization(self, batch_img): return batch_img / 255. def _print_train_status(self, model): print("epoch:%s steps:%s Train_Loss:%2.5f Train_Acc:%2.5f" % (model["epochs"], model["steps"], model["train_loss"], model["train_accuracy"])) def _print_test_status(self, model): print("E:%s S:%s Train_Loss:%2.5f Test_Loss:%2.5f Train_Acc:%2.5f Test_Acc:%2.5f gap:%2.5f Train_Speed:%s Test_Speed:%s" % (model["epochs"], model["steps"], model["train_loss"], model["test_loss"], model["train_accuracy"], model["test_accuracy"], model["train_test_gap"], model["train_speed"], model["test_speed"])) print("best_epoch:%s best_test_acc:%s" % (model["best_epoch"], model["best_test_accuracy"])) def _test_update_model(self, model, avg_loss, accuracy): if accuracy > model["best_test_accuracy"]: model["best_test_accuracy"] = accuracy model["best_epoch"] = model["epochs"] model["test_loss"] = avg_loss model["test_accuracy"] = accuracy model["train_test_gap"] = model["train_accuracy"] - model["test_accuracy"] return model def _record_model_status(self, model): steps_dic = {} steps_dic["epochs"] = model["epochs"] steps_dic["steps"] = model["steps"] steps_dic["train_loss"] = model["train_loss"] steps_dic["train_accuracy"] = model["train_accuracy"] steps_dic["test_loss"] = model["test_loss"] steps_dic["test_accuracy"] = model["test_accuracy"] steps_dic["train_test_gap"] = model["train_test_gap"] record = model["record"] record[model["steps"]] = steps_dic def _plot_record(self, model): self._plot_a_key(model, "train_loss", "test_loss") self._plot_a_key(model, "train_accuracy", "test_accuracy") def _plot_a_key(self, model, train_key, test_key): record = model["record"] train = [] test = [] steps = [] for key in record: steps.append([key]) steps.sort() for i in range(len(steps)): step_dic = record[steps[i][0]] train_value = step_dic[train_key] train.append(train_value) test_value = step_dic[test_key] test.append(test_value) train = np.array(train) steps = np.array(steps) plt.plot(steps, train) plt.plot(steps, test) plt.show() def run_with_epoch(self): model = None while 1: model = self._train_model_with_epochs(model) self._test_model(model) if model["epochs"] > self.hyper["max_epochs"]: break if model["epochs"] > model["best_epoch"] + 3: print("early stop") break self._plot_record(model) def run_with_steps(self): model = None data_generator = None while 1: model, data_generator = self._train_model_with_steps(model, data_generator) model = self._test_model(model) if model["steps"] > self.hyper["max_steps"]: break if model["epochs"] > model["best_epoch"] + 3: print("early stop") break self._plot_record(model)
#!/usr/bin/python # -*- coding: utf-8 -*- ##################################### # File name : main.py # Create date : 2018-12-23 16:53 # Modified date : 2018-12-31 15:37 # Author : DARREN # Describe : not set # Email : lzygzh@126.com ##################################### from __future__ import division from __future__ import print_function import forward_neural_network def test_deep_model_with_epochs(): neural_model = forward_neural_network.DeepModel() neural_model.show_hyperparameters() neural_model.run_with_epoch() def test_deep_model_with_steps(): neural_model = forward_neural_network.DeepModel() neural_model.show_hyperparameters() neural_model.run_with_steps() def run(): test_deep_model_with_epochs() #test_deep_model_with_steps() run()
输出
loss:
accuracy:
shell 输出
hyperparameters: learn_rate:0.005 epsilon:1e-07 output_dim:10 batch_size:128 file_path:./data/ reg_lambda:0.05 activation:sigmoid hidden_dim:2048 input_dim:3072 max_epochs:50 max_steps:50000 train_steps:50 architecture:[3072, 2048, 10] E:0 S:50 Train_Loss:2.40109 Test_Loss:2.36403 Train_Acc:0.08594 Test_Acc:0.11798 gap:-0.03204 Train_Speed:1196.95652911 Test_Speed:4396.79711724 best_epoch:0 best_test_acc:0.117978639241 E:0 S:100 Train_Loss:2.73293 Test_Loss:2.86622 Train_Acc:0.12500 Test_Acc:0.14735 gap:-0.02235 Train_Speed:1186.44981017 Test_Speed:3955.98670705 best_epoch:0 best_test_acc:0.147349683544 E:0 S:150 Train_Loss:3.19594 Test_Loss:3.08659 Train_Acc:0.21875 Test_Acc:0.19027 gap:0.02848 Train_Speed:1139.08183629 Test_Speed:4421.85690165 best_epoch:0 best_test_acc:0.190268987342 E:0 S:200 Train_Loss:3.84240 Test_Loss:3.45429 Train_Acc:0.17969 Test_Acc:0.20144 gap:-0.02176 Train_Speed:1020.29470611 Test_Speed:2919.56969226 best_epoch:0 best_test_acc:0.201443829114 E:0 S:250 Train_Loss:2.59963 Test_Loss:3.05986 Train_Acc:0.21094 Test_Acc:0.19759 gap:0.01335 Train_Speed:1047.98650759 Test_Speed:3492.77473668 best_epoch:0 best_test_acc:0.201443829114 E:0 S:300 Train_Loss:3.47834 Test_Loss:3.65382 Train_Acc:0.14844 Test_Acc:0.20688 gap:-0.05845 Train_Speed:957.839422518 Test_Speed:3265.30049812 best_epoch:0 best_test_acc:0.206882911392 E:0 S:350 Train_Loss:2.79823 Test_Loss:3.33488 Train_Acc:0.24219 Test_Acc:0.22063 gap:0.02156 Train_Speed:1107.5829123 Test_Speed:4329.32481775 best_epoch:0 best_test_acc:0.220628955696 E:1 S:400 Train_Loss:3.29347 Test_Loss:3.32084 Train_Acc:0.24219 Test_Acc:0.23022 gap:0.01197 Train_Speed:1022.81387551 Test_Speed:3619.8262605 best_epoch:1 best_test_acc:0.230221518987 E:1 S:450 Train_Loss:3.66320 Test_Loss:3.18382 Train_Acc:0.17188 Test_Acc:0.22083 gap:-0.04895 Train_Speed:1033.77604221 Test_Speed:3801.50193307 best_epoch:1 best_test_acc:0.230221518987 E:1 S:500 Train_Loss:3.46018 Test_Loss:3.05441 Train_Acc:0.17969 Test_Acc:0.23200 gap:-0.05231 Train_Speed:1054.55731728 Test_Speed:2861.9988272 best_epoch:1 best_test_acc:0.232001582278 E:1 S:550 Train_Loss:3.19501 Test_Loss:3.25045 Train_Acc:0.21094 Test_Acc:0.22884 gap:-0.01790 Train_Speed:1078.60200463 Test_Speed:2156.15137653 best_epoch:1 best_test_acc:0.232001582278 E:1 S:600 Train_Loss:3.27641 Test_Loss:3.39094 Train_Acc:0.25781 Test_Acc:0.22785 gap:0.02996 Train_Speed:1104.35453162 Test_Speed:1596.927071 best_epoch:1 best_test_acc:0.232001582278 E:1 S:650 Train_Loss:3.69020 Test_Loss:3.44635 Train_Acc:0.30469 Test_Acc:0.23299 gap:0.07170 Train_Speed:1148.32311358 Test_Speed:4115.0885455 best_epoch:1 best_test_acc:0.232990506329 E:1 S:700 Train_Loss:3.41978 Test_Loss:3.65220 Train_Acc:0.21094 Test_Acc:0.22439 gap:-0.01345 Train_Speed:1127.96300568 Test_Speed:4346.32345393 best_epoch:1 best_test_acc:0.232990506329 E:1 S:750 Train_Loss:4.83669 Test_Loss:3.69271 Train_Acc:0.17188 Test_Acc:0.22725 gap:-0.05538 Train_Speed:1152.00192261 Test_Speed:4489.18750418 best_epoch:1 best_test_acc:0.232990506329 E:2 S:800 Train_Loss:4.27469 Test_Loss:3.61413 Train_Acc:0.28125 Test_Acc:0.22112 gap:0.06013 Train_Speed:1220.90758214 Test_Speed:4418.07246723 best_epoch:1 best_test_acc:0.232990506329 E:2 S:850 Train_Loss:3.60413 Test_Loss:3.57808 Train_Acc:0.23438 Test_Acc:0.23329 gap:0.00109 Train_Speed:1153.19462744 Test_Speed:4376.51043849 best_epoch:2 best_test_acc:0.233287183544 E:2 S:900 Train_Loss:4.30450 Test_Loss:4.45190 Train_Acc:0.16406 Test_Acc:0.19482 gap:-0.03076 Train_Speed:1208.02599343 Test_Speed:4029.47335555 best_epoch:2 best_test_acc:0.233287183544 E:2 S:950 Train_Loss:3.46003 Test_Loss:3.87522 Train_Acc:0.23438 Test_Acc:0.22627 gap:0.00811 Train_Speed:1052.51883423 Test_Speed:4551.45064261 best_epoch:2 best_test_acc:0.233287183544 E:2 S:1000 Train_Loss:4.13542 Test_Loss:3.67394 Train_Acc:0.15625 Test_Acc:0.22627 gap:-0.07002 Train_Speed:1064.42174718 Test_Speed:4086.7086245 best_epoch:2 best_test_acc:0.233287183544 E:2 S:1050 Train_Loss:4.47440 Test_Loss:3.42450 Train_Acc:0.20312 Test_Acc:0.23240 gap:-0.02927 Train_Speed:1202.16152696 Test_Speed:1849.06650318 best_epoch:2 best_test_acc:0.233287183544 E:2 S:1100 Train_Loss:3.42983 Test_Loss:3.46037 Train_Acc:0.24219 Test_Acc:0.23378 gap:0.00841 Train_Speed:1192.55089996 Test_Speed:4545.28524501 best_epoch:2 best_test_acc:0.23378164557 E:2 S:1150 Train_Loss:3.74305 Test_Loss:4.28859 Train_Acc:0.27344 Test_Acc:0.22824 gap:0.04519 Train_Speed:1212.62634449 Test_Speed:4841.90937951 best_epoch:2 best_test_acc:0.23378164557 E:3 S:1200 Train_Loss:3.76206 Test_Loss:3.55129 Train_Acc:0.19531 Test_Acc:0.22310 gap:-0.02779 Train_Speed:1127.22646533 Test_Speed:4313.81001816 best_epoch:2 best_test_acc:0.23378164557 E:3 S:1250 Train_Loss:4.28606 Test_Loss:3.66876 Train_Acc:0.20312 Test_Acc:0.22112 gap:-0.01800 Train_Speed:729.008353702 Test_Speed:4447.38818384 best_epoch:2 best_test_acc:0.23378164557 E:3 S:1300 Train_Loss:3.53226 Test_Loss:3.73910 Train_Acc:0.28906 Test_Acc:0.22597 gap:0.06309 Train_Speed:1211.21469148 Test_Speed:4358.03680464 best_epoch:2 best_test_acc:0.23378164557 E:3 S:1350 Train_Loss:5.38663 Test_Loss:4.59524 Train_Acc:0.17969 Test_Acc:0.18018 gap:-0.00049 Train_Speed:1144.28712527 Test_Speed:4323.43277741 best_epoch:2 best_test_acc:0.23378164557 E:3 S:1400 Train_Loss:2.33098 Test_Loss:3.50315 Train_Acc:0.33594 Test_Acc:0.23883 gap:0.09711 Train_Speed:1176.40427201 Test_Speed:4242.49610418 best_epoch:3 best_test_acc:0.238825158228 E:3 S:1450 Train_Loss:3.68373 Test_Loss:3.56768 Train_Acc:0.25781 Test_Acc:0.23438 gap:0.02344 Train_Speed:1181.28091322 Test_Speed:4338.52609802 best_epoch:3 best_test_acc:0.238825158228 E:3 S:1500 Train_Loss:3.55359 Test_Loss:3.43764 Train_Acc:0.27344 Test_Acc:0.23803 gap:0.03540 Train_Speed:1195.1898449 Test_Speed:4333.3783618 best_epoch:3 best_test_acc:0.238825158228 E:3 S:1550 Train_Loss:3.81624 Test_Loss:3.67806 Train_Acc:0.21094 Test_Acc:0.22834 gap:-0.01741 Train_Speed:1177.21682882 Test_Speed:3657.78172032 best_epoch:3 best_test_acc:0.238825158228 E:4 S:1600 Train_Loss:4.00349 Test_Loss:3.45820 Train_Acc:0.19531 Test_Acc:0.22716 gap:-0.03184 Train_Speed:1172.09718738 Test_Speed:2576.48981394 best_epoch:3 best_test_acc:0.238825158228 E:4 S:1650 Train_Loss:3.18337 Test_Loss:3.47405 Train_Acc:0.28906 Test_Acc:0.24080 gap:0.04826 Train_Speed:1212.5578746 Test_Speed:4253.04924266 best_epoch:4 best_test_acc:0.240803006329 E:4 S:1700 Train_Loss:3.31928 Test_Loss:3.50022 Train_Acc:0.25000 Test_Acc:0.22814 gap:0.02186 Train_Speed:1132.56259493 Test_Speed:4611.6195401 best_epoch:4 best_test_acc:0.240803006329 E:4 S:1750 Train_Loss:2.78207 Test_Loss:3.33416 Train_Acc:0.32031 Test_Acc:0.24021 gap:0.08010 Train_Speed:1101.30303372 Test_Speed:3714.98399474 best_epoch:4 best_test_acc:0.240803006329 E:4 S:1800 Train_Loss:3.21437 Test_Loss:3.06041 Train_Acc:0.23438 Test_Acc:0.25524 gap:-0.02087 Train_Speed:1123.88272803 Test_Speed:2048.52357132 best_epoch:4 best_test_acc:0.255241297468 E:4 S:1850 Train_Loss:2.76578 Test_Loss:3.30497 Train_Acc:0.28125 Test_Acc:0.23012 gap:0.05113 Train_Speed:1155.85118013 Test_Speed:4870.46096344 best_epoch:4 best_test_acc:0.255241297468 E:4 S:1900 Train_Loss:4.05632 Test_Loss:3.19663 Train_Acc:0.19531 Test_Acc:0.24397 gap:-0.04866 Train_Speed:1083.59823353 Test_Speed:3997.13292732 best_epoch:4 best_test_acc:0.255241297468 E:4 S:1950 Train_Loss:2.91405 Test_Loss:3.38560 Train_Acc:0.21875 Test_Acc:0.23230 gap:-0.01355 Train_Speed:1171.46547383 Test_Speed:4092.19104532 best_epoch:4 best_test_acc:0.255241297468 E:5 S:2000 Train_Loss:3.53234 Test_Loss:3.26789 Train_Acc:0.18750 Test_Acc:0.23586 gap:-0.04836 Train_Speed:1186.14049755 Test_Speed:4045.50525967 best_epoch:4 best_test_acc:0.255241297468 E:5 S:2050 Train_Loss:2.98674 Test_Loss:2.92163 Train_Acc:0.27344 Test_Acc:0.25475 gap:0.01869 Train_Speed:1127.84215657 Test_Speed:4047.30463102 best_epoch:4 best_test_acc:0.255241297468 E:5 S:2100 Train_Loss:3.03509 Test_Loss:3.14473 Train_Acc:0.24219 Test_Acc:0.24031 gap:0.00188 Train_Speed:1121.19529361 Test_Speed:4454.65787137 best_epoch:4 best_test_acc:0.255241297468 E:5 S:2150 Train_Loss:3.29612 Test_Loss:3.11643 Train_Acc:0.23438 Test_Acc:0.23507 gap:-0.00069 Train_Speed:1052.83050452 Test_Speed:4373.05252183 best_epoch:4 best_test_acc:0.255241297468 E:5 S:2200 Train_Loss:3.33340 Test_Loss:2.87789 Train_Acc:0.18750 Test_Acc:0.25326 gap:-0.06576 Train_Speed:1209.29128403 Test_Speed:4496.10505159 best_epoch:4 best_test_acc:0.255241297468 E:5 S:2250 Train_Loss:2.99067 Test_Loss:2.95525 Train_Acc:0.20312 Test_Acc:0.24644 gap:-0.04331 Train_Speed:1172.8576809 Test_Speed:4567.67581272 best_epoch:4 best_test_acc:0.255241297468 E:5 S:2300 Train_Loss:2.87697 Test_Loss:3.11992 Train_Acc:0.30469 Test_Acc:0.24080 gap:0.06388 Train_Speed:1168.53324047 Test_Speed:4591.269462 best_epoch:4 best_test_acc:0.255241297468 E:6 S:2350 Train_Loss:3.13953 Test_Loss:3.04756 Train_Acc:0.21875 Test_Acc:0.24802 gap:-0.02927 Train_Speed:1207.44457689 Test_Speed:4524.10412155 best_epoch:4 best_test_acc:0.255241297468 E:6 S:2400 Train_Loss:2.98216 Test_Loss:3.08932 Train_Acc:0.25781 Test_Acc:0.22261 gap:0.03521 Train_Speed:1200.55930445 Test_Speed:4667.95562202 best_epoch:4 best_test_acc:0.255241297468 E:6 S:2450 Train_Loss:2.64439 Test_Loss:2.69463 Train_Acc:0.20312 Test_Acc:0.26147 gap:-0.05835 Train_Speed:1114.4111443 Test_Speed:4953.36911934 best_epoch:6 best_test_acc:0.261471518987 E:6 S:2500 Train_Loss:2.96564 Test_Loss:2.77728 Train_Acc:0.26562 Test_Acc:0.25415 gap:0.01147 Train_Speed:1167.10560389 Test_Speed:4997.86736176 best_epoch:6 best_test_acc:0.261471518987 E:6 S:2550 Train_Loss:2.79762 Test_Loss:2.60256 Train_Acc:0.27344 Test_Acc:0.26948 gap:0.00396 Train_Speed:1188.54321294 Test_Speed:4400.29270212 best_epoch:6 best_test_acc:0.269481803797 E:6 S:2600 Train_Loss:3.05216 Test_Loss:2.65896 Train_Acc:0.27344 Test_Acc:0.24219 gap:0.03125 Train_Speed:1179.79864323 Test_Speed:3713.36518696 best_epoch:6 best_test_acc:0.269481803797 E:6 S:2650 Train_Loss:2.64152 Test_Loss:2.74660 Train_Acc:0.29688 Test_Acc:0.24268 gap:0.05419 Train_Speed:1183.3456663 Test_Speed:4086.98861915 best_epoch:6 best_test_acc:0.269481803797 E:6 S:2700 Train_Loss:2.67641 Test_Loss:2.69596 Train_Acc:0.32031 Test_Acc:0.25752 gap:0.06280 Train_Speed:1191.00852319 Test_Speed:4357.89530419 best_epoch:6 best_test_acc:0.269481803797 E:7 S:2750 Train_Loss:3.38848 Test_Loss:3.04510 Train_Acc:0.21875 Test_Acc:0.23081 gap:-0.01206 Train_Speed:1178.92038507 Test_Speed:4261.82732671 best_epoch:6 best_test_acc:0.269481803797 E:7 S:2800 Train_Loss:2.92383 Test_Loss:2.62347 Train_Acc:0.30469 Test_Acc:0.25069 gap:0.05400 Train_Speed:1024.04883045 Test_Speed:2370.3642585 best_epoch:6 best_test_acc:0.269481803797 E:7 S:2850 Train_Loss:2.77859 Test_Loss:2.46160 Train_Acc:0.25781 Test_Acc:0.27611 gap:-0.01830 Train_Speed:1216.7983754 Test_Speed:3980.09409218 best_epoch:7 best_test_acc:0.276107594937 E:7 S:2900 Train_Loss:3.13493 Test_Loss:2.72927 Train_Acc:0.18750 Test_Acc:0.26315 gap:-0.07565 Train_Speed:1172.26866037 Test_Speed:3305.10235968 best_epoch:7 best_test_acc:0.276107594937 E:7 S:2950 Train_Loss:2.52136 Test_Loss:2.53657 Train_Acc:0.22656 Test_Acc:0.26909 gap:-0.04252 Train_Speed:1078.20342979 Test_Speed:3911.25730896 best_epoch:7 best_test_acc:0.276107594937 E:7 S:3000 Train_Loss:2.28862 Test_Loss:2.50843 Train_Acc:0.31250 Test_Acc:0.24911 gap:0.06339 Train_Speed:1191.64033552 Test_Speed:4173.30704891 best_epoch:7 best_test_acc:0.276107594937 E:7 S:3050 Train_Loss:2.69223 Test_Loss:2.59592 Train_Acc:0.26562 Test_Acc:0.25613 gap:0.00949 Train_Speed:1166.09668115 Test_Speed:4135.18379419 best_epoch:7 best_test_acc:0.276107594937 E:7 S:3100 Train_Loss:2.40791 Test_Loss:2.45942 Train_Acc:0.27344 Test_Acc:0.27225 gap:0.00119 Train_Speed:1125.90736399 Test_Speed:2323.13093147 best_epoch:7 best_test_acc:0.276107594937 E:8 S:3150 Train_Loss:2.34843 Test_Loss:2.48130 Train_Acc:0.28125 Test_Acc:0.26849 gap:0.01276 Train_Speed:1194.38683856 Test_Speed:3490.02738087 best_epoch:7 best_test_acc:0.276107594937 E:8 S:3200 Train_Loss:2.43000 Test_Loss:2.42790 Train_Acc:0.25781 Test_Acc:0.26404 gap:-0.00623 Train_Speed:1221.80504679 Test_Speed:4578.15356278 best_epoch:7 best_test_acc:0.276107594937 E:8 S:3250 Train_Loss:2.51207 Test_Loss:2.39278 Train_Acc:0.22656 Test_Acc:0.27611 gap:-0.04955 Train_Speed:1215.40176083 Test_Speed:4282.53082648 best_epoch:7 best_test_acc:0.276107594937 E:8 S:3300 Train_Loss:2.28570 Test_Loss:2.33994 Train_Acc:0.32031 Test_Acc:0.28738 gap:0.03293 Train_Speed:1221.17696084 Test_Speed:4558.25192732 best_epoch:8 best_test_acc:0.287381329114 E:8 S:3350 Train_Loss:2.16668 Test_Loss:2.37339 Train_Acc:0.31250 Test_Acc:0.27858 gap:0.03392 Train_Speed:1119.23421803 Test_Speed:4051.91710064 best_epoch:8 best_test_acc:0.287381329114 E:8 S:3400 Train_Loss:2.33357 Test_Loss:2.39103 Train_Acc:0.32031 Test_Acc:0.27739 gap:0.04292 Train_Speed:1179.14565401 Test_Speed:4524.10412155 best_epoch:8 best_test_acc:0.287381329114 E:8 S:3450 Train_Loss:2.57082 Test_Loss:2.32535 Train_Acc:0.21875 Test_Acc:0.28135 gap:-0.06260 Train_Speed:1147.57201731 Test_Speed:4240.38505951 best_epoch:8 best_test_acc:0.287381329114 E:8 S:3500 Train_Loss:2.48471 Test_Loss:2.42561 Train_Acc:0.30469 Test_Acc:0.27413 gap:0.03056 Train_Speed:1167.33907069 Test_Speed:4199.7505515 best_epoch:8 best_test_acc:0.287381329114 E:9 S:3550 Train_Loss:2.42667 Test_Loss:2.41197 Train_Acc:0.29688 Test_Acc:0.27769 gap:0.01919 Train_Speed:1165.60805744 Test_Speed:4193.6815004 best_epoch:8 best_test_acc:0.287381329114 E:9 S:3600 Train_Loss:2.51808 Test_Loss:2.34828 Train_Acc:0.22656 Test_Acc:0.27383 gap:-0.04727 Train_Speed:916.42625945 Test_Speed:4169.51493076 best_epoch:8 best_test_acc:0.287381329114 E:9 S:3650 Train_Loss:2.47517 Test_Loss:2.29117 Train_Acc:0.21875 Test_Acc:0.28471 gap:-0.06596 Train_Speed:894.810206289 Test_Speed:2389.84233931 best_epoch:8 best_test_acc:0.287381329114 E:9 S:3700 Train_Loss:2.30268 Test_Loss:2.28041 Train_Acc:0.28125 Test_Acc:0.27838 gap:0.00287 Train_Speed:1198.9051208 Test_Speed:4522.65588381 best_epoch:8 best_test_acc:0.287381329114 E:9 S:3750 Train_Loss:2.17532 Test_Loss:2.27726 Train_Acc:0.31250 Test_Acc:0.29678 gap:0.01572 Train_Speed:1188.26699719 Test_Speed:4831.27777978 best_epoch:9 best_test_acc:0.296776107595 E:9 S:3800 Train_Loss:2.46029 Test_Loss:2.26493 Train_Acc:0.31250 Test_Acc:0.28135 gap:0.03115 Train_Speed:1273.45526654 Test_Speed:4857.90084604 best_epoch:9 best_test_acc:0.296776107595 E:9 S:3850 Train_Loss:2.04513 Test_Loss:2.38796 Train_Acc:0.32031 Test_Acc:0.26701 gap:0.05330 Train_Speed:1103.55321043 Test_Speed:4519.15346089 best_epoch:9 best_test_acc:0.296776107595 E:9 S:3900 Train_Loss:2.66478 Test_Loss:2.31545 Train_Acc:0.23438 Test_Acc:0.28610 gap:-0.05172 Train_Speed:1060.8943929 Test_Speed:4518.46883863 best_epoch:9 best_test_acc:0.296776107595 E:10 S:3950 Train_Loss:2.51084 Test_Loss:2.40000 Train_Acc:0.30469 Test_Acc:0.27077 gap:0.03392 Train_Speed:1210.99612478 Test_Speed:4583.70397691 best_epoch:9 best_test_acc:0.296776107595 E:10 S:4000 Train_Loss:2.33085 Test_Loss:2.30818 Train_Acc:0.27344 Test_Acc:0.28333 gap:-0.00989 Train_Speed:1091.57377316 Test_Speed:3022.49633779 best_epoch:9 best_test_acc:0.296776107595 E:10 S:4050 Train_Loss:2.81832 Test_Loss:2.24091 Train_Acc:0.22656 Test_Acc:0.28768 gap:-0.06112 Train_Speed:1113.64367694 Test_Speed:4559.37454459 best_epoch:9 best_test_acc:0.296776107595 E:10 S:4100 Train_Loss:2.28684 Test_Loss:2.44056 Train_Acc:0.29688 Test_Acc:0.27205 gap:0.02482 Train_Speed:749.814473363 Test_Speed:4034.80318653 best_epoch:9 best_test_acc:0.296776107595 E:10 S:4150 Train_Loss:2.43879 Test_Loss:2.24455 Train_Acc:0.27344 Test_Acc:0.28026 gap:-0.00682 Train_Speed:1191.75143455 Test_Speed:4189.85228195 best_epoch:9 best_test_acc:0.296776107595 E:10 S:4200 Train_Loss:2.37320 Test_Loss:2.17202 Train_Acc:0.32031 Test_Acc:0.29945 gap:0.02087 Train_Speed:1156.89479423 Test_Speed:3642.27213026 best_epoch:10 best_test_acc:0.299446202532 E:10 S:4250 Train_Loss:2.32131 Test_Loss:2.30443 Train_Acc:0.35156 Test_Acc:0.29173 gap:0.05983 Train_Speed:676.393254095 Test_Speed:4281.06240531 best_epoch:10 best_test_acc:0.299446202532 E:10 S:4300 Train_Loss:2.44296 Test_Loss:2.24309 Train_Acc:0.28906 Test_Acc:0.29173 gap:-0.00267 Train_Speed:1041.9619835 Test_Speed:4106.77828774 best_epoch:10 best_test_acc:0.299446202532 E:11 S:4350 Train_Loss:2.15289 Test_Loss:2.24860 Train_Acc:0.27344 Test_Acc:0.27828 gap:-0.00485 Train_Speed:1090.70889152 Test_Speed:4325.4879389 best_epoch:10 best_test_acc:0.299446202532 E:11 S:4400 Train_Loss:2.66482 Test_Loss:2.22693 Train_Acc:0.25000 Test_Acc:0.28214 gap:-0.03214 Train_Speed:1193.13927083 Test_Speed:4414.69379163 best_epoch:10 best_test_acc:0.299446202532 E:11 S:4450 Train_Loss:2.36792 Test_Loss:2.22178 Train_Acc:0.30469 Test_Acc:0.28165 gap:0.02304 Train_Speed:722.323011183 Test_Speed:4381.76122229 best_epoch:10 best_test_acc:0.299446202532 E:11 S:4500 Train_Loss:2.14378 Test_Loss:2.21626 Train_Acc:0.28125 Test_Acc:0.27393 gap:0.00732 Train_Speed:1052.85734569 Test_Speed:4148.73276355 best_epoch:10 best_test_acc:0.299446202532 E:11 S:4550 Train_Loss:2.09879 Test_Loss:2.14655 Train_Acc:0.31250 Test_Acc:0.28046 gap:0.03204 Train_Speed:1064.5272759 Test_Speed:4004.49709474 best_epoch:10 best_test_acc:0.299446202532 E:11 S:4600 Train_Loss:2.43194 Test_Loss:2.16936 Train_Acc:0.27344 Test_Acc:0.28066 gap:-0.00722 Train_Speed:1010.70800835 Test_Speed:3921.22728136 best_epoch:10 best_test_acc:0.299446202532 E:11 S:4650 Train_Loss:2.31174 Test_Loss:2.19297 Train_Acc:0.34375 Test_Acc:0.28748 gap:0.05627 Train_Speed:1081.38069725 Test_Speed:4239.54792552 best_epoch:10 best_test_acc:0.299446202532 E:12 S:4700 Train_Loss:2.87508 Test_Loss:2.36022 Train_Acc:0.28125 Test_Acc:0.27275 gap:0.00850 Train_Speed:1180.24734435 Test_Speed:1132.95456119 best_epoch:10 best_test_acc:0.299446202532 E:12 S:4750 Train_Loss:2.18191 Test_Loss:2.26949 Train_Acc:0.31250 Test_Acc:0.28590 gap:0.02660 Train_Speed:1057.97378274 Test_Speed:4217.86472876 best_epoch:10 best_test_acc:0.299446202532 E:12 S:4800 Train_Loss:1.92283 Test_Loss:2.23566 Train_Acc:0.32031 Test_Acc:0.28352 gap:0.03679 Train_Speed:1215.3549871 Test_Speed:4259.02115743 best_epoch:10 best_test_acc:0.299446202532 E:12 S:4850 Train_Loss:2.24625 Test_Loss:2.14552 Train_Acc:0.21094 Test_Acc:0.28867 gap:-0.07773 Train_Speed:1131.89401069 Test_Speed:4383.26376121 best_epoch:10 best_test_acc:0.299446202532 E:12 S:4900 Train_Loss:2.12096 Test_Loss:2.23744 Train_Acc:0.27344 Test_Acc:0.29242 gap:-0.01899 Train_Speed:1141.6955427 Test_Speed:3354.75128255 best_epoch:10 best_test_acc:0.299446202532 E:12 S:4950 Train_Loss:2.29347 Test_Loss:2.22825 Train_Acc:0.26562 Test_Acc:0.28441 gap:-0.01879 Train_Speed:1160.74424839 Test_Speed:3632.85726272 best_epoch:10 best_test_acc:0.299446202532 E:12 S:5000 Train_Loss:2.23441 Test_Loss:2.16348 Train_Acc:0.25000 Test_Acc:0.28669 gap:-0.03669 Train_Speed:1190.02381063 Test_Speed:4523.45611108 best_epoch:10 best_test_acc:0.299446202532 E:12 S:5050 Train_Loss:2.30547 Test_Loss:2.17714 Train_Acc:0.27344 Test_Acc:0.29707 gap:-0.02364 Train_Speed:957.589988727 Test_Speed:4973.00695647 best_epoch:10 best_test_acc:0.299446202532 E:13 S:5100 Train_Loss:2.34201 Test_Loss:2.23939 Train_Acc:0.25000 Test_Acc:0.27008 gap:-0.02008 Train_Speed:1159.12531279 Test_Speed:4573.08397077 best_epoch:10 best_test_acc:0.299446202532 E:13 S:5150 Train_Loss:2.31656 Test_Loss:2.16713 Train_Acc:0.28125 Test_Acc:0.29371 gap:-0.01246 Train_Speed:927.912890029 Test_Speed:3734.15670536 best_epoch:10 best_test_acc:0.299446202532 E:13 S:5200 Train_Loss:2.24781 Test_Loss:2.19491 Train_Acc:0.28125 Test_Acc:0.29312 gap:-0.01187 Train_Speed:1197.11132914 Test_Speed:4471.91190631 best_epoch:10 best_test_acc:0.299446202532 E:13 S:5250 Train_Loss:2.11684 Test_Loss:2.09564 Train_Acc:0.32031 Test_Acc:0.30380 gap:0.01652 Train_Speed:1194.64198185 Test_Speed:4327.68459151 best_epoch:13 best_test_acc:0.303797468354 E:13 S:5300 Train_Loss:2.12072 Test_Loss:2.11669 Train_Acc:0.28906 Test_Acc:0.30113 gap:-0.01206 Train_Speed:1217.7119022 Test_Speed:4333.55325418 best_epoch:13 best_test_acc:0.303797468354 E:13 S:5350 Train_Loss:2.33529 Test_Loss:2.15970 Train_Acc:0.29688 Test_Acc:0.28511 gap:0.01177 Train_Speed:1158.78005728 Test_Speed:4201.26234075 best_epoch:13 best_test_acc:0.303797468354 E:13 S:5400 Train_Loss:2.37300 Test_Loss:2.07269 Train_Acc:0.26562 Test_Acc:0.30953 gap:-0.04391 Train_Speed:1079.58453551 Test_Speed:4171.68564191 best_epoch:13 best_test_acc:0.309533227848 E:13 S:5450 Train_Loss:1.91987 Test_Loss:2.06280 Train_Acc:0.35156 Test_Acc:0.30538 gap:0.04618 Train_Speed:912.336690701 Test_Speed:4424.29838643 best_epoch:13 best_test_acc:0.309533227848 E:14 S:5500 Train_Loss:2.27801 Test_Loss:2.11238 Train_Acc:0.26562 Test_Acc:0.30231 gap:-0.03669 Train_Speed:1053.17748052 Test_Speed:4062.58730231 best_epoch:13 best_test_acc:0.309533227848 E:14 S:5550 Train_Loss:1.97383 Test_Loss:2.08985 Train_Acc:0.30469 Test_Acc:0.29826 gap:0.00643 Train_Speed:1172.31217642 Test_Speed:4422.36683992 best_epoch:13 best_test_acc:0.309533227848 E:14 S:5600 Train_Loss:2.49850 Test_Loss:2.10744 Train_Acc:0.28125 Test_Acc:0.29905 gap:-0.01780 Train_Speed:1137.51549258 Test_Speed:4432.59036155 best_epoch:13 best_test_acc:0.309533227848 E:14 S:5650 Train_Loss:2.00086 Test_Loss:2.15149 Train_Acc:0.35156 Test_Acc:0.29826 gap:0.05330 Train_Speed:1198.23081644 Test_Speed:4246.01918681 best_epoch:13 best_test_acc:0.309533227848 E:14 S:5700 Train_Loss:2.08668 Test_Loss:2.06769 Train_Acc:0.34375 Test_Acc:0.29974 gap:0.04401 Train_Speed:1185.27109274 Test_Speed:4277.78770059 best_epoch:13 best_test_acc:0.309533227848 E:14 S:5750 Train_Loss:2.26744 Test_Loss:2.04210 Train_Acc:0.31250 Test_Acc:0.30953 gap:0.00297 Train_Speed:1186.84323713 Test_Speed:4899.17243393 best_epoch:13 best_test_acc:0.309533227848 E:14 S:5800 Train_Loss:2.42013 Test_Loss:2.05375 Train_Acc:0.25000 Test_Acc:0.30063 gap:-0.05063 Train_Speed:1212.24574991 Test_Speed:4548.21172484 best_epoch:13 best_test_acc:0.309533227848 E:14 S:5850 Train_Loss:2.10737 Test_Loss:2.07919 Train_Acc:0.33594 Test_Acc:0.30696 gap:0.02898 Train_Speed:1065.12073676 Test_Speed:4101.38206264 best_epoch:13 best_test_acc:0.309533227848 E:15 S:5900 Train_Loss:2.12405 Test_Loss:2.08203 Train_Acc:0.33594 Test_Acc:0.29272 gap:0.04322 Train_Speed:1091.60928477 Test_Speed:4205.4089079 best_epoch:13 best_test_acc:0.309533227848 E:15 S:5950 Train_Loss:2.24514 Test_Loss:2.05870 Train_Acc:0.30469 Test_Acc:0.29579 gap:0.00890 Train_Speed:1156.53093441 Test_Speed:4327.68459151 best_epoch:13 best_test_acc:0.309533227848 E:15 S:6000 Train_Loss:2.03245 Test_Loss:2.09055 Train_Acc:0.31250 Test_Acc:0.28797 gap:0.02453 Train_Speed:1155.93330577 Test_Speed:4050.38862902 best_epoch:13 best_test_acc:0.309533227848 E:15 S:6050 Train_Loss:2.24727 Test_Loss:2.10345 Train_Acc:0.32031 Test_Acc:0.30775 gap:0.01256 Train_Speed:1166.74543621 Test_Speed:4533.99976353 best_epoch:13 best_test_acc:0.309533227848 E:15 S:6100 Train_Loss:1.92576 Test_Loss:2.07067 Train_Acc:0.28125 Test_Acc:0.29688 gap:-0.01562 Train_Speed:1136.7471728 Test_Speed:4419.41810998 best_epoch:13 best_test_acc:0.309533227848 E:15 S:6150 Train_Loss:2.08536 Test_Loss:2.05159 Train_Acc:0.33594 Test_Acc:0.31477 gap:0.02116 Train_Speed:1177.81341758 Test_Speed:4039.90392198 best_epoch:15 best_test_acc:0.314774525316 E:15 S:6200 Train_Loss:1.90426 Test_Loss:2.09866 Train_Acc:0.29688 Test_Acc:0.29529 gap:0.00158 Train_Speed:1219.97539466 Test_Speed:4544.93893757 best_epoch:15 best_test_acc:0.314774525316 E:15 S:6250 Train_Loss:2.19895 Test_Loss:2.07545 Train_Acc:0.28906 Test_Acc:0.30133 gap:-0.01226 Train_Speed:1187.86999763 Test_Speed:3007.37691437 best_epoch:15 best_test_acc:0.314774525316 E:16 S:6300 Train_Loss:2.09298 Test_Loss:2.06988 Train_Acc:0.25000 Test_Acc:0.29153 gap:-0.04153 Train_Speed:1104.79086651 Test_Speed:2219.63969521 best_epoch:15 best_test_acc:0.314774525316 E:16 S:6350 Train_Loss:2.27335 Test_Loss:2.03631 Train_Acc:0.25781 Test_Acc:0.30439 gap:-0.04658 Train_Speed:1195.23507935 Test_Speed:4413.7864266 best_epoch:15 best_test_acc:0.314774525316 E:16 S:6400 Train_Loss:2.01116 Test_Loss:2.04440 Train_Acc:0.27344 Test_Acc:0.29529 gap:-0.02186 Train_Speed:1213.38550865 Test_Speed:3909.83309665 best_epoch:15 best_test_acc:0.314774525316 E:16 S:6450 Train_Loss:1.89976 Test_Loss:2.05661 Train_Acc:0.34375 Test_Acc:0.30597 gap:0.03778 Train_Speed:884.8873879 Test_Speed:3824.87487443 best_epoch:15 best_test_acc:0.314774525316 E:16 S:6500 Train_Loss:1.93772 Test_Loss:2.03792 Train_Acc:0.32031 Test_Acc:0.29935 gap:0.02097 Train_Speed:1203.35791135 Test_Speed:4516.11228224 best_epoch:15 best_test_acc:0.314774525316 E:16 S:6550 Train_Loss:2.47803 Test_Loss:2.05896 Train_Acc:0.26562 Test_Acc:0.30172 gap:-0.03610 Train_Speed:1062.20416239 Test_Speed:3959.89667864 best_epoch:15 best_test_acc:0.314774525316 E:16 S:6600 Train_Loss:2.27755 Test_Loss:2.01047 Train_Acc:0.24219 Test_Acc:0.30845 gap:-0.06626 Train_Speed:1142.23479101 Test_Speed:4090.22689837 best_epoch:15 best_test_acc:0.314774525316 E:17 S:6650 Train_Loss:2.56299 Test_Loss:2.43982 Train_Acc:0.27344 Test_Acc:0.27888 gap:-0.00544 Train_Speed:973.91018643 Test_Speed:3597.82411323 best_epoch:15 best_test_acc:0.314774525316 E:17 S:6700 Train_Loss:1.98020 Test_Loss:2.07314 Train_Acc:0.25781 Test_Acc:0.29717 gap:-0.03936 Train_Speed:1261.96667795 Test_Speed:4582.7258154 best_epoch:15 best_test_acc:0.314774525316 E:17 S:6750 Train_Loss:2.05947 Test_Loss:2.04610 Train_Acc:0.28906 Test_Acc:0.30182 gap:-0.01276 Train_Speed:1256.0323044 Test_Speed:4453.88179857 best_epoch:15 best_test_acc:0.314774525316 E:17 S:6800 Train_Loss:1.94956 Test_Loss:2.03221 Train_Acc:0.32812 Test_Acc:0.30103 gap:0.02710 Train_Speed:1281.53810091 Test_Speed:4259.83220001 best_epoch:15 best_test_acc:0.314774525316 E:17 S:6850 Train_Loss:1.93869 Test_Loss:2.05692 Train_Acc:0.33594 Test_Acc:0.30508 gap:0.03085 Train_Speed:1217.63181028 Test_Speed:4405.31153944 best_epoch:15 best_test_acc:0.314774525316 E:17 S:6900 Train_Loss:2.07925 Test_Loss:2.03596 Train_Acc:0.28906 Test_Acc:0.29341 gap:-0.00435 Train_Speed:1283.28416161 Test_Speed:1913.09847521 best_epoch:15 best_test_acc:0.314774525316 E:17 S:6950 Train_Loss:2.03528 Test_Loss:2.00767 Train_Acc:0.28125 Test_Acc:0.30696 gap:-0.02571 Train_Speed:1282.26964231 Test_Speed:4485.06217106 best_epoch:15 best_test_acc:0.314774525316 E:17 S:7000 Train_Loss:1.76588 Test_Loss:2.01119 Train_Acc:0.38281 Test_Acc:0.30746 gap:0.07536 Train_Speed:948.857053983 Test_Speed:3732.75471226 best_epoch:15 best_test_acc:0.314774525316 E:18 S:7050 Train_Loss:2.37053 Test_Loss:2.15638 Train_Acc:0.34375 Test_Acc:0.29035 gap:0.05340 Train_Speed:1263.3624157 Test_Speed:4704.52436951 best_epoch:15 best_test_acc:0.314774525316 E:18 S:7100 Train_Loss:1.94380 Test_Loss:2.05977 Train_Acc:0.32031 Test_Acc:0.29836 gap:0.02195 Train_Speed:1201.46204526 Test_Speed:3903.38019485 best_epoch:15 best_test_acc:0.314774525316 E:18 S:7150 Train_Loss:2.10513 Test_Loss:2.04257 Train_Acc:0.26562 Test_Acc:0.30103 gap:-0.03540 Train_Speed:1274.54664847 Test_Speed:5034.18736931 best_epoch:15 best_test_acc:0.314774525316 E:18 S:7200 Train_Loss:2.02944 Test_Loss:2.00461 Train_Acc:0.33594 Test_Acc:0.30756 gap:0.02838 Train_Speed:1299.57084969 Test_Speed:3688.64292634 best_epoch:15 best_test_acc:0.314774525316 E:18 S:7250 Train_Loss:1.96929 Test_Loss:2.06279 Train_Acc:0.31250 Test_Acc:0.30093 gap:0.01157 Train_Speed:1328.24071431 Test_Speed:4367.71597325 best_epoch:15 best_test_acc:0.314774525316 E:18 S:7300 Train_Loss:2.02918 Test_Loss:2.00739 Train_Acc:0.30469 Test_Acc:0.30113 gap:0.00356 Train_Speed:1311.75762077 Test_Speed:3894.12195812 best_epoch:15 best_test_acc:0.314774525316 E:18 S:7350 Train_Loss:2.13662 Test_Loss:1.99577 Train_Acc:0.33594 Test_Acc:0.31260 gap:0.02334 Train_Speed:1302.78433856 Test_Speed:4919.82434661 best_epoch:15 best_test_acc:0.314774525316 E:18 S:7400 Train_Loss:1.94479 Test_Loss:2.01498 Train_Acc:0.32812 Test_Acc:0.31181 gap:0.01632 Train_Speed:761.886458718 Test_Speed:4576.82658437 best_epoch:15 best_test_acc:0.314774525316 E:19 S:7450 Train_Loss:2.11968 Test_Loss:2.06430 Train_Acc:0.35156 Test_Acc:0.30756 gap:0.04401 Train_Speed:1241.10193816 Test_Speed:4498.4784616 best_epoch:15 best_test_acc:0.314774525316 early stop
单层神经网络处理cifar10 确实比较难 正确率才能到0.31. 但是这个比随机猜测要高.所以神经网络是学到了东西的.
Copyright © 2003-2013 www.wpsshop.cn 版权所有,并保留所有权利。