From 5dbdfc1865adce0fadd85ed7a3c4156edada26cd Mon Sep 17 00:00:00 2001 From: Matthew Moocarme Date: Fri, 10 Jan 2020 18:54:48 -0500 Subject: [PATCH] cleanup --- .../Activity4.02/Activity4.02_Unit_test.py | 6 +- Chapter04/Activity4.02/Activity4_02.ipynb | 48 +- Chapter04/Activity4.03/Activity4_03.ipynb | 26 +- Chapter04/Exercise4.02/Exercise4_02.ipynb | 1020 ++++++++--------- .../Exercise4.03/Exercise4.03_Unit_test.py | 6 +- Chapter04/Exercise4.03/Exercise4_03.ipynb | 44 +- Chapter05/Activity5.01/Activity5_01.ipynb | 14 +- .../Activity5.03/Activity5.03_Unit_test.py | 8 +- Chapter05/Exercise5.01/Exercise5_01.ipynb | 6 +- Chapter05/Exercise5.02/Exercise5_02.ipynb | 6 +- 10 files changed, 588 insertions(+), 596 deletions(-) diff --git a/Chapter04/Activity4.02/Activity4.02_Unit_test.py b/Chapter04/Activity4.02/Activity4.02_Unit_test.py index 2fa23a20..5fd11a23 100755 --- a/Chapter04/Activity4.02/Activity4.02_Unit_test.py +++ b/Chapter04/Activity4.02/Activity4.02_Unit_test.py @@ -80,7 +80,7 @@ def test_model_iter(self): self.results_1 =[] models = [build_model_1, build_model_2, build_model_3] for m in range(len(models)): - classifier = KerasClassifier(build_fn=models[m], epochs=epochs, batch_size=batch_size, verbose=0) + classifier = KerasClassifier(build_fn=models[m], epochs=epochs, batch_size=batch_size, verbose=0, shuffle=False) kfold = StratifiedKFold(n_splits=self.n_folds, shuffle=True, random_state=self.seed) result = cross_val_score(classifier, self.X, self.y, cv=kfold) self.results_1.append(result) @@ -99,7 +99,7 @@ def test_batch_epoch_iter(self): self.results_2 =[] for e in range(len(epochs)): for b in range(len(batches)): - classifier = KerasClassifier(build_fn=build_model_2, epochs=epochs[e], batch_size=batches[b], verbose=0) + classifier = KerasClassifier(build_fn=build_model_2, epochs=epochs[e], batch_size=batches[b], verbose=0, shuffle=False) kfold = StratifiedKFold(n_splits=self.n_folds, shuffle=True, random_state=self.seed) result = cross_val_score(classifier, self.X, self.y, cv=kfold) self.results_2.append(result) @@ -122,7 +122,7 @@ def test_opt_act_iter(self): for a in range(len(activations)): optimizer = optimizers[o] activation = activations[a] - classifier = KerasClassifier(build_fn=build_model_2, epochs=epochs, batch_size=batch_size, verbose=0) + classifier = KerasClassifier(build_fn=build_model_2, epochs=epochs, batch_size=batch_size, verbose=0, shuffle=False) kfold = StratifiedKFold(n_splits=self.n_folds, shuffle=True, random_state=self.seed) result = cross_val_score(classifier, self.X, self.y, cv=kfold) self.results_3.append(result) diff --git a/Chapter04/Activity4.02/Activity4_02.ipynb b/Chapter04/Activity4.02/Activity4_02.ipynb index 5d9e3dfd..bc2e31ad 100644 --- a/Chapter04/Activity4.02/Activity4_02.ipynb +++ b/Chapter04/Activity4.02/Activity4_02.ipynb @@ -29,9 +29,7 @@ { "cell_type": "code", "execution_count": 2, - "metadata": { - "collapsed": true - }, + "metadata": {}, "outputs": [], "source": [ "# Load the dataset\n", @@ -45,9 +43,7 @@ { "cell_type": "code", "execution_count": 3, - "metadata": { - "collapsed": true - }, + "metadata": {}, "outputs": [], "source": [ "# Create the function that returns the keras model 1\n", @@ -88,9 +84,7 @@ { "cell_type": "code", "execution_count": 4, - "metadata": { - "collapsed": true - }, + "metadata": {}, "outputs": [], "source": [ "# define a seed for random number generator so the result will be reproducible\n", @@ -109,7 +103,7 @@ "# loop over models\n", "for m in range(len(models)):\n", " # build the Scikit-Learn interface for the keras model\n", - " classifier = KerasClassifier(build_fn=models[m], epochs=epochs, batch_size=batch_size, verbose=0)\n", + " classifier = KerasClassifier(build_fn=models[m], epochs=epochs, batch_size=batch_size, verbose=0, shuffle=False)\n", " # define the cross validation iterator\n", " kfold = StratifiedKFold(n_splits=n_folds, shuffle=True, random_state=seed)\n", " # perform the k-fold cross validation and store the scores in result\n", @@ -127,9 +121,9 @@ "name": "stdout", "output_type": "stream", "text": [ - "Model 1 Test Accuracy = 0.49602887630462644\n", - "Model 2 Test Accuracy = 0.5090252697467804\n", - "Model 3 Test Accuracy = 0.513357400894165\n" + "Model 1 Test Accuracy = 0.5046931445598603\n", + "Model 2 Test Accuracy = 0.5054151535034179\n", + "Model 3 Test Accuracy = 0.507581228017807\n" ] } ], @@ -159,7 +153,7 @@ "for e in range(len(epochs)):\n", " for b in range(len(batches)):\n", " # build the Scikit-Learn interface for the keras model\n", - " classifier = KerasClassifier(build_fn=build_model_2, epochs=epochs[e], batch_size=batches[b], verbose=0)\n", + " classifier = KerasClassifier(build_fn=build_model_2, epochs=epochs[e], batch_size=batches[b], verbose=0, shuffle=False)\n", " # define the cross validation iterator\n", " kfold = StratifiedKFold(n_splits=n_folds, shuffle=True, random_state=seed)\n", " # perform the k-fold cross validation. \n", @@ -178,10 +172,10 @@ "name": "stdout", "output_type": "stream", "text": [ - "batch_size = 10 , epochs = 100 , Test Accuracy = 0.4916967451572418\n", - "batch_size = 20 , epochs = 100 , Test Accuracy = 0.5364620923995972\n", - "batch_size = 10 , epochs = 200 , Test Accuracy = 0.5061371803283692\n", - "batch_size = 20 , epochs = 200 , Test Accuracy = 0.505415165424347\n" + "batch_size = 10 , epochs = 100 , Test Accuracy = 0.5371841073036194\n", + "batch_size = 20 , epochs = 100 , Test Accuracy = 0.5025270760059357\n", + "batch_size = 10 , epochs = 200 , Test Accuracy = 0.49386281371116636\n", + "batch_size = 20 , epochs = 200 , Test Accuracy = 0.49891695380210876\n" ] } ], @@ -197,9 +191,7 @@ { "cell_type": "code", "execution_count": 8, - "metadata": { - "collapsed": true - }, + "metadata": {}, "outputs": [], "source": [ "# define a seed for random number generator so the result will be reproducible\n", @@ -220,7 +212,7 @@ " optimizer = optimizers[o]\n", " activation = activations[a]\n", " # build the Scikit-Learn interface for the keras model\n", - " classifier = KerasClassifier(build_fn=build_model_2, epochs=epochs, batch_size=batch_size, verbose=0)\n", + " classifier = KerasClassifier(build_fn=build_model_2, epochs=epochs, batch_size=batch_size, verbose=0, shuffle=False)\n", " # define the cross validation iterator\n", " kfold = StratifiedKFold(n_splits=n_folds, shuffle=True, random_state=seed)\n", " # perform the k-fold cross validation. \n", @@ -239,12 +231,12 @@ "name": "stdout", "output_type": "stream", "text": [ - "activation = relu , optimizer = rmsprop , Test accuracy = 0.4815884530544281\n", - "activation = tanh , optimizer = rmsprop , Test accuracy = 0.5046931385993958\n", - "activation = relu , optimizer = adam , Test accuracy = 0.4974729120731354\n", - "activation = tanh , optimizer = adam , Test accuracy = 0.5090252697467804\n", - "activation = relu , optimizer = sgd , Test accuracy = 0.5075812220573426\n", - "activation = tanh , optimizer = sgd , Test accuracy = 0.5140794217586517\n" + "activation = relu , optimizer = rmsprop , Test accuracy = 0.5523465752601624\n", + "activation = tanh , optimizer = rmsprop , Test accuracy = 0.5032491028308869\n", + "activation = relu , optimizer = adam , Test accuracy = 0.5039711177349091\n", + "activation = tanh , optimizer = adam , Test accuracy = 0.5119133591651917\n", + "activation = relu , optimizer = sgd , Test accuracy = 0.5046931326389312\n", + "activation = tanh , optimizer = sgd , Test accuracy = 0.4830324828624725\n" ] } ], diff --git a/Chapter04/Activity4.03/Activity4_03.ipynb b/Chapter04/Activity4.03/Activity4_03.ipynb index 8fce7f34..2025e8e4 100644 --- a/Chapter04/Activity4.03/Activity4_03.ipynb +++ b/Chapter04/Activity4.03/Activity4_03.ipynb @@ -113,7 +113,7 @@ "# loop over models\n", "for i in range(len(models)):\n", " # build the Scikit-Learn interface for the keras model\n", - " regressor = KerasRegressor(build_fn=models[i], epochs=100, batch_size=50, verbose=0)\n", + " regressor = KerasRegressor(build_fn=models[i], epochs=100, batch_size=50, verbose=0, shuffle=False)\n", " # build the pipeline of transformations so for each fold trainind set will be scaled \n", " # and test set will be scaled accordingly.\n", " model = make_pipeline(StandardScaler(), regressor)\n", @@ -135,9 +135,9 @@ "name": "stdout", "output_type": "stream", "text": [ - "Model 1 test error rate = 25.667806420326237\n", - "Model 2 test error rate = 25.12541407108307\n", - "Model 3 test error rate = 25.41868120670319\n" + "Model 1 test error rate = 25.5814275598526\n", + "Model 2 test error rate = 24.995588760375977\n", + "Model 3 test error rate = 25.413555855751035\n" ] } ], @@ -167,7 +167,7 @@ "for i in range(len(epochs)):\n", " for j in range(len(batches)):\n", " # build the Scikit-Learn interface for the keras model\n", - " regressor = KerasRegressor(build_fn=build_model_2, epochs=epochs[i], batch_size=batches[j], verbose=0)\n", + " regressor = KerasRegressor(build_fn=build_model_2, epochs=epochs[i], batch_size=batches[j], verbose=0, shuffle=False)\n", " # build the pipeline of transformations so for each fold trainind set will be scaled \n", " # and test set will be scaled accordingly.\n", " model = make_pipeline(StandardScaler(), regressor)\n", @@ -189,10 +189,10 @@ "name": "stdout", "output_type": "stream", "text": [ - "batch_size = 50 , epochs = 80 , Test error rate = 25.261057367324828\n", - "batch_size = 25 , epochs = 80 , Test error rate = 25.33181362867355\n", - "batch_size = 50 , epochs = 100 , Test error rate = 25.049520788192748\n", - "batch_size = 25 , epochs = 100 , Test error rate = 25.159788784980773\n" + "batch_size = 50 , epochs = 80 , Test error rate = 24.9977765750885\n", + "batch_size = 25 , epochs = 80 , Test error rate = 25.185024490356447\n", + "batch_size = 50 , epochs = 100 , Test error rate = 25.031687560081483\n", + "batch_size = 25 , epochs = 100 , Test error rate = 25.14500049829483\n" ] } ], @@ -224,7 +224,7 @@ "for i in range(len(optimizers)):\n", " optimizer=optimizers[i]\n", " # build the Scikit-Learn interface for the keras model\n", - " regressor = KerasRegressor(build_fn=build_model_2, epochs=100, batch_size=50, verbose=0)\n", + " regressor = KerasRegressor(build_fn=build_model_2, epochs=100, batch_size=50, verbose=0, shuffle=False)\n", " # build the pipeline of transformations so for each fold trainind set will be scaled \n", " # and test set will be scaled accordingly.\n", " model = make_pipeline(StandardScaler(), regressor)\n", @@ -246,9 +246,9 @@ "name": "stdout", "output_type": "stream", "text": [ - "optimizer= adam test error rate = 25.30731126308441\n", - "optimizer= sgd test error rate = 25.128445773124696\n", - "optimizer= rmsprop test error rate = 25.159849672317506\n" + "optimizer= adam test error rate = 25.013688187599183\n", + "optimizer= sgd test error rate = 25.04590551376343\n", + "optimizer= rmsprop test error rate = 25.028708510398864\n" ] } ], diff --git a/Chapter04/Exercise4.02/Exercise4_02.ipynb b/Chapter04/Exercise4.02/Exercise4_02.ipynb index 619fc655..aea3bbc9 100644 --- a/Chapter04/Exercise4.02/Exercise4_02.ipynb +++ b/Chapter04/Exercise4.02/Exercise4_02.ipynb @@ -72,7 +72,7 @@ "seed = 1\n", "np.random.seed(seed)\n", "random.set_seed(seed)\n", - "YourModel = KerasRegressor(build_fn= build_model, epochs=100, batch_size=20, verbose=1)" + "YourModel = KerasRegressor(build_fn= build_model, epochs=100, batch_size=20, verbose=1, shuffle=False)" ] }, { @@ -96,199 +96,199 @@ "output_type": "stream", "text": [ "Epoch 1/100\n", - "726/726 [==============================] - 0s 220us/step - loss: 13.5596\n", + "726/726 [==============================] - 0s 154us/step - loss: 13.5507\n", "Epoch 2/100\n", - "726/726 [==============================] - 0s 79us/step - loss: 9.0390\n", + "726/726 [==============================] - 0s 57us/step - loss: 9.0845\n", "Epoch 3/100\n", - "726/726 [==============================] - 0s 66us/step - loss: 5.5038\n", + "726/726 [==============================] - 0s 53us/step - loss: 5.5874\n", "Epoch 4/100\n", - "726/726 [==============================] - 0s 194us/step - loss: 3.4147\n", + "726/726 [==============================] - 0s 66us/step - loss: 3.4597\n", "Epoch 5/100\n", - "726/726 [==============================] - 0s 113us/step - loss: 2.5435\n", + "726/726 [==============================] - 0s 51us/step - loss: 2.5885\n", "Epoch 6/100\n", - "726/726 [==============================] - 0s 53us/step - loss: 2.2349\n", + "726/726 [==============================] - 0s 57us/step - loss: 2.2735\n", "Epoch 7/100\n", - "726/726 [==============================] - 0s 48us/step - loss: 2.0583\n", + "726/726 [==============================] - 0s 50us/step - loss: 2.0983\n", "Epoch 8/100\n", - "726/726 [==============================] - 0s 46us/step - loss: 1.9152\n", + "726/726 [==============================] - 0s 59us/step - loss: 1.9587\n", "Epoch 9/100\n", - "726/726 [==============================] - 0s 50us/step - loss: 1.7944\n", + "726/726 [==============================] - 0s 51us/step - loss: 1.8397\n", "Epoch 10/100\n", - "726/726 [==============================] - 0s 47us/step - loss: 1.6910\n", + "726/726 [==============================] - 0s 60us/step - loss: 1.7368\n", "Epoch 11/100\n", - "726/726 [==============================] - 0s 52us/step - loss: 1.6004\n", + "726/726 [==============================] - 0s 52us/step - loss: 1.6473\n", "Epoch 12/100\n", - "726/726 [==============================] - 0s 51us/step - loss: 1.5244\n", + "726/726 [==============================] - 0s 54us/step - loss: 1.5691\n", "Epoch 13/100\n", - "726/726 [==============================] - 0s 64us/step - loss: 1.4553\n", + "726/726 [==============================] - 0s 74us/step - loss: 1.4996\n", "Epoch 14/100\n", - "726/726 [==============================] - 0s 97us/step - loss: 1.3943\n", + "726/726 [==============================] - 0s 61us/step - loss: 1.4346\n", "Epoch 15/100\n", - "726/726 [==============================] - 0s 109us/step - loss: 1.3093\n", + "726/726 [==============================] - 0s 72us/step - loss: 1.3551\n", "Epoch 16/100\n", - "726/726 [==============================] - 0s 80us/step - loss: 1.2268\n", + "726/726 [==============================] - 0s 85us/step - loss: 1.2680\n", "Epoch 17/100\n", - "726/726 [==============================] - 0s 62us/step - loss: 1.1600\n", + "726/726 [==============================] - 0s 69us/step - loss: 1.1970\n", "Epoch 18/100\n", - "726/726 [==============================] - 0s 82us/step - loss: 1.1084\n", + "726/726 [==============================] - 0s 56us/step - loss: 1.1390\n", "Epoch 19/100\n", - "726/726 [==============================] - 0s 77us/step - loss: 1.0672\n", + "726/726 [==============================] - 0s 63us/step - loss: 1.0936\n", "Epoch 20/100\n", - "726/726 [==============================] - 0s 80us/step - loss: 1.0401\n", + "726/726 [==============================] - 0s 70us/step - loss: 1.0576\n", "Epoch 21/100\n", - "726/726 [==============================] - 0s 88us/step - loss: 1.0170\n", + "726/726 [==============================] - 0s 67us/step - loss: 1.0301\n", "Epoch 22/100\n", - "726/726 [==============================] - 0s 68us/step - loss: 1.0008\n", + "726/726 [==============================] - 0s 68us/step - loss: 1.0092\n", "Epoch 23/100\n", - "726/726 [==============================] - 0s 65us/step - loss: 0.9858\n", + "726/726 [==============================] - 0s 64us/step - loss: 0.9932\n", "Epoch 24/100\n", - "726/726 [==============================] - 0s 73us/step - loss: 0.9776\n", + "726/726 [==============================] - 0s 70us/step - loss: 0.9816\n", "Epoch 25/100\n", - "726/726 [==============================] - 0s 67us/step - loss: 0.9735\n", + "726/726 [==============================] - 0s 68us/step - loss: 0.9728\n", "Epoch 26/100\n", - "726/726 [==============================] - 0s 62us/step - loss: 0.9625\n", + "726/726 [==============================] - 0s 68us/step - loss: 0.9661\n", "Epoch 27/100\n", - "726/726 [==============================] - 0s 62us/step - loss: 0.9570\n", + "726/726 [==============================] - 0s 66us/step - loss: 0.9603\n", "Epoch 28/100\n", - "726/726 [==============================] - 0s 67us/step - loss: 0.9529\n", + "726/726 [==============================] - 0s 62us/step - loss: 0.9545\n", "Epoch 29/100\n", - "726/726 [==============================] - 0s 71us/step - loss: 0.9468\n", + "726/726 [==============================] - 0s 72us/step - loss: 0.9488\n", "Epoch 30/100\n", - "726/726 [==============================] - 0s 67us/step - loss: 0.9421\n", + "726/726 [==============================] - 0s 66us/step - loss: 0.9442\n", "Epoch 31/100\n", - "726/726 [==============================] - 0s 67us/step - loss: 0.9378\n", + "726/726 [==============================] - 0s 59us/step - loss: 0.9402\n", "Epoch 32/100\n", - "726/726 [==============================] - 0s 64us/step - loss: 0.9357\n", + "726/726 [==============================] - 0s 67us/step - loss: 0.9371\n", "Epoch 33/100\n", - "726/726 [==============================] - 0s 66us/step - loss: 0.9355\n", + "726/726 [==============================] - 0s 68us/step - loss: 0.9345\n", "Epoch 34/100\n", - "726/726 [==============================] - 0s 57us/step - loss: 0.9308\n", + "726/726 [==============================] - 0s 85us/step - loss: 0.9323\n", "Epoch 35/100\n", - "726/726 [==============================] - 0s 64us/step - loss: 0.9287\n", + "726/726 [==============================] - 0s 62us/step - loss: 0.9303\n", "Epoch 36/100\n", - "726/726 [==============================] - 0s 56us/step - loss: 0.9285\n", + "726/726 [==============================] - 0s 54us/step - loss: 0.9285\n", "Epoch 37/100\n", - "726/726 [==============================] - 0s 53us/step - loss: 0.9263\n", + "726/726 [==============================] - 0s 88us/step - loss: 0.9267\n", "Epoch 38/100\n", - "726/726 [==============================] - 0s 55us/step - loss: 0.9265\n", + "726/726 [==============================] - 0s 62us/step - loss: 0.9251\n", "Epoch 39/100\n", - "726/726 [==============================] - 0s 49us/step - loss: 0.9260\n", + "726/726 [==============================] - 0s 70us/step - loss: 0.9237\n", "Epoch 40/100\n", - "726/726 [==============================] - 0s 49us/step - loss: 0.9229\n", + "726/726 [==============================] - 0s 61us/step - loss: 0.9223\n", "Epoch 41/100\n", - "726/726 [==============================] - 0s 50us/step - loss: 0.9210\n", + "726/726 [==============================] - 0s 63us/step - loss: 0.9210\n", "Epoch 42/100\n", - "726/726 [==============================] - 0s 49us/step - loss: 0.9200\n", + "726/726 [==============================] - 0s 57us/step - loss: 0.9198\n", "Epoch 43/100\n", - "726/726 [==============================] - 0s 52us/step - loss: 0.9183\n", + "726/726 [==============================] - 0s 60us/step - loss: 0.9187\n", "Epoch 44/100\n", - "726/726 [==============================] - 0s 49us/step - loss: 0.9167\n", + "726/726 [==============================] - 0s 66us/step - loss: 0.9176\n", "Epoch 45/100\n", - "726/726 [==============================] - 0s 49us/step - loss: 0.9142\n", + "726/726 [==============================] - 0s 60us/step - loss: 0.9165\n", "Epoch 46/100\n", - "726/726 [==============================] - 0s 50us/step - loss: 0.9158\n", + "726/726 [==============================] - 0s 58us/step - loss: 0.9156\n", "Epoch 47/100\n", - "726/726 [==============================] - 0s 49us/step - loss: 0.9114\n", + "726/726 [==============================] - 0s 63us/step - loss: 0.9146\n", "Epoch 48/100\n", - "726/726 [==============================] - 0s 57us/step - loss: 0.9112\n", + "726/726 [==============================] - 0s 84us/step - loss: 0.9137\n", "Epoch 49/100\n", - "726/726 [==============================] - 0s 50us/step - loss: 0.9113\n", + "726/726 [==============================] - 0s 72us/step - loss: 0.9127\n", "Epoch 50/100\n", - "726/726 [==============================] - 0s 50us/step - loss: 0.9089\n", + "726/726 [==============================] - 0s 64us/step - loss: 0.9118\n", "Epoch 51/100\n", - "726/726 [==============================] - 0s 48us/step - loss: 0.9104\n", + "726/726 [==============================] - 0s 70us/step - loss: 0.9108\n", "Epoch 52/100\n", - "726/726 [==============================] - 0s 47us/step - loss: 0.9080\n", + "726/726 [==============================] - 0s 61us/step - loss: 0.9099\n", "Epoch 53/100\n", - "726/726 [==============================] - 0s 51us/step - loss: 0.9055\n", + "726/726 [==============================] - 0s 54us/step - loss: 0.9090\n", "Epoch 54/100\n", - "726/726 [==============================] - 0s 51us/step - loss: 0.9074\n", + "726/726 [==============================] - 0s 56us/step - loss: 0.9080\n", "Epoch 55/100\n", - "726/726 [==============================] - 0s 52us/step - loss: 0.9065\n", + "726/726 [==============================] - 0s 77us/step - loss: 0.9070\n", "Epoch 56/100\n", - "726/726 [==============================] - 0s 54us/step - loss: 0.9043\n", + "726/726 [==============================] - 0s 52us/step - loss: 0.9060\n", "Epoch 57/100\n", - "726/726 [==============================] - 0s 83us/step - loss: 0.9077\n", + "726/726 [==============================] - 0s 51us/step - loss: 0.9049\n", "Epoch 58/100\n", - "726/726 [==============================] - 0s 100us/step - loss: 0.9041\n", + "726/726 [==============================] - 0s 48us/step - loss: 0.9037\n", "Epoch 59/100\n", - "726/726 [==============================] - 0s 92us/step - loss: 0.9024\n", + "726/726 [==============================] - 0s 52us/step - loss: 0.9027\n", "Epoch 60/100\n", - "726/726 [==============================] - 0s 83us/step - loss: 0.9023\n", + "726/726 [==============================] - 0s 60us/step - loss: 0.9012\n", "Epoch 61/100\n", - "726/726 [==============================] - 0s 65us/step - loss: 0.9018\n", + "726/726 [==============================] - 0s 57us/step - loss: 0.9001\n", "Epoch 62/100\n", - "726/726 [==============================] - 0s 60us/step - loss: 0.8992\n", + "726/726 [==============================] - 0s 61us/step - loss: 0.8986\n", "Epoch 63/100\n", - "726/726 [==============================] - 0s 57us/step - loss: 0.9000\n", + "726/726 [==============================] - 0s 60us/step - loss: 0.8973\n", "Epoch 64/100\n", - "726/726 [==============================] - 0s 51us/step - loss: 0.8989\n", + "726/726 [==============================] - 0s 52us/step - loss: 0.8961\n", "Epoch 65/100\n", - "726/726 [==============================] - 0s 55us/step - loss: 0.8992\n", + "726/726 [==============================] - 0s 47us/step - loss: 0.8946\n", "Epoch 66/100\n", - "726/726 [==============================] - 0s 65us/step - loss: 0.8969\n", + "726/726 [==============================] - 0s 49us/step - loss: 0.8933\n", "Epoch 67/100\n", - "726/726 [==============================] - 0s 66us/step - loss: 0.8970\n", + "726/726 [==============================] - 0s 49us/step - loss: 0.8919\n", "Epoch 68/100\n", - "726/726 [==============================] - 0s 63us/step - loss: 0.8970\n", + "726/726 [==============================] - 0s 53us/step - loss: 0.8907\n", "Epoch 69/100\n", - "726/726 [==============================] - 0s 66us/step - loss: 0.8962\n", + "726/726 [==============================] - 0s 62us/step - loss: 0.8896\n", "Epoch 70/100\n", - "726/726 [==============================] - 0s 70us/step - loss: 0.8970\n", + "726/726 [==============================] - 0s 59us/step - loss: 0.8885\n", "Epoch 71/100\n", - "726/726 [==============================] - 0s 64us/step - loss: 0.8963\n", + "726/726 [==============================] - 0s 61us/step - loss: 0.8875\n", "Epoch 72/100\n", - "726/726 [==============================] - 0s 58us/step - loss: 0.8957\n", + "726/726 [==============================] - 0s 79us/step - loss: 0.8864\n", "Epoch 73/100\n", - "726/726 [==============================] - 0s 59us/step - loss: 0.8908\n", + "726/726 [==============================] - 0s 62us/step - loss: 0.8854\n", "Epoch 74/100\n", - "726/726 [==============================] - 0s 70us/step - loss: 0.8930\n", + "726/726 [==============================] - 0s 53us/step - loss: 0.8844\n", "Epoch 75/100\n", - "726/726 [==============================] - 0s 81us/step - loss: 0.8906\n", + "726/726 [==============================] - 0s 54us/step - loss: 0.8835\n", "Epoch 76/100\n", - "726/726 [==============================] - 0s 71us/step - loss: 0.8898\n", + "726/726 [==============================] - 0s 57us/step - loss: 0.8825\n", "Epoch 77/100\n", - "726/726 [==============================] - 0s 56us/step - loss: 0.8901\n", + "726/726 [==============================] - 0s 52us/step - loss: 0.8816\n", "Epoch 78/100\n", - "726/726 [==============================] - 0s 57us/step - loss: 0.8900\n", + "726/726 [==============================] - 0s 55us/step - loss: 0.8807\n", "Epoch 79/100\n", - "726/726 [==============================] - 0s 54us/step - loss: 0.8890\n", + "726/726 [==============================] - 0s 67us/step - loss: 0.8800\n", "Epoch 80/100\n", - "726/726 [==============================] - 0s 63us/step - loss: 0.8876\n", + "726/726 [==============================] - 0s 52us/step - loss: 0.8791\n", "Epoch 81/100\n", - "726/726 [==============================] - 0s 55us/step - loss: 0.8868\n", + "726/726 [==============================] - 0s 51us/step - loss: 0.8783\n", "Epoch 82/100\n", - "726/726 [==============================] - 0s 51us/step - loss: 0.8880\n", + "726/726 [==============================] - 0s 48us/step - loss: 0.8775\n", "Epoch 83/100\n", - "726/726 [==============================] - 0s 55us/step - loss: 0.8862\n", + "726/726 [==============================] - 0s 53us/step - loss: 0.8768\n", "Epoch 84/100\n", - "726/726 [==============================] - 0s 51us/step - loss: 0.8891\n", + "726/726 [==============================] - 0s 53us/step - loss: 0.8760\n", "Epoch 85/100\n", - "726/726 [==============================] - 0s 53us/step - loss: 0.8862\n", + "726/726 [==============================] - 0s 49us/step - loss: 0.8752\n", "Epoch 86/100\n", - "726/726 [==============================] - 0s 55us/step - loss: 0.8825\n", + "726/726 [==============================] - 0s 53us/step - loss: 0.8746\n", "Epoch 87/100\n", - "726/726 [==============================] - 0s 49us/step - loss: 0.8845\n", + "726/726 [==============================] - 0s 55us/step - loss: 0.8738\n", "Epoch 88/100\n", - "726/726 [==============================] - 0s 49us/step - loss: 0.8858\n", + "726/726 [==============================] - 0s 51us/step - loss: 0.8731\n", "Epoch 89/100\n", - "726/726 [==============================] - 0s 46us/step - loss: 0.8838\n", + "726/726 [==============================] - 0s 51us/step - loss: 0.8724\n", "Epoch 90/100\n", - "726/726 [==============================] - 0s 47us/step - loss: 0.8830\n", + "726/726 [==============================] - 0s 49us/step - loss: 0.8716\n", "Epoch 91/100\n", - "726/726 [==============================] - 0s 45us/step - loss: 0.8819\n", + "726/726 [==============================] - 0s 59us/step - loss: 0.8710\n", "Epoch 92/100\n", - "726/726 [==============================] - 0s 44us/step - loss: 0.8803\n", + "726/726 [==============================] - 0s 55us/step - loss: 0.8700\n", "Epoch 93/100\n", - "726/726 [==============================] - 0s 47us/step - loss: 0.8814\n", + "726/726 [==============================] - 0s 48us/step - loss: 0.8695\n", "Epoch 94/100\n", - "726/726 [==============================] - 0s 52us/step - loss: 0.8825\n", + "726/726 [==============================] - 0s 49us/step - loss: 0.8688\n", "Epoch 95/100\n", - "726/726 [==============================] - 0s 63us/step - loss: 0.8793\n", + "726/726 [==============================] - 0s 49us/step - loss: 0.8680\n", "Epoch 96/100\n", - "726/726 [==============================] - 0s 63us/step - loss: 0.8791\n", + "726/726 [==============================] - 0s 57us/step - loss: 0.8674\n", "Epoch 97/100\n", - "726/726 [==============================] - 0s 58us/step - loss: 0.8804\n", + "726/726 [==============================] - 0s 55us/step - loss: 0.8668\n", "Epoch 98/100\n" ] }, @@ -296,198 +296,198 @@ "name": "stdout", "output_type": "stream", "text": [ - "726/726 [==============================] - 0s 62us/step - loss: 0.8790\n", + "726/726 [==============================] - 0s 58us/step - loss: 0.8660\n", "Epoch 99/100\n", - "726/726 [==============================] - 0s 64us/step - loss: 0.8791\n", + "726/726 [==============================] - 0s 55us/step - loss: 0.8654\n", "Epoch 100/100\n", - "726/726 [==============================] - 0s 66us/step - loss: 0.8764\n", - "182/182 [==============================] - 0s 251us/step\n", + "726/726 [==============================] - 0s 52us/step - loss: 0.8648\n", + "182/182 [==============================] - 0s 99us/step\n", "Epoch 1/100\n", - "726/726 [==============================] - 0s 212us/step - loss: 13.6690\n", + "726/726 [==============================] - 0s 154us/step - loss: 4.5522\n", "Epoch 2/100\n", - "726/726 [==============================] - ETA: 0s - loss: 10.31 - 0s 61us/step - loss: 10.0055\n", + "726/726 [==============================] - 0s 49us/step - loss: 2.9388\n", "Epoch 3/100\n", - "726/726 [==============================] - 0s 59us/step - loss: 5.9825\n", + "726/726 [==============================] - 0s 49us/step - loss: 2.3166\n", "Epoch 4/100\n", - "726/726 [==============================] - 0s 60us/step - loss: 2.9531\n", + "726/726 [==============================] - 0s 49us/step - loss: 2.1064\n", "Epoch 5/100\n", - "726/726 [==============================] - 0s 51us/step - loss: 1.6202\n", + "726/726 [==============================] - 0s 49us/step - loss: 1.9919\n", "Epoch 6/100\n", - "726/726 [==============================] - 0s 50us/step - loss: 1.3303\n", + "726/726 [==============================] - 0s 48us/step - loss: 1.9075\n", "Epoch 7/100\n", - "726/726 [==============================] - 0s 53us/step - loss: 1.2658\n", + "726/726 [==============================] - 0s 47us/step - loss: 1.8352\n", "Epoch 8/100\n", - "726/726 [==============================] - 0s 55us/step - loss: 1.2286\n", + "726/726 [==============================] - 0s 52us/step - loss: 1.7665\n", "Epoch 9/100\n", - "726/726 [==============================] - 0s 54us/step - loss: 1.1985\n", + "726/726 [==============================] - 0s 48us/step - loss: 1.6571\n", "Epoch 10/100\n", - "726/726 [==============================] - 0s 64us/step - loss: 1.1737\n", + "726/726 [==============================] - 0s 49us/step - loss: 1.4782\n", "Epoch 11/100\n", - "726/726 [==============================] - 0s 51us/step - loss: 1.1504\n", + "726/726 [==============================] - 0s 53us/step - loss: 1.3579\n", "Epoch 12/100\n", - "726/726 [==============================] - 0s 49us/step - loss: 1.1337\n", + "726/726 [==============================] - 0s 52us/step - loss: 1.2821\n", "Epoch 13/100\n", - "726/726 [==============================] - ETA: 0s - loss: 1.385 - 0s 51us/step - loss: 1.1154\n", + "726/726 [==============================] - 0s 50us/step - loss: 1.2206\n", "Epoch 14/100\n", - "726/726 [==============================] - 0s 51us/step - loss: 1.1029\n", + "726/726 [==============================] - 0s 51us/step - loss: 1.1695\n", "Epoch 15/100\n", - "726/726 [==============================] - 0s 57us/step - loss: 1.0904\n", + "726/726 [==============================] - 0s 50us/step - loss: 1.1261\n", "Epoch 16/100\n", - "726/726 [==============================] - 0s 52us/step - loss: 1.0790\n", + "726/726 [==============================] - 0s 58us/step - loss: 1.0899\n", "Epoch 17/100\n", - "726/726 [==============================] - 0s 62us/step - loss: 1.0694\n", + "726/726 [==============================] - 0s 53us/step - loss: 1.0599\n", "Epoch 18/100\n", - "726/726 [==============================] - 0s 69us/step - loss: 1.0600\n", + "726/726 [==============================] - 0s 69us/step - loss: 1.0340\n", "Epoch 19/100\n", - "726/726 [==============================] - 0s 54us/step - loss: 1.0547\n", + "726/726 [==============================] - 0s 56us/step - loss: 1.0128\n", "Epoch 20/100\n", - "726/726 [==============================] - 0s 56us/step - loss: 1.0456\n", + "726/726 [==============================] - 0s 57us/step - loss: 0.9953\n", "Epoch 21/100\n", - "726/726 [==============================] - 0s 50us/step - loss: 1.0398\n", + "726/726 [==============================] - 0s 55us/step - loss: 0.9805\n", "Epoch 22/100\n", - "726/726 [==============================] - 0s 49us/step - loss: 1.0321\n", + "726/726 [==============================] - 0s 48us/step - loss: 0.9675\n", "Epoch 23/100\n", - "726/726 [==============================] - 0s 50us/step - loss: 1.0272\n", + "726/726 [==============================] - 0s 56us/step - loss: 0.9557\n", "Epoch 24/100\n", - "726/726 [==============================] - 0s 51us/step - loss: 1.0209\n", + "726/726 [==============================] - 0s 73us/step - loss: 0.9451\n", "Epoch 25/100\n", - "726/726 [==============================] - 0s 52us/step - loss: 1.0171\n", + "726/726 [==============================] - 0s 52us/step - loss: 0.9360\n", "Epoch 26/100\n", - "726/726 [==============================] - 0s 56us/step - loss: 1.0134\n", + "726/726 [==============================] - 0s 53us/step - loss: 0.9289\n", "Epoch 27/100\n", - "726/726 [==============================] - 0s 59us/step - loss: 1.0053\n", + "726/726 [==============================] - 0s 60us/step - loss: 0.9224\n", "Epoch 28/100\n", - "726/726 [==============================] - 0s 53us/step - loss: 1.0021\n", + "726/726 [==============================] - 0s 51us/step - loss: 0.9170\n", "Epoch 29/100\n", - "726/726 [==============================] - 0s 56us/step - loss: 0.9966\n", + "726/726 [==============================] - 0s 53us/step - loss: 0.9125\n", "Epoch 30/100\n", - "726/726 [==============================] - 0s 59us/step - loss: 0.9924\n", + "726/726 [==============================] - 0s 50us/step - loss: 0.9080\n", "Epoch 31/100\n", - "726/726 [==============================] - 0s 62us/step - loss: 0.9870\n", + "726/726 [==============================] - 0s 52us/step - loss: 0.9043\n", "Epoch 32/100\n", - "726/726 [==============================] - 0s 54us/step - loss: 0.9826\n", + "726/726 [==============================] - 0s 54us/step - loss: 0.9002\n", "Epoch 33/100\n", - "726/726 [==============================] - 0s 58us/step - loss: 0.9789\n", + "726/726 [==============================] - 0s 48us/step - loss: 0.8969\n", "Epoch 34/100\n", - "726/726 [==============================] - 0s 63us/step - loss: 0.9762\n", + "726/726 [==============================] - 0s 52us/step - loss: 0.8931\n", "Epoch 35/100\n", - "726/726 [==============================] - 0s 61us/step - loss: 0.9710\n", + "726/726 [==============================] - 0s 60us/step - loss: 0.8901\n", "Epoch 36/100\n", - "726/726 [==============================] - 0s 57us/step - loss: 0.9675\n", + "726/726 [==============================] - 0s 55us/step - loss: 0.8866\n", "Epoch 37/100\n", - "726/726 [==============================] - 0s 58us/step - loss: 0.9628\n", + "726/726 [==============================] - 0s 50us/step - loss: 0.8835\n", "Epoch 38/100\n", - "726/726 [==============================] - 0s 57us/step - loss: 0.9594\n", + "726/726 [==============================] - 0s 56us/step - loss: 0.8798\n", "Epoch 39/100\n", - "726/726 [==============================] - 0s 60us/step - loss: 0.9586\n", + "726/726 [==============================] - 0s 53us/step - loss: 0.8765\n", "Epoch 40/100\n", - "726/726 [==============================] - 0s 65us/step - loss: 0.9588\n", + "726/726 [==============================] - 0s 53us/step - loss: 0.8725\n", "Epoch 41/100\n", - "726/726 [==============================] - 0s 98us/step - loss: 0.9530\n", + "726/726 [==============================] - 0s 51us/step - loss: 0.8690\n", "Epoch 42/100\n", - "726/726 [==============================] - 0s 54us/step - loss: 0.9477\n", + "726/726 [==============================] - 0s 59us/step - loss: 0.8658\n", "Epoch 43/100\n", - "726/726 [==============================] - 0s 54us/step - loss: 0.9444\n", + "726/726 [==============================] - 0s 177us/step - loss: 0.8633\n", "Epoch 44/100\n", - "726/726 [==============================] - 0s 55us/step - loss: 0.9409\n", + "726/726 [==============================] - 0s 69us/step - loss: 0.8604\n", "Epoch 45/100\n", - "726/726 [==============================] - 0s 84us/step - loss: 0.9380\n", + "726/726 [==============================] - 0s 55us/step - loss: 0.8581\n", "Epoch 46/100\n", - "726/726 [==============================] - 0s 88us/step - loss: 0.9383\n", + "726/726 [==============================] - 0s 58us/step - loss: 0.8554\n", "Epoch 47/100\n", - "726/726 [==============================] - 0s 59us/step - loss: 0.9356\n", + "726/726 [==============================] - 0s 51us/step - loss: 0.8530\n", "Epoch 48/100\n", - "726/726 [==============================] - 0s 59us/step - loss: 0.9297\n", + "726/726 [==============================] - 0s 51us/step - loss: 0.8504\n", "Epoch 49/100\n", - "726/726 [==============================] - 0s 56us/step - loss: 0.9281\n", + "726/726 [==============================] - 0s 65us/step - loss: 0.8476\n", "Epoch 50/100\n", - "726/726 [==============================] - 0s 58us/step - loss: 0.9242\n", + "726/726 [==============================] - 0s 50us/step - loss: 0.8454\n", "Epoch 51/100\n", - "726/726 [==============================] - 0s 61us/step - loss: 0.9220\n", + "726/726 [==============================] - 0s 53us/step - loss: 0.8428\n", "Epoch 52/100\n", - "726/726 [==============================] - 0s 59us/step - loss: 0.9211\n", + "726/726 [==============================] - 0s 57us/step - loss: 0.8406\n", "Epoch 53/100\n", - "726/726 [==============================] - 0s 68us/step - loss: 0.9171\n", + "726/726 [==============================] - 0s 56us/step - loss: 0.8383\n", "Epoch 54/100\n", - "726/726 [==============================] - 0s 58us/step - loss: 0.9147\n", + "726/726 [==============================] - 0s 53us/step - loss: 0.8357\n", "Epoch 55/100\n", - "726/726 [==============================] - 0s 58us/step - loss: 0.9121\n", + "726/726 [==============================] - 0s 56us/step - loss: 0.8335\n", "Epoch 56/100\n", - "726/726 [==============================] - 0s 56us/step - loss: 0.9119\n", + "726/726 [==============================] - ETA: 0s - loss: 0.541 - 0s 52us/step - loss: 0.8311\n", "Epoch 57/100\n", - "726/726 [==============================] - 0s 53us/step - loss: 0.9096\n", + "726/726 [==============================] - 0s 68us/step - loss: 0.8284\n", "Epoch 58/100\n", - "726/726 [==============================] - 0s 56us/step - loss: 0.9054\n", + "726/726 [==============================] - 0s 53us/step - loss: 0.8260\n", "Epoch 59/100\n", - "726/726 [==============================] - 0s 55us/step - loss: 0.9071\n", + "726/726 [==============================] - 0s 56us/step - loss: 0.8237\n", "Epoch 60/100\n", - "726/726 [==============================] - 0s 62us/step - loss: 0.9002\n", + "726/726 [==============================] - 0s 57us/step - loss: 0.8210\n", "Epoch 61/100\n", - "726/726 [==============================] - 0s 59us/step - loss: 0.8979\n", + "726/726 [==============================] - 0s 51us/step - loss: 0.8192\n", "Epoch 62/100\n", - "726/726 [==============================] - 0s 74us/step - loss: 0.8962\n", + "726/726 [==============================] - 0s 50us/step - loss: 0.8173\n", "Epoch 63/100\n", - "726/726 [==============================] - 0s 72us/step - loss: 0.8947\n", + "726/726 [==============================] - 0s 52us/step - loss: 0.8154\n", "Epoch 64/100\n", - "726/726 [==============================] - 0s 81us/step - loss: 0.8934\n", + "726/726 [==============================] - 0s 67us/step - loss: 0.8133\n", "Epoch 65/100\n", - "726/726 [==============================] - 0s 63us/step - loss: 0.8938\n", + "726/726 [==============================] - 0s 61us/step - loss: 0.8116\n", "Epoch 66/100\n", - "726/726 [==============================] - 0s 55us/step - loss: 0.8937\n", + "726/726 [==============================] - 0s 60us/step - loss: 0.8099\n", "Epoch 67/100\n", - "726/726 [==============================] - 0s 56us/step - loss: 0.8838\n", + "726/726 [==============================] - 0s 58us/step - loss: 0.8079\n", "Epoch 68/100\n", - "726/726 [==============================] - 0s 60us/step - loss: 0.8835\n", + "726/726 [==============================] - 0s 58us/step - loss: 0.8065\n", "Epoch 69/100\n", - "726/726 [==============================] - 0s 53us/step - loss: 0.8799\n", + "726/726 [==============================] - 0s 61us/step - loss: 0.8047\n", "Epoch 70/100\n", - "726/726 [==============================] - 0s 52us/step - loss: 0.8800\n", + "726/726 [==============================] - 0s 61us/step - loss: 0.8036\n", "Epoch 71/100\n", - "726/726 [==============================] - 0s 54us/step - loss: 0.8886\n", + "726/726 [==============================] - 0s 63us/step - loss: 0.8019\n", "Epoch 72/100\n", - "726/726 [==============================] - 0s 53us/step - loss: 0.8763\n", + "726/726 [==============================] - 0s 70us/step - loss: 0.8008\n", "Epoch 73/100\n", - "726/726 [==============================] - 0s 51us/step - loss: 0.8749\n", + "726/726 [==============================] - 0s 60us/step - loss: 0.7993\n", "Epoch 74/100\n", - "726/726 [==============================] - 0s 51us/step - loss: 0.8721\n", + "726/726 [==============================] - 0s 58us/step - loss: 0.7982\n", "Epoch 75/100\n", - "726/726 [==============================] - 0s 54us/step - loss: 0.8687\n", + "726/726 [==============================] - 0s 83us/step - loss: 0.7968\n", "Epoch 76/100\n", - "726/726 [==============================] - 0s 51us/step - loss: 0.8680\n", + "726/726 [==============================] - 0s 52us/step - loss: 0.7954\n", "Epoch 77/100\n", - "726/726 [==============================] - 0s 51us/step - loss: 0.8669\n", + "726/726 [==============================] - 0s 68us/step - loss: 0.7946\n", "Epoch 78/100\n", - "726/726 [==============================] - 0s 56us/step - loss: 0.8652\n", + "726/726 [==============================] - 0s 65us/step - loss: 0.7932\n", "Epoch 79/100\n", - "726/726 [==============================] - 0s 56us/step - loss: 0.8627\n", + "726/726 [==============================] - 0s 66us/step - loss: 0.7920\n", "Epoch 80/100\n", - "726/726 [==============================] - 0s 53us/step - loss: 0.8594\n", + "726/726 [==============================] - 0s 67us/step - loss: 0.7911\n", "Epoch 81/100\n", - "726/726 [==============================] - 0s 56us/step - loss: 0.8592\n", + "726/726 [==============================] - 0s 65us/step - loss: 0.7900\n", "Epoch 82/100\n", - "726/726 [==============================] - 0s 61us/step - loss: 0.8577\n", + "726/726 [==============================] - 0s 71us/step - loss: 0.7889\n", "Epoch 83/100\n", - "726/726 [==============================] - 0s 54us/step - loss: 0.8574\n", + "726/726 [==============================] - 0s 70us/step - loss: 0.7876\n", "Epoch 84/100\n", - "726/726 [==============================] - 0s 51us/step - loss: 0.8524\n", + "726/726 [==============================] - 0s 68us/step - loss: 0.7867\n", "Epoch 85/100\n", - "726/726 [==============================] - 0s 49us/step - loss: 0.8532\n", + "726/726 [==============================] - 0s 53us/step - loss: 0.7854\n", "Epoch 86/100\n", - "726/726 [==============================] - 0s 51us/step - loss: 0.8548\n", + "726/726 [==============================] - 0s 60us/step - loss: 0.7846\n", "Epoch 87/100\n", - "726/726 [==============================] - 0s 50us/step - loss: 0.8513\n", + "726/726 [==============================] - 0s 56us/step - loss: 0.7833\n", "Epoch 88/100\n", - "726/726 [==============================] - 0s 65us/step - loss: 0.8571\n", + "726/726 [==============================] - 0s 50us/step - loss: 0.7824\n", "Epoch 89/100\n", - "726/726 [==============================] - 0s 62us/step - loss: 0.8491\n", + "726/726 [==============================] - 0s 65us/step - loss: 0.7811\n", "Epoch 90/100\n", - "726/726 [==============================] - 0s 61us/step - loss: 0.8508\n", + "726/726 [==============================] - 0s 59us/step - loss: 0.7803\n", "Epoch 91/100\n", - "726/726 [==============================] - 0s 58us/step - loss: 0.8499\n", + "726/726 [==============================] - 0s 62us/step - loss: 0.7791\n", "Epoch 92/100\n", - "726/726 [==============================] - 0s 60us/step - loss: 0.8435\n", + "726/726 [==============================] - 0s 61us/step - loss: 0.7784\n", "Epoch 93/100\n", - "726/726 [==============================] - 0s 57us/step - loss: 0.8455\n", + "726/726 [==============================] - 0s 62us/step - loss: 0.7772\n", "Epoch 94/100\n" ] }, @@ -495,399 +495,399 @@ "name": "stdout", "output_type": "stream", "text": [ - "726/726 [==============================] - 0s 55us/step - loss: 0.8409\n", + "726/726 [==============================] - 0s 51us/step - loss: 0.7764\n", "Epoch 95/100\n", - "726/726 [==============================] - 0s 57us/step - loss: 0.8400\n", + "726/726 [==============================] - 0s 61us/step - loss: 0.7752\n", "Epoch 96/100\n", - "726/726 [==============================] - 0s 53us/step - loss: 0.8385\n", + "726/726 [==============================] - 0s 53us/step - loss: 0.7746\n", "Epoch 97/100\n", - "726/726 [==============================] - 0s 56us/step - loss: 0.8376\n", + "726/726 [==============================] - 0s 58us/step - loss: 0.7734\n", "Epoch 98/100\n", - "726/726 [==============================] - 0s 54us/step - loss: 0.8378\n", + "726/726 [==============================] - 0s 55us/step - loss: 0.7728\n", "Epoch 99/100\n", - "726/726 [==============================] - 0s 58us/step - loss: 0.8370\n", + "726/726 [==============================] - 0s 60us/step - loss: 0.7717\n", "Epoch 100/100\n", - "726/726 [==============================] - 0s 55us/step - loss: 0.8372\n", - "182/182 [==============================] - 0s 117us/step\n", + "726/726 [==============================] - 0s 55us/step - loss: 0.7711\n", + "182/182 [==============================] - 0s 91us/step\n", "Epoch 1/100\n", - "726/726 [==============================] - 0s 211us/step - loss: 9.4218\n", + "726/726 [==============================] - 0s 162us/step - loss: 16.3760\n", "Epoch 2/100\n", - "726/726 [==============================] - 0s 64us/step - loss: 5.5441\n", + "726/726 [==============================] - 0s 49us/step - loss: 11.0709\n", "Epoch 3/100\n", - "726/726 [==============================] - 0s 56us/step - loss: 3.4082\n", + "726/726 [==============================] - 0s 55us/step - loss: 6.3973\n", "Epoch 4/100\n", - "726/726 [==============================] - 0s 57us/step - loss: 2.4394\n", + "726/726 [==============================] - 0s 55us/step - loss: 3.2553\n", "Epoch 5/100\n", - "726/726 [==============================] - 0s 57us/step - loss: 2.0473\n", + "726/726 [==============================] - 0s 52us/step - loss: 1.8951\n", "Epoch 6/100\n", - "726/726 [==============================] - 0s 52us/step - loss: 1.8340\n", + "726/726 [==============================] - 0s 53us/step - loss: 1.5231\n", "Epoch 7/100\n", - "726/726 [==============================] - 0s 58us/step - loss: 1.6820\n", + "726/726 [==============================] - 0s 57us/step - loss: 1.4299\n", "Epoch 8/100\n", - "726/726 [==============================] - 0s 60us/step - loss: 1.5621\n", + "726/726 [==============================] - 0s 50us/step - loss: 1.3841\n", "Epoch 9/100\n", - "726/726 [==============================] - 0s 58us/step - loss: 1.4708\n", + "726/726 [==============================] - 0s 54us/step - loss: 1.3459\n", "Epoch 10/100\n", - "726/726 [==============================] - 0s 67us/step - loss: 1.3883\n", + "726/726 [==============================] - 0s 50us/step - loss: 1.3106\n", "Epoch 11/100\n", - "726/726 [==============================] - 0s 70us/step - loss: 1.3259\n", + "726/726 [==============================] - 0s 50us/step - loss: 1.2778\n", "Epoch 12/100\n", - "726/726 [==============================] - 0s 58us/step - loss: 1.2768\n", + "726/726 [==============================] - 0s 64us/step - loss: 1.2478\n", "Epoch 13/100\n", - "726/726 [==============================] - 0s 58us/step - loss: 1.2359\n", + "726/726 [==============================] - 0s 52us/step - loss: 1.2203\n", "Epoch 14/100\n", - "726/726 [==============================] - 0s 51us/step - loss: 1.2005\n", + "726/726 [==============================] - 0s 50us/step - loss: 1.1945\n", "Epoch 15/100\n", - "726/726 [==============================] - 0s 52us/step - loss: 1.1710\n", + "726/726 [==============================] - 0s 48us/step - loss: 1.1707\n", "Epoch 16/100\n", - "726/726 [==============================] - 0s 51us/step - loss: 1.1502\n", + "726/726 [==============================] - 0s 51us/step - loss: 1.1478\n", "Epoch 17/100\n", - "726/726 [==============================] - 0s 50us/step - loss: 1.1265\n", + "726/726 [==============================] - 0s 52us/step - loss: 1.1259\n", "Epoch 18/100\n", - "726/726 [==============================] - 0s 144us/step - loss: 1.1078\n", + "726/726 [==============================] - 0s 53us/step - loss: 1.1051\n", "Epoch 19/100\n", - "726/726 [==============================] - 0s 61us/step - loss: 1.0913\n", + "726/726 [==============================] - 0s 54us/step - loss: 1.0871\n", "Epoch 20/100\n", - "726/726 [==============================] - 0s 51us/step - loss: 1.0754\n", + "726/726 [==============================] - 0s 77us/step - loss: 1.0711\n", "Epoch 21/100\n", - "726/726 [==============================] - 0s 53us/step - loss: 1.0616\n", + "726/726 [==============================] - 0s 59us/step - loss: 1.0553\n", "Epoch 22/100\n", - "726/726 [==============================] - 0s 49us/step - loss: 1.0485\n", + "726/726 [==============================] - 0s 55us/step - loss: 1.0394\n", "Epoch 23/100\n", - "726/726 [==============================] - 0s 53us/step - loss: 1.0393\n", + "726/726 [==============================] - 0s 55us/step - loss: 1.0249\n", "Epoch 24/100\n", - "726/726 [==============================] - 0s 54us/step - loss: 1.0251\n", + "726/726 [==============================] - 0s 66us/step - loss: 1.0131\n", "Epoch 25/100\n", - "726/726 [==============================] - 0s 49us/step - loss: 1.0149\n", + "726/726 [==============================] - 0s 52us/step - loss: 1.0028\n", "Epoch 26/100\n", - "726/726 [==============================] - 0s 54us/step - loss: 1.0037\n", + "726/726 [==============================] - 0s 54us/step - loss: 0.9945\n", "Epoch 27/100\n", - "726/726 [==============================] - 0s 60us/step - loss: 0.9932\n", + "726/726 [==============================] - 0s 50us/step - loss: 0.9869\n", "Epoch 28/100\n", - "726/726 [==============================] - 0s 283us/step - loss: 0.9834\n", + "726/726 [==============================] - 0s 57us/step - loss: 0.9802\n", "Epoch 29/100\n", - "726/726 [==============================] - 0s 116us/step - loss: 0.9779\n", + "726/726 [==============================] - 0s 51us/step - loss: 0.9742\n", "Epoch 30/100\n", - "726/726 [==============================] - 0s 54us/step - loss: 0.9649\n", + "726/726 [==============================] - 0s 51us/step - loss: 0.9685\n", "Epoch 31/100\n", - "726/726 [==============================] - 0s 73us/step - loss: 0.9583\n", + "726/726 [==============================] - 0s 52us/step - loss: 0.9633\n", "Epoch 32/100\n", - "726/726 [==============================] - 0s 63us/step - loss: 0.9553\n", + "726/726 [==============================] - 0s 55us/step - loss: 0.9584\n", "Epoch 33/100\n", - "726/726 [==============================] - 0s 69us/step - loss: 0.9406\n", + "726/726 [==============================] - 0s 63us/step - loss: 0.9540\n", "Epoch 34/100\n", - "726/726 [==============================] - 0s 69us/step - loss: 0.9429\n", + "726/726 [==============================] - 0s 53us/step - loss: 0.9500\n", "Epoch 35/100\n", - "726/726 [==============================] - 0s 71us/step - loss: 0.9280\n", + "726/726 [==============================] - 0s 52us/step - loss: 0.9460\n", "Epoch 36/100\n", - "726/726 [==============================] - 0s 69us/step - loss: 0.9243\n", + "726/726 [==============================] - 0s 57us/step - loss: 0.9419\n", "Epoch 37/100\n", - "726/726 [==============================] - 0s 72us/step - loss: 0.9144\n", + "726/726 [==============================] - 0s 53us/step - loss: 0.9384\n", "Epoch 38/100\n", - "726/726 [==============================] - 0s 68us/step - loss: 0.9094\n", + "726/726 [==============================] - 0s 50us/step - loss: 0.9349\n", "Epoch 39/100\n", - "726/726 [==============================] - 0s 67us/step - loss: 0.9038\n", + "726/726 [==============================] - 0s 52us/step - loss: 0.9313\n", "Epoch 40/100\n", - "726/726 [==============================] - 0s 70us/step - loss: 0.8984\n", + "726/726 [==============================] - 0s 67us/step - loss: 0.9282\n", "Epoch 41/100\n", - "726/726 [==============================] - 0s 63us/step - loss: 0.8948\n", + "726/726 [==============================] - 0s 56us/step - loss: 0.9250\n", "Epoch 42/100\n", - "726/726 [==============================] - 0s 60us/step - loss: 0.8949\n", + "726/726 [==============================] - 0s 56us/step - loss: 0.9219\n", "Epoch 43/100\n", - "726/726 [==============================] - 0s 65us/step - loss: 0.8849\n", + "726/726 [==============================] - 0s 51us/step - loss: 0.9192\n", "Epoch 44/100\n", - "726/726 [==============================] - 0s 71us/step - loss: 0.8769\n", + "726/726 [==============================] - 0s 60us/step - loss: 0.9164\n", "Epoch 45/100\n", - "726/726 [==============================] - 0s 72us/step - loss: 0.8727\n", + "726/726 [==============================] - 0s 55us/step - loss: 0.9136\n", "Epoch 46/100\n", - "726/726 [==============================] - 0s 77us/step - loss: 0.8710\n", + "726/726 [==============================] - 0s 56us/step - loss: 0.9110\n", "Epoch 47/100\n", - "726/726 [==============================] - 0s 55us/step - loss: 0.8675\n", + "726/726 [==============================] - 0s 65us/step - loss: 0.9084\n", "Epoch 48/100\n", - "726/726 [==============================] - 0s 53us/step - loss: 0.8668\n", + "726/726 [==============================] - 0s 78us/step - loss: 0.9060\n", "Epoch 49/100\n", - "726/726 [==============================] - 0s 61us/step - loss: 0.8604\n", + "726/726 [==============================] - 0s 57us/step - loss: 0.9036\n", "Epoch 50/100\n", - "726/726 [==============================] - 0s 57us/step - loss: 0.8552\n", + "726/726 [==============================] - 0s 55us/step - loss: 0.9012\n", "Epoch 51/100\n", - "726/726 [==============================] - 0s 56us/step - loss: 0.8502\n", + "726/726 [==============================] - 0s 53us/step - loss: 0.8990\n", "Epoch 52/100\n", - "726/726 [==============================] - 0s 56us/step - loss: 0.8589\n", + "726/726 [==============================] - 0s 55us/step - loss: 0.8968\n", "Epoch 53/100\n", - "726/726 [==============================] - 0s 58us/step - loss: 0.8450\n", + "726/726 [==============================] - 0s 51us/step - loss: 0.8946\n", "Epoch 54/100\n", - "726/726 [==============================] - 0s 53us/step - loss: 0.8464\n", + "726/726 [==============================] - 0s 53us/step - loss: 0.8927\n", "Epoch 55/100\n", - "726/726 [==============================] - 0s 48us/step - loss: 0.8384\n", + "726/726 [==============================] - 0s 51us/step - loss: 0.8906\n", "Epoch 56/100\n", - "726/726 [==============================] - 0s 50us/step - loss: 0.8404\n", + "726/726 [==============================] - 0s 57us/step - loss: 0.8887\n", "Epoch 57/100\n", - "726/726 [==============================] - 0s 51us/step - loss: 0.8405\n", + "726/726 [==============================] - 0s 48us/step - loss: 0.8868\n", "Epoch 58/100\n", - "726/726 [==============================] - 0s 51us/step - loss: 0.8342\n", + "726/726 [==============================] - 0s 53us/step - loss: 0.8851\n", "Epoch 59/100\n", - "726/726 [==============================] - 0s 55us/step - loss: 0.8362\n", + "726/726 [==============================] - 0s 52us/step - loss: 0.8832\n", "Epoch 60/100\n", - "726/726 [==============================] - 0s 54us/step - loss: 0.8326\n", + "726/726 [==============================] - 0s 49us/step - loss: 0.8815\n", "Epoch 61/100\n", - "726/726 [==============================] - 0s 53us/step - loss: 0.8247\n", + "726/726 [==============================] - 0s 61us/step - loss: 0.8796\n", "Epoch 62/100\n", - "726/726 [==============================] - 0s 58us/step - loss: 0.8254\n", + "726/726 [==============================] - 0s 53us/step - loss: 0.8781\n", "Epoch 63/100\n", - "726/726 [==============================] - 0s 53us/step - loss: 0.8252\n", + "726/726 [==============================] - 0s 56us/step - loss: 0.8765\n", "Epoch 64/100\n", - "726/726 [==============================] - 0s 53us/step - loss: 0.8196\n", + "726/726 [==============================] - 0s 62us/step - loss: 0.8749\n", "Epoch 65/100\n", - "726/726 [==============================] - 0s 55us/step - loss: 0.8180\n", + "726/726 [==============================] - 0s 50us/step - loss: 0.8735\n", "Epoch 66/100\n", - "726/726 [==============================] - 0s 56us/step - loss: 0.8190\n", + "726/726 [==============================] - 0s 50us/step - loss: 0.8720\n", "Epoch 67/100\n", - "726/726 [==============================] - 0s 64us/step - loss: 0.8172\n", + "726/726 [==============================] - 0s 59us/step - loss: 0.8705\n", "Epoch 68/100\n", - "726/726 [==============================] - 0s 100us/step - loss: 0.8132\n", + "726/726 [==============================] - 0s 63us/step - loss: 0.8693\n", "Epoch 69/100\n", - "726/726 [==============================] - 0s 88us/step - loss: 0.8125\n", + "726/726 [==============================] - 0s 57us/step - loss: 0.8678\n", "Epoch 70/100\n", - "726/726 [==============================] - 0s 77us/step - loss: 0.8146\n", + "726/726 [==============================] - 0s 55us/step - loss: 0.8666\n", "Epoch 71/100\n", - "726/726 [==============================] - 0s 79us/step - loss: 0.8119\n", + "726/726 [==============================] - 0s 58us/step - loss: 0.8653\n", "Epoch 72/100\n", - "726/726 [==============================] - 0s 61us/step - loss: 0.8169\n", + "726/726 [==============================] - 0s 53us/step - loss: 0.8640\n", "Epoch 73/100\n", - "726/726 [==============================] - 0s 67us/step - loss: 0.8097\n", + "726/726 [==============================] - 0s 83us/step - loss: 0.8627\n", "Epoch 74/100\n", - "726/726 [==============================] - 0s 65us/step - loss: 0.8060\n", + "726/726 [==============================] - 0s 51us/step - loss: 0.8617\n", "Epoch 75/100\n", - "726/726 [==============================] - ETA: 0s - loss: 0.611 - 0s 73us/step - loss: 0.8090\n", + "726/726 [==============================] - 0s 149us/step - loss: 0.8607\n", "Epoch 76/100\n", - "726/726 [==============================] - 0s 68us/step - loss: 0.8024\n", + "726/726 [==============================] - 0s 91us/step - loss: 0.8595\n", "Epoch 77/100\n", - "726/726 [==============================] - 0s 78us/step - loss: 0.8055\n", + "726/726 [==============================] - 0s 55us/step - loss: 0.8585\n", "Epoch 78/100\n", - "726/726 [==============================] - 0s 58us/step - loss: 0.7991\n", + "726/726 [==============================] - 0s 79us/step - loss: 0.8573\n", "Epoch 79/100\n", - "726/726 [==============================] - 0s 58us/step - loss: 0.8008\n", + "726/726 [==============================] - 0s 51us/step - loss: 0.8565\n", "Epoch 80/100\n", - "726/726 [==============================] - 0s 58us/step - loss: 0.8011\n", + "726/726 [==============================] - 0s 57us/step - loss: 0.8557\n", "Epoch 81/100\n", - "726/726 [==============================] - 0s 66us/step - loss: 0.8103\n", + "726/726 [==============================] - 0s 62us/step - loss: 0.8548\n", "Epoch 82/100\n", - "726/726 [==============================] - 0s 66us/step - loss: 0.7985\n", + "726/726 [==============================] - 0s 58us/step - loss: 0.8537\n", "Epoch 83/100\n", - "726/726 [==============================] - 0s 70us/step - loss: 0.7968\n", + "726/726 [==============================] - 0s 61us/step - loss: 0.8529\n", "Epoch 84/100\n", - "726/726 [==============================] - 0s 83us/step - loss: 0.7985\n", + "726/726 [==============================] - 0s 55us/step - loss: 0.8522\n", "Epoch 85/100\n", - "726/726 [==============================] - 0s 149us/step - loss: 0.7955\n", + "726/726 [==============================] - 0s 54us/step - loss: 0.8514\n", "Epoch 86/100\n", - "726/726 [==============================] - 0s 89us/step - loss: 0.8009\n", + "726/726 [==============================] - 0s 54us/step - loss: 0.8504\n", "Epoch 87/100\n", - "726/726 [==============================] - 0s 53us/step - loss: 0.7951\n", + "726/726 [==============================] - 0s 52us/step - loss: 0.8498\n", "Epoch 88/100\n", - "726/726 [==============================] - 0s 49us/step - loss: 0.7982\n", + "726/726 [==============================] - 0s 51us/step - loss: 0.8493\n", "Epoch 89/100\n", - "726/726 [==============================] - 0s 50us/step - loss: 0.7949\n", - "Epoch 90/100\n" + "726/726 [==============================] - 0s 52us/step - loss: 0.8485\n", + "Epoch 90/100\n", + "726/726 [==============================] - 0s 56us/step - loss: 0.8478\n", + "Epoch 91/100\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ - "726/726 [==============================] - 0s 56us/step - loss: 0.7922\n", - "Epoch 91/100\n", - "726/726 [==============================] - 0s 52us/step - loss: 0.7932\n", + "726/726 [==============================] - 0s 52us/step - loss: 0.8469\n", "Epoch 92/100\n", - "726/726 [==============================] - 0s 52us/step - loss: 0.7927\n", + "726/726 [==============================] - 0s 52us/step - loss: 0.8466\n", "Epoch 93/100\n", - "726/726 [==============================] - 0s 50us/step - loss: 0.7974\n", + "726/726 [==============================] - 0s 50us/step - loss: 0.8459\n", "Epoch 94/100\n", - "726/726 [==============================] - 0s 48us/step - loss: 0.7991\n", + "726/726 [==============================] - 0s 55us/step - loss: 0.8451\n", "Epoch 95/100\n", - "726/726 [==============================] - 0s 50us/step - loss: 0.7977\n", + "726/726 [==============================] - 0s 52us/step - loss: 0.8447\n", "Epoch 96/100\n", - "726/726 [==============================] - 0s 51us/step - loss: 0.7910\n", + "726/726 [==============================] - 0s 59us/step - loss: 0.8443\n", "Epoch 97/100\n", - "726/726 [==============================] - 0s 52us/step - loss: 0.7906\n", + "726/726 [==============================] - 0s 51us/step - loss: 0.8437\n", "Epoch 98/100\n", - "726/726 [==============================] - 0s 50us/step - loss: 0.7905\n", + "726/726 [==============================] - 0s 51us/step - loss: 0.8428\n", "Epoch 99/100\n", - "726/726 [==============================] - 0s 50us/step - loss: 0.7962\n", + "726/726 [==============================] - 0s 52us/step - loss: 0.8422\n", "Epoch 100/100\n", - "726/726 [==============================] - 0s 54us/step - loss: 0.7955\n", - "182/182 [==============================] - 0s 114us/step\n", + "726/726 [==============================] - 0s 51us/step - loss: 0.8417\n", + "182/182 [==============================] - 0s 92us/step\n", "Epoch 1/100\n", - "727/727 [==============================] - 0s 269us/step - loss: 6.9244\n", + "727/727 [==============================] - 0s 158us/step - loss: 21.5484\n", "Epoch 2/100\n", - "727/727 [==============================] - 0s 76us/step - loss: 4.0054\n", + "727/727 [==============================] - 0s 55us/step - loss: 14.5283\n", "Epoch 3/100\n", - "727/727 [==============================] - 0s 72us/step - loss: 2.2267\n", + "727/727 [==============================] - 0s 55us/step - loss: 9.2746\n", "Epoch 4/100\n", - "727/727 [==============================] - 0s 68us/step - loss: 1.5728\n", + "727/727 [==============================] - 0s 52us/step - loss: 5.6396\n", "Epoch 5/100\n", - "727/727 [==============================] - 0s 74us/step - loss: 1.3981\n", + "727/727 [==============================] - 0s 55us/step - loss: 3.5013\n", "Epoch 6/100\n", - "727/727 [==============================] - 0s 96us/step - loss: 1.3159\n", + "727/727 [==============================] - 0s 51us/step - loss: 2.4853\n", "Epoch 7/100\n", - "727/727 [==============================] - 0s 87us/step - loss: 1.2627\n", + "727/727 [==============================] - 0s 52us/step - loss: 2.0817\n", "Epoch 8/100\n", - "727/727 [==============================] - 0s 69us/step - loss: 1.2247\n", + "727/727 [==============================] - 0s 58us/step - loss: 1.9193\n", "Epoch 9/100\n", - "727/727 [==============================] - 0s 68us/step - loss: 1.1992\n", + "727/727 [==============================] - 0s 64us/step - loss: 1.8289\n", "Epoch 10/100\n", - "727/727 [==============================] - 0s 63us/step - loss: 1.1759\n", + "727/727 [==============================] - 0s 52us/step - loss: 1.7591\n", "Epoch 11/100\n", - "727/727 [==============================] - 0s 55us/step - loss: 1.1588\n", + "727/727 [==============================] - 0s 53us/step - loss: 1.6977\n", "Epoch 12/100\n", - "727/727 [==============================] - 0s 64us/step - loss: 1.1420\n", + "727/727 [==============================] - 0s 53us/step - loss: 1.6426\n", "Epoch 13/100\n", - "727/727 [==============================] - 0s 60us/step - loss: 1.1290\n", + "727/727 [==============================] - 0s 50us/step - loss: 1.5921\n", "Epoch 14/100\n", - "727/727 [==============================] - 0s 58us/step - loss: 1.1187\n", + "727/727 [==============================] - 0s 53us/step - loss: 1.5463\n", "Epoch 15/100\n", - "727/727 [==============================] - 0s 61us/step - loss: 1.1035\n", + "727/727 [==============================] - 0s 56us/step - loss: 1.5045\n", "Epoch 16/100\n", - "727/727 [==============================] - 0s 55us/step - loss: 1.0944\n", + "727/727 [==============================] - 0s 54us/step - loss: 1.4656\n", "Epoch 17/100\n", - "727/727 [==============================] - 0s 56us/step - loss: 1.0863\n", + "727/727 [==============================] - 0s 55us/step - loss: 1.4295\n", "Epoch 18/100\n", - "727/727 [==============================] - 0s 46us/step - loss: 1.0753\n", + "727/727 [==============================] - 0s 55us/step - loss: 1.3964\n", "Epoch 19/100\n", - "727/727 [==============================] - 0s 50us/step - loss: 1.0677\n", + "727/727 [==============================] - 0s 58us/step - loss: 1.3652\n", "Epoch 20/100\n", - "727/727 [==============================] - 0s 51us/step - loss: 1.0585\n", + "727/727 [==============================] - 0s 53us/step - loss: 1.3361\n", "Epoch 21/100\n", - "727/727 [==============================] - 0s 52us/step - loss: 1.0502\n", + "727/727 [==============================] - 0s 61us/step - loss: 1.3089\n", "Epoch 22/100\n", - "727/727 [==============================] - 0s 48us/step - loss: 1.0508\n", + "727/727 [==============================] - 0s 56us/step - loss: 1.2828\n", "Epoch 23/100\n", - "727/727 [==============================] - 0s 52us/step - loss: 1.0369\n", + "727/727 [==============================] - 0s 59us/step - loss: 1.2563\n", "Epoch 24/100\n", - "727/727 [==============================] - 0s 48us/step - loss: 1.0318\n", + "727/727 [==============================] - 0s 53us/step - loss: 1.2266\n", "Epoch 25/100\n", - "727/727 [==============================] - 0s 50us/step - loss: 1.0269\n", + "727/727 [==============================] - 0s 54us/step - loss: 1.1967\n", "Epoch 26/100\n", - "727/727 [==============================] - 0s 49us/step - loss: 1.0217\n", + "727/727 [==============================] - 0s 59us/step - loss: 1.1702\n", "Epoch 27/100\n", - "727/727 [==============================] - 0s 52us/step - loss: 1.0140\n", + "727/727 [==============================] - 0s 66us/step - loss: 1.1475\n", "Epoch 28/100\n", - "727/727 [==============================] - 0s 48us/step - loss: 1.0088\n", + "727/727 [==============================] - 0s 58us/step - loss: 1.1275\n", "Epoch 29/100\n", - "727/727 [==============================] - 0s 49us/step - loss: 1.0122\n", + "727/727 [==============================] - 0s 58us/step - loss: 1.1086\n", "Epoch 30/100\n", - "727/727 [==============================] - 0s 48us/step - loss: 1.0005\n", + "727/727 [==============================] - 0s 54us/step - loss: 1.0934\n", "Epoch 31/100\n", - "727/727 [==============================] - 0s 54us/step - loss: 0.9952\n", + "727/727 [==============================] - 0s 58us/step - loss: 1.0805\n", "Epoch 32/100\n", - "727/727 [==============================] - 0s 50us/step - loss: 0.9934\n", + "727/727 [==============================] - 0s 52us/step - loss: 1.0697\n", "Epoch 33/100\n", - "727/727 [==============================] - 0s 49us/step - loss: 0.9878\n", + "727/727 [==============================] - 0s 52us/step - loss: 1.0601\n", "Epoch 34/100\n", - "727/727 [==============================] - 0s 47us/step - loss: 0.9854\n", + "727/727 [==============================] - 0s 56us/step - loss: 1.0516\n", "Epoch 35/100\n", - "727/727 [==============================] - 0s 51us/step - loss: 0.9852\n", + "727/727 [==============================] - 0s 52us/step - loss: 1.0443\n", "Epoch 36/100\n", - "727/727 [==============================] - 0s 56us/step - loss: 0.9795\n", + "727/727 [==============================] - 0s 56us/step - loss: 1.0379\n", "Epoch 37/100\n", - "727/727 [==============================] - 0s 93us/step - loss: 0.9780\n", + "727/727 [==============================] - 0s 51us/step - loss: 1.0322\n", "Epoch 38/100\n", - "727/727 [==============================] - 0s 107us/step - loss: 0.9736\n", + "727/727 [==============================] - 0s 54us/step - loss: 1.0271\n", "Epoch 39/100\n", - "727/727 [==============================] - 0s 85us/step - loss: 0.9701\n", + "727/727 [==============================] - ETA: 0s - loss: 0.766 - 0s 56us/step - loss: 1.0225\n", "Epoch 40/100\n", - "727/727 [==============================] - 0s 71us/step - loss: 0.9684\n", + "727/727 [==============================] - 0s 61us/step - loss: 1.0183\n", "Epoch 41/100\n", - "727/727 [==============================] - 0s 69us/step - loss: 0.9642\n", + "727/727 [==============================] - 0s 52us/step - loss: 1.0146\n", "Epoch 42/100\n", - "727/727 [==============================] - 0s 65us/step - loss: 0.9630\n", + "727/727 [==============================] - 0s 58us/step - loss: 1.0112\n", "Epoch 43/100\n", - "727/727 [==============================] - 0s 69us/step - loss: 0.9594\n", + "727/727 [==============================] - 0s 50us/step - loss: 1.0080\n", "Epoch 44/100\n", - "727/727 [==============================] - 0s 63us/step - loss: 0.9568\n", + "727/727 [==============================] - 0s 52us/step - loss: 1.0051\n", "Epoch 45/100\n", - "727/727 [==============================] - 0s 80us/step - loss: 0.9529\n", + "727/727 [==============================] - 0s 54us/step - loss: 1.0024\n", "Epoch 46/100\n", - "727/727 [==============================] - 0s 66us/step - loss: 0.9551\n", + "727/727 [==============================] - 0s 55us/step - loss: 0.9999\n", "Epoch 47/100\n", - "727/727 [==============================] - 0s 64us/step - loss: 0.9534\n", + "727/727 [==============================] - 0s 58us/step - loss: 0.9976\n", "Epoch 48/100\n", - "727/727 [==============================] - 0s 62us/step - loss: 0.9479\n", + "727/727 [==============================] - 0s 60us/step - loss: 0.9953\n", "Epoch 49/100\n", - "727/727 [==============================] - 0s 62us/step - loss: 0.9458\n", + "727/727 [==============================] - 0s 55us/step - loss: 0.9932\n", "Epoch 50/100\n", - "727/727 [==============================] - 0s 65us/step - loss: 0.9447\n", + "727/727 [==============================] - 0s 53us/step - loss: 0.9911\n", "Epoch 51/100\n", - "727/727 [==============================] - 0s 64us/step - loss: 0.9422\n", + "727/727 [==============================] - 0s 56us/step - loss: 0.9891\n", "Epoch 52/100\n", - "727/727 [==============================] - 0s 63us/step - loss: 0.9415\n", + "727/727 [==============================] - 0s 60us/step - loss: 0.9872\n", "Epoch 53/100\n", - "727/727 [==============================] - 0s 61us/step - loss: 0.9419\n", + "727/727 [==============================] - 0s 59us/step - loss: 0.9853\n", "Epoch 54/100\n", - "727/727 [==============================] - 0s 62us/step - loss: 0.9459\n", + "727/727 [==============================] - 0s 69us/step - loss: 0.9835\n", "Epoch 55/100\n", - "727/727 [==============================] - 0s 70us/step - loss: 0.9451\n", + "727/727 [==============================] - 0s 122us/step - loss: 0.9818\n", "Epoch 56/100\n", - "727/727 [==============================] - 0s 57us/step - loss: 0.9383\n", + "727/727 [==============================] - 0s 97us/step - loss: 0.9802\n", "Epoch 57/100\n", - "727/727 [==============================] - 0s 55us/step - loss: 0.9414\n", + "727/727 [==============================] - 0s 52us/step - loss: 0.9786\n", "Epoch 58/100\n", - "727/727 [==============================] - 0s 50us/step - loss: 0.9314\n", + "727/727 [==============================] - 0s 56us/step - loss: 0.9771\n", "Epoch 59/100\n", - "727/727 [==============================] - 0s 48us/step - loss: 0.9339\n", + "727/727 [==============================] - 0s 50us/step - loss: 0.9756\n", "Epoch 60/100\n", - "727/727 [==============================] - 0s 52us/step - loss: 0.9285\n", + "727/727 [==============================] - 0s 54us/step - loss: 0.9741\n", "Epoch 61/100\n", - "727/727 [==============================] - 0s 47us/step - loss: 0.9271\n", + "727/727 [==============================] - 0s 51us/step - loss: 0.9728\n", "Epoch 62/100\n", - "727/727 [==============================] - 0s 48us/step - loss: 0.9250\n", + "727/727 [==============================] - 0s 50us/step - loss: 0.9714\n", "Epoch 63/100\n", - "727/727 [==============================] - 0s 49us/step - loss: 0.9267\n", + "727/727 [==============================] - 0s 52us/step - loss: 0.9701\n", "Epoch 64/100\n", - "727/727 [==============================] - 0s 48us/step - loss: 0.9244\n", + "727/727 [==============================] - 0s 52us/step - loss: 0.9688\n", "Epoch 65/100\n", - "727/727 [==============================] - 0s 51us/step - loss: 0.9241\n", + "727/727 [==============================] - 0s 53us/step - loss: 0.9675\n", "Epoch 66/100\n", - "727/727 [==============================] - 0s 46us/step - loss: 0.9209\n", + "727/727 [==============================] - 0s 50us/step - loss: 0.9663\n", "Epoch 67/100\n", - "727/727 [==============================] - 0s 47us/step - loss: 0.9214\n", + "727/727 [==============================] - 0s 49us/step - loss: 0.9651\n", "Epoch 68/100\n", - "727/727 [==============================] - 0s 49us/step - loss: 0.9215\n", + "727/727 [==============================] - 0s 55us/step - loss: 0.9639\n", "Epoch 69/100\n", - "727/727 [==============================] - 0s 49us/step - loss: 0.9168\n", + "727/727 [==============================] - 0s 55us/step - loss: 0.9628\n", "Epoch 70/100\n", - "727/727 [==============================] - 0s 55us/step - loss: 0.9166\n", + "727/727 [==============================] - 0s 53us/step - loss: 0.9616\n", "Epoch 71/100\n", - "727/727 [==============================] - 0s 59us/step - loss: 0.9173\n", + "727/727 [==============================] - 0s 53us/step - loss: 0.9605\n", "Epoch 72/100\n", - "727/727 [==============================] - 0s 54us/step - loss: 0.9121\n", + "727/727 [==============================] - 0s 49us/step - loss: 0.9594\n", "Epoch 73/100\n", - "727/727 [==============================] - 0s 54us/step - loss: 0.9127\n", + "727/727 [==============================] - 0s 54us/step - loss: 0.9583\n", "Epoch 74/100\n", - "727/727 [==============================] - 0s 53us/step - loss: 0.9123\n", + "727/727 [==============================] - 0s 57us/step - loss: 0.9572\n", "Epoch 75/100\n", - "727/727 [==============================] - 0s 56us/step - loss: 0.9114\n", + "727/727 [==============================] - 0s 64us/step - loss: 0.9562\n", "Epoch 76/100\n", - "727/727 [==============================] - 0s 58us/step - loss: 0.9140\n", + "727/727 [==============================] - 0s 56us/step - loss: 0.9551\n", "Epoch 77/100\n", - "727/727 [==============================] - 0s 53us/step - loss: 0.9100\n", + "727/727 [==============================] - 0s 58us/step - loss: 0.9541\n", "Epoch 78/100\n", - "727/727 [==============================] - 0s 50us/step - loss: 0.9081\n", + "727/727 [==============================] - 0s 53us/step - loss: 0.9531\n", "Epoch 79/100\n", - "727/727 [==============================] - 0s 49us/step - loss: 0.9071\n", + "727/727 [==============================] - 0s 54us/step - loss: 0.9521\n", "Epoch 80/100\n", - "727/727 [==============================] - 0s 47us/step - loss: 0.9063\n", + "727/727 [==============================] - 0s 66us/step - loss: 0.9511\n", "Epoch 81/100\n", - "727/727 [==============================] - 0s 49us/step - loss: 0.9037\n", + "727/727 [==============================] - 0s 55us/step - loss: 0.9501\n", "Epoch 82/100\n", - "727/727 [==============================] - 0s 61us/step - loss: 0.9024\n", + "727/727 [==============================] - 0s 52us/step - loss: 0.9491\n", "Epoch 83/100\n", - "727/727 [==============================] - 0s 49us/step - loss: 0.9057\n", + "727/727 [==============================] - 0s 54us/step - loss: 0.9481\n", "Epoch 84/100\n", - "727/727 [==============================] - 0s 46us/step - loss: 0.9060\n", + "727/727 [==============================] - 0s 60us/step - loss: 0.9472\n", "Epoch 85/100\n", - "727/727 [==============================] - 0s 48us/step - loss: 0.9117\n", + "727/727 [==============================] - 0s 59us/step - loss: 0.9463\n", "Epoch 86/100\n", - "727/727 [==============================] - 0s 53us/step - loss: 0.8989\n", + "727/727 [==============================] - 0s 59us/step - loss: 0.9454\n", "Epoch 87/100\n" ] }, @@ -895,241 +895,241 @@ "name": "stdout", "output_type": "stream", "text": [ - "727/727 [==============================] - 0s 47us/step - loss: 0.9010\n", + "727/727 [==============================] - 0s 54us/step - loss: 0.9445\n", "Epoch 88/100\n", - "727/727 [==============================] - 0s 49us/step - loss: 0.9117\n", + "727/727 [==============================] - 0s 56us/step - loss: 0.9437\n", "Epoch 89/100\n", - "727/727 [==============================] - 0s 47us/step - loss: 0.8979\n", + "727/727 [==============================] - 0s 52us/step - loss: 0.9429\n", "Epoch 90/100\n", - "727/727 [==============================] - 0s 48us/step - loss: 0.8943\n", + "727/727 [==============================] - 0s 49us/step - loss: 0.9422\n", "Epoch 91/100\n", - "727/727 [==============================] - 0s 50us/step - loss: 0.9032\n", + "727/727 [==============================] - 0s 49us/step - loss: 0.9415\n", "Epoch 92/100\n", - "727/727 [==============================] - 0s 47us/step - loss: 0.8988\n", + "727/727 [==============================] - 0s 53us/step - loss: 0.9408\n", "Epoch 93/100\n", - "727/727 [==============================] - 0s 49us/step - loss: 0.8956\n", + "727/727 [==============================] - 0s 54us/step - loss: 0.9402\n", "Epoch 94/100\n", - "727/727 [==============================] - 0s 47us/step - loss: 0.8947\n", + "727/727 [==============================] - 0s 51us/step - loss: 0.9395\n", "Epoch 95/100\n", - "727/727 [==============================] - 0s 48us/step - loss: 0.8896\n", + "727/727 [==============================] - 0s 52us/step - loss: 0.9389\n", "Epoch 96/100\n", - "727/727 [==============================] - 0s 49us/step - loss: 0.8927\n", + "727/727 [==============================] - 0s 48us/step - loss: 0.9383\n", "Epoch 97/100\n", - "727/727 [==============================] - 0s 51us/step - loss: 0.8906\n", + "727/727 [==============================] - 0s 53us/step - loss: 0.9378\n", "Epoch 98/100\n", - "727/727 [==============================] - 0s 48us/step - loss: 0.8903\n", + "727/727 [==============================] - 0s 52us/step - loss: 0.9373\n", "Epoch 99/100\n", - "727/727 [==============================] - 0s 47us/step - loss: 0.8900\n", + "727/727 [==============================] - 0s 59us/step - loss: 0.9368\n", "Epoch 100/100\n", - "727/727 [==============================] - 0s 46us/step - loss: 0.8919\n", - "181/181 [==============================] - 0s 104us/step\n", + "727/727 [==============================] - ETA: 0s - loss: 0.564 - 0s 62us/step - loss: 0.9363\n", + "181/181 [==============================] - 0s 109us/step\n", "Epoch 1/100\n", - "727/727 [==============================] - 0s 245us/step - loss: 41.0806\n", + "727/727 [==============================] - 0s 170us/step - loss: 30.2340\n", "Epoch 2/100\n", - "727/727 [==============================] - 0s 67us/step - loss: 32.3391\n", + "727/727 [==============================] - 0s 57us/step - loss: 22.1595\n", "Epoch 3/100\n", - "727/727 [==============================] - 0s 73us/step - loss: 24.6962\n", + "727/727 [==============================] - 0s 56us/step - loss: 15.9794\n", "Epoch 4/100\n", - "727/727 [==============================] - 0s 55us/step - loss: 17.9724\n", + "727/727 [==============================] - 0s 52us/step - loss: 11.0835\n", "Epoch 5/100\n", - "727/727 [==============================] - 0s 72us/step - loss: 12.2436\n", + "727/727 [==============================] - 0s 102us/step - loss: 7.2851\n", "Epoch 6/100\n", - "727/727 [==============================] - 0s 64us/step - loss: 7.3969\n", + "727/727 [==============================] - 0s 56us/step - loss: 4.5818\n", "Epoch 7/100\n", - "727/727 [==============================] - 0s 58us/step - loss: 3.8910\n", + "727/727 [==============================] - 0s 55us/step - loss: 2.8925\n", "Epoch 8/100\n", - "727/727 [==============================] - 0s 59us/step - loss: 2.1084\n", + "727/727 [==============================] - 0s 58us/step - loss: 2.0167\n", "Epoch 9/100\n", - "727/727 [==============================] - 0s 69us/step - loss: 1.4921\n", + "727/727 [==============================] - 0s 53us/step - loss: 1.6366\n", "Epoch 10/100\n", - "727/727 [==============================] - 0s 63us/step - loss: 1.3456\n", + "727/727 [==============================] - 0s 53us/step - loss: 1.4912\n", "Epoch 11/100\n", - "727/727 [==============================] - 0s 57us/step - loss: 1.3028\n", + "727/727 [==============================] - 0s 56us/step - loss: 1.4259\n", "Epoch 12/100\n", - "727/727 [==============================] - 0s 50us/step - loss: 1.2783\n", + "727/727 [==============================] - 0s 54us/step - loss: 1.3855\n", "Epoch 13/100\n", - "727/727 [==============================] - 0s 55us/step - loss: 1.2573\n", + "727/727 [==============================] - 0s 54us/step - loss: 1.3539\n", "Epoch 14/100\n", - "727/727 [==============================] - 0s 55us/step - loss: 1.2384\n", + "727/727 [==============================] - 0s 51us/step - loss: 1.3268\n", "Epoch 15/100\n", - "727/727 [==============================] - 0s 50us/step - loss: 1.2191\n", + "727/727 [==============================] - 0s 55us/step - loss: 1.3030\n", "Epoch 16/100\n", - "727/727 [==============================] - 0s 57us/step - loss: 1.2025\n", + "727/727 [==============================] - 0s 51us/step - loss: 1.2815\n", "Epoch 17/100\n", - "727/727 [==============================] - 0s 59us/step - loss: 1.1860\n", + "727/727 [==============================] - 0s 52us/step - loss: 1.2622\n", "Epoch 18/100\n", - "727/727 [==============================] - 0s 56us/step - loss: 1.1717\n", + "727/727 [==============================] - 0s 52us/step - loss: 1.2448\n", "Epoch 19/100\n", - "727/727 [==============================] - 0s 113us/step - loss: 1.1586\n", + "727/727 [==============================] - 0s 52us/step - loss: 1.2293\n", "Epoch 20/100\n", - "727/727 [==============================] - 0s 99us/step - loss: 1.1468\n", + "727/727 [==============================] - 0s 53us/step - loss: 1.2153\n", "Epoch 21/100\n", - "727/727 [==============================] - 0s 78us/step - loss: 1.1387\n", + "727/727 [==============================] - 0s 55us/step - loss: 1.2027\n", "Epoch 22/100\n", - "727/727 [==============================] - 0s 71us/step - loss: 1.1298\n", + "727/727 [==============================] - 0s 52us/step - loss: 1.1915\n", "Epoch 23/100\n", - "727/727 [==============================] - 0s 79us/step - loss: 1.1196\n", + "727/727 [==============================] - 0s 50us/step - loss: 1.1809\n", "Epoch 24/100\n", - "727/727 [==============================] - 0s 62us/step - loss: 1.1141\n", + "727/727 [==============================] - 0s 49us/step - loss: 1.1709\n", "Epoch 25/100\n", - "727/727 [==============================] - 0s 58us/step - loss: 1.1057\n", + "727/727 [==============================] - 0s 50us/step - loss: 1.1613\n", "Epoch 26/100\n", - "727/727 [==============================] - 0s 57us/step - loss: 1.1010\n", + "727/727 [==============================] - 0s 53us/step - loss: 1.1521\n", "Epoch 27/100\n", - "727/727 [==============================] - 0s 65us/step - loss: 1.0946\n", + "727/727 [==============================] - 0s 53us/step - loss: 1.1435\n", "Epoch 28/100\n", - "727/727 [==============================] - 0s 61us/step - loss: 1.0892\n", + "727/727 [==============================] - 0s 60us/step - loss: 1.1353\n", "Epoch 29/100\n", - "727/727 [==============================] - 0s 65us/step - loss: 1.0847\n", + "727/727 [==============================] - 0s 52us/step - loss: 1.1275\n", "Epoch 30/100\n", - "727/727 [==============================] - 0s 69us/step - loss: 1.0812\n", + "727/727 [==============================] - 0s 64us/step - loss: 1.1200\n", "Epoch 31/100\n", - "727/727 [==============================] - 0s 65us/step - loss: 1.0765\n", + "727/727 [==============================] - 0s 66us/step - loss: 1.1126\n", "Epoch 32/100\n", - "727/727 [==============================] - 0s 61us/step - loss: 1.0728\n", + "727/727 [==============================] - 0s 90us/step - loss: 1.1055\n", "Epoch 33/100\n", - "727/727 [==============================] - 0s 58us/step - loss: 1.0690\n", + "727/727 [==============================] - 0s 90us/step - loss: 1.0987\n", "Epoch 34/100\n", - "727/727 [==============================] - 0s 62us/step - loss: 1.0670\n", + "727/727 [==============================] - 0s 66us/step - loss: 1.0920\n", "Epoch 35/100\n", - "727/727 [==============================] - 0s 62us/step - loss: 1.0625\n", + "727/727 [==============================] - 0s 52us/step - loss: 1.0854\n", "Epoch 36/100\n", - "727/727 [==============================] - 0s 63us/step - loss: 1.0599\n", + "727/727 [==============================] - 0s 51us/step - loss: 1.0790\n", "Epoch 37/100\n", - "727/727 [==============================] - 0s 62us/step - loss: 1.0549\n", + "727/727 [==============================] - 0s 55us/step - loss: 1.0727\n", "Epoch 38/100\n", - "727/727 [==============================] - 0s 51us/step - loss: 1.0561\n", + "727/727 [==============================] - 0s 69us/step - loss: 1.0665\n", "Epoch 39/100\n", - "727/727 [==============================] - 0s 51us/step - loss: 1.0493\n", + "727/727 [==============================] - 0s 46us/step - loss: 1.0605\n", "Epoch 40/100\n", - "727/727 [==============================] - 0s 71us/step - loss: 1.0478\n", + "727/727 [==============================] - 0s 54us/step - loss: 1.0547\n", "Epoch 41/100\n", - "727/727 [==============================] - 0s 71us/step - loss: 1.0435\n", + "727/727 [==============================] - 0s 57us/step - loss: 1.0491\n", "Epoch 42/100\n", - "727/727 [==============================] - 0s 64us/step - loss: 1.0408\n", + "727/727 [==============================] - 0s 60us/step - loss: 1.0438\n", "Epoch 43/100\n", - "727/727 [==============================] - 0s 55us/step - loss: 1.0366\n", + "727/727 [==============================] - 0s 58us/step - loss: 1.0386\n", "Epoch 44/100\n", - "727/727 [==============================] - 0s 50us/step - loss: 1.0352\n", + "727/727 [==============================] - 0s 59us/step - loss: 1.0335\n", "Epoch 45/100\n", - "727/727 [==============================] - 0s 55us/step - loss: 1.0336\n", + "727/727 [==============================] - 0s 56us/step - loss: 1.0286\n", "Epoch 46/100\n", - "727/727 [==============================] - 0s 48us/step - loss: 1.0323\n", + "727/727 [==============================] - 0s 61us/step - loss: 1.0239\n", "Epoch 47/100\n", - "727/727 [==============================] - 0s 48us/step - loss: 1.0301\n", + "727/727 [==============================] - 0s 55us/step - loss: 1.0193\n", "Epoch 48/100\n", - "727/727 [==============================] - 0s 49us/step - loss: 1.0279\n", + "727/727 [==============================] - 0s 55us/step - loss: 1.0149\n", "Epoch 49/100\n", - "727/727 [==============================] - 0s 50us/step - loss: 1.0241\n", + "727/727 [==============================] - 0s 57us/step - loss: 1.0107\n", "Epoch 50/100\n", - "727/727 [==============================] - 0s 114us/step - loss: 1.0202\n", + "727/727 [==============================] - 0s 60us/step - loss: 1.0067\n", "Epoch 51/100\n", - "727/727 [==============================] - 0s 71us/step - loss: 1.0183\n", + "727/727 [==============================] - 0s 52us/step - loss: 1.0030\n", "Epoch 52/100\n", - "727/727 [==============================] - 0s 61us/step - loss: 1.0179\n", + "727/727 [==============================] - 0s 62us/step - loss: 0.9996\n", "Epoch 53/100\n", - "727/727 [==============================] - 0s 81us/step - loss: 1.0144\n", + "727/727 [==============================] - 0s 62us/step - loss: 0.9963\n", "Epoch 54/100\n", - "727/727 [==============================] - 0s 69us/step - loss: 1.0118\n", + "727/727 [==============================] - 0s 56us/step - loss: 0.9931\n", "Epoch 55/100\n", - "727/727 [==============================] - 0s 68us/step - loss: 1.0093\n", + "727/727 [==============================] - 0s 56us/step - loss: 0.9900\n", "Epoch 56/100\n", - "727/727 [==============================] - 0s 72us/step - loss: 1.0102\n", + "727/727 [==============================] - 0s 62us/step - loss: 0.9869\n", "Epoch 57/100\n", - "727/727 [==============================] - 0s 69us/step - loss: 1.0081\n", + "727/727 [==============================] - 0s 51us/step - loss: 0.9839\n", "Epoch 58/100\n", - "727/727 [==============================] - 0s 70us/step - loss: 1.0051\n", + "727/727 [==============================] - 0s 69us/step - loss: 0.9809\n", "Epoch 59/100\n", - "727/727 [==============================] - 0s 69us/step - loss: 1.0022\n", + "727/727 [==============================] - 0s 49us/step - loss: 0.9781\n", "Epoch 60/100\n", - "727/727 [==============================] - 0s 70us/step - loss: 1.0033\n", + "727/727 [==============================] - 0s 52us/step - loss: 0.9753\n", "Epoch 61/100\n", - "727/727 [==============================] - 0s 71us/step - loss: 0.9995\n", + "727/727 [==============================] - 0s 52us/step - loss: 0.9727\n", "Epoch 62/100\n", - "727/727 [==============================] - 0s 62us/step - loss: 0.9973\n", + "727/727 [==============================] - 0s 50us/step - loss: 0.9701\n", "Epoch 63/100\n", - "727/727 [==============================] - 0s 76us/step - loss: 0.9973\n", + "727/727 [==============================] - 0s 56us/step - loss: 0.9677\n", "Epoch 64/100\n", - "727/727 [==============================] - 0s 255us/step - loss: 0.9957\n", + "727/727 [==============================] - 0s 60us/step - loss: 0.9653\n", "Epoch 65/100\n", - "727/727 [==============================] - 0s 50us/step - loss: 0.9919\n", + "727/727 [==============================] - 0s 49us/step - loss: 0.9631\n", "Epoch 66/100\n", - "727/727 [==============================] - 0s 52us/step - loss: 0.9938\n", + "727/727 [==============================] - 0s 52us/step - loss: 0.9608\n", "Epoch 67/100\n", - "727/727 [==============================] - 0s 52us/step - loss: 0.9903\n", + "727/727 [==============================] - 0s 51us/step - loss: 0.9586\n", "Epoch 68/100\n", - "727/727 [==============================] - 0s 82us/step - loss: 0.9847\n", + "727/727 [==============================] - 0s 53us/step - loss: 0.9565\n", "Epoch 69/100\n", - "727/727 [==============================] - 0s 56us/step - loss: 0.9840\n", + "727/727 [==============================] - 0s 50us/step - loss: 0.9545\n", "Epoch 70/100\n", - "727/727 [==============================] - 0s 87us/step - loss: 0.9808\n", + "727/727 [==============================] - 0s 49us/step - loss: 0.9526\n", "Epoch 71/100\n", - "727/727 [==============================] - 0s 124us/step - loss: 0.9830\n", + "727/727 [==============================] - 0s 66us/step - loss: 0.9507\n", "Epoch 72/100\n", - "727/727 [==============================] - 0s 59us/step - loss: 0.9789\n", + "727/727 [==============================] - 0s 49us/step - loss: 0.9488\n", "Epoch 73/100\n", - "727/727 [==============================] - 0s 69us/step - loss: 0.9805\n", + "727/727 [==============================] - 0s 50us/step - loss: 0.9471\n", "Epoch 74/100\n", - "727/727 [==============================] - 0s 72us/step - loss: 0.9762\n", + "727/727 [==============================] - 0s 47us/step - loss: 0.9453\n", "Epoch 75/100\n", - "727/727 [==============================] - 0s 58us/step - loss: 0.9762\n", + "727/727 [==============================] - 0s 84us/step - loss: 0.9437\n", "Epoch 76/100\n", - "727/727 [==============================] - 0s 61us/step - loss: 0.9742\n", + "727/727 [==============================] - 0s 54us/step - loss: 0.9420\n", "Epoch 77/100\n", - "727/727 [==============================] - 0s 74us/step - loss: 0.9716\n", + "727/727 [==============================] - 0s 58us/step - loss: 0.9402\n", "Epoch 78/100\n", - "727/727 [==============================] - 0s 49us/step - loss: 0.9697\n", + "727/727 [==============================] - 0s 68us/step - loss: 0.9386\n", "Epoch 79/100\n", - "727/727 [==============================] - 0s 58us/step - loss: 0.9700\n", + "727/727 [==============================] - 0s 64us/step - loss: 0.9370\n", "Epoch 80/100\n", - "727/727 [==============================] - 0s 50us/step - loss: 0.9684\n", + "727/727 [==============================] - 0s 65us/step - loss: 0.9354\n", "Epoch 81/100\n", - "727/727 [==============================] - 0s 57us/step - loss: 0.9683\n", + "727/727 [==============================] - 0s 55us/step - loss: 0.9338\n", "Epoch 82/100\n", - "727/727 [==============================] - 0s 60us/step - loss: 0.9669\n", - "Epoch 83/100\n", - "727/727 [==============================] - 0s 68us/step - loss: 0.9685\n" + "727/727 [==============================] - 0s 61us/step - loss: 0.9323\n", + "Epoch 83/100\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ + "727/727 [==============================] - 0s 56us/step - loss: 0.9308\n", "Epoch 84/100\n", - "727/727 [==============================] - 0s 68us/step - loss: 0.9613\n", + "727/727 [==============================] - 0s 64us/step - loss: 0.9293\n", "Epoch 85/100\n", - "727/727 [==============================] - 0s 58us/step - loss: 0.9634\n", + "727/727 [==============================] - 0s 60us/step - loss: 0.9278\n", "Epoch 86/100\n", - "727/727 [==============================] - 0s 52us/step - loss: 0.9577\n", + "727/727 [==============================] - 0s 58us/step - loss: 0.9263\n", "Epoch 87/100\n", - "727/727 [==============================] - 0s 52us/step - loss: 0.9581\n", + "727/727 [==============================] - 0s 54us/step - loss: 0.9247\n", "Epoch 88/100\n", - "727/727 [==============================] - 0s 59us/step - loss: 0.9565\n", + "727/727 [==============================] - 0s 56us/step - loss: 0.9234\n", "Epoch 89/100\n", - "727/727 [==============================] - 0s 72us/step - loss: 0.9563\n", + "727/727 [==============================] - 0s 53us/step - loss: 0.9220\n", "Epoch 90/100\n", - "727/727 [==============================] - 0s 111us/step - loss: 0.9534\n", + "727/727 [==============================] - 0s 56us/step - loss: 0.9205\n", "Epoch 91/100\n", - "727/727 [==============================] - 0s 94us/step - loss: 0.9532\n", + "727/727 [==============================] - 0s 57us/step - loss: 0.9192\n", "Epoch 92/100\n", - "727/727 [==============================] - 0s 73us/step - loss: 0.9509\n", + "727/727 [==============================] - 0s 54us/step - loss: 0.9178\n", "Epoch 93/100\n", - "727/727 [==============================] - 0s 59us/step - loss: 0.9516\n", + "727/727 [==============================] - 0s 56us/step - loss: 0.9164\n", "Epoch 94/100\n", - "727/727 [==============================] - 0s 68us/step - loss: 0.9494\n", + "727/727 [==============================] - 0s 58us/step - loss: 0.9150\n", "Epoch 95/100\n", - "727/727 [==============================] - 0s 55us/step - loss: 0.9465\n", + "727/727 [==============================] - 0s 59us/step - loss: 0.9137\n", "Epoch 96/100\n", - "727/727 [==============================] - 0s 55us/step - loss: 0.9490\n", + "727/727 [==============================] - 0s 62us/step - loss: 0.9124\n", "Epoch 97/100\n", - "727/727 [==============================] - 0s 53us/step - loss: 0.9446\n", + "727/727 [==============================] - 0s 60us/step - loss: 0.9111\n", "Epoch 98/100\n", - "727/727 [==============================] - 0s 77us/step - loss: 0.9441\n", + "727/727 [==============================] - 0s 55us/step - loss: 0.9099\n", "Epoch 99/100\n", - "727/727 [==============================] - 0s 60us/step - loss: 0.9435\n", + "727/727 [==============================] - 0s 54us/step - loss: 0.9086\n", "Epoch 100/100\n", - "727/727 [==============================] - 0s 66us/step - loss: 0.9443\n", - "181/181 [==============================] - 0s 131us/step\n" + "727/727 [==============================] - 0s 50us/step - loss: 0.9076\n", + "181/181 [==============================] - 0s 84us/step\n" ] } ], @@ -1148,7 +1148,7 @@ "name": "stdout", "output_type": "stream", "text": [ - "Final Cross Validation Loss = 0.9297\n" + "Final Cross Validation Loss = 0.9010\n" ] } ], diff --git a/Chapter04/Exercise4.03/Exercise4.03_Unit_test.py b/Chapter04/Exercise4.03/Exercise4.03_Unit_test.py index 807b4a12..15407034 100755 --- a/Chapter04/Exercise4.03/Exercise4.03_Unit_test.py +++ b/Chapter04/Exercise4.03/Exercise4.03_Unit_test.py @@ -79,7 +79,7 @@ def test_model_iter(self): models = [build_model_1, build_model_2, build_model_3] # loop over three models for m in range(len(models)): - model = KerasRegressor(build_fn=models[m], epochs=100, batch_size=20, verbose=0) + model = KerasRegressor(build_fn=models[m], epochs=100, batch_size=20, verbose=0, shuffle=False) kf = KFold(n_splits=3) result = cross_val_score(model, self.X, self.y, cv=kf) self.results_1.append(result) @@ -102,7 +102,7 @@ def test_batch_epoch_iter(self): # Loop over pairs of epochs and batch_size for e in range(len(epochs)): for b in range(len(batches)): - model = KerasRegressor(build_fn= build_model_2, epochs= epochs[e], batch_size= batches[b], verbose=0) + model = KerasRegressor(build_fn= build_model_2, epochs= epochs[e], batch_size= batches[b], verbose=0, shuffle=False) kf = KFold(n_splits=3) result = cross_val_score(model, self.X, self.y, cv=kf) self.results_2.append(result) @@ -123,7 +123,7 @@ def test_opt_act_iter(self): for a in range(len(activations)): optimizer = optimizers[o] activation = activations[a] - model = KerasRegressor(build_fn= build_model_2_mod, epochs=100, batch_size=20, verbose=0) + model = KerasRegressor(build_fn= build_model_2_mod, epochs=100, batch_size=20, verbose=0, shuffle=False) kf = KFold(n_splits=3) result = cross_val_score(model, self.X, self.y, cv=kf) self.results_3.append(result) diff --git a/Chapter04/Exercise4.03/Exercise4_03.ipynb b/Chapter04/Exercise4.03/Exercise4_03.ipynb index a6735e5b..f176b637 100644 --- a/Chapter04/Exercise4.03/Exercise4_03.ipynb +++ b/Chapter04/Exercise4.03/Exercise4_03.ipynb @@ -84,7 +84,7 @@ }, { "cell_type": "code", - "execution_count": 4, + "execution_count": 3, "metadata": {}, "outputs": [], "source": [ @@ -102,7 +102,7 @@ "models = [build_model_1, build_model_2, build_model_3]\n", "# loop over three models\n", "for m in range(len(models)):\n", - " model = KerasRegressor(build_fn=models[m], epochs=100, batch_size=20, verbose=0)\n", + " model = KerasRegressor(build_fn=models[m], epochs=100, batch_size=20, verbose=0, shuffle=False)\n", " kf = KFold(n_splits=3)\n", " result = cross_val_score(model, X, y, cv=kf)\n", " results_1.append(result)" @@ -110,16 +110,16 @@ }, { "cell_type": "code", - "execution_count": 5, + "execution_count": 4, "metadata": {}, "outputs": [ { "name": "stdout", "output_type": "stream", "text": [ - "Cross Validation Loss for Model 1 = 0.990475798256843\n", - "Cross Validation Loss for Model 2 = 0.926532513151634\n", - "Cross Validation Loss for Model 3 = 0.9735719371528117\n" + "Cross Validation Loss for Model 1 = 1.0304891858847403\n", + "Cross Validation Loss for Model 2 = 0.9583476809877253\n", + "Cross Validation Loss for Model 3 = 1.7218542151593959\n" ] } ], @@ -132,7 +132,7 @@ }, { "cell_type": "code", - "execution_count": 6, + "execution_count": 5, "metadata": {}, "outputs": [], "source": [ @@ -146,7 +146,7 @@ "# Loop over pairs of epochs and batch_size\n", "for e in range(len(epochs)):\n", " for b in range(len(batches)):\n", - " model = KerasRegressor(build_fn= build_model_2, epochs= epochs[e], batch_size= batches[b], verbose=0)\n", + " model = KerasRegressor(build_fn= build_model_2, epochs= epochs[e], batch_size= batches[b], verbose=0, shuffle=False)\n", " kf = KFold(n_splits=3)\n", " result = cross_val_score(model, X, y, cv=kf)\n", " results_2.append(result)" @@ -154,17 +154,17 @@ }, { "cell_type": "code", - "execution_count": 7, + "execution_count": 6, "metadata": {}, "outputs": [ { "name": "stdout", "output_type": "stream", "text": [ - "batch_size = 20 , epochs = 100 , Test Loss = 0.9359159401008821\n", - "batch_size = 15 , epochs = 100 , Test Loss = 0.9642481369794683\n", - "batch_size = 20 , epochs = 150 , Test Loss = 0.9561188386646661\n", - "batch_size = 15 , epochs = 150 , Test Loss = 0.9359079093029896\n" + "batch_size = 20 , epochs = 100 , Test Loss = 0.9574524474039251\n", + "batch_size = 15 , epochs = 100 , Test Loss = 0.9478728140969382\n", + "batch_size = 20 , epochs = 150 , Test Loss = 0.8896153777520778\n", + "batch_size = 15 , epochs = 150 , Test Loss = 0.9501340622995701\n" ] } ], @@ -179,7 +179,7 @@ }, { "cell_type": "code", - "execution_count": 10, + "execution_count": 7, "metadata": {}, "outputs": [], "source": [ @@ -206,7 +206,7 @@ " for a in range(len(activations)):\n", " optimizer = optimizers[o]\n", " activation = activations[a]\n", - " model = KerasRegressor(build_fn= build_model_2, epochs=100, batch_size=20, verbose=0)\n", + " model = KerasRegressor(build_fn= build_model_2, epochs=100, batch_size=20, verbose=0, shuffle=False)\n", " kf = KFold(n_splits=3)\n", " result = cross_val_score(model, X, y, cv=kf)\n", " results_3.append(result)" @@ -214,19 +214,19 @@ }, { "cell_type": "code", - "execution_count": 11, + "execution_count": 8, "metadata": {}, "outputs": [ { "name": "stdout", "output_type": "stream", "text": [ - "activation = relu , optimizer = sgd , Test Loss = 1.0123592540516995\n", - "activation = tanh , optimizer = sgd , Test Loss = 3.393908379781118\n", - "activation = relu , optimizer = adam , Test Loss = 0.9662686089392641\n", - "activation = tanh , optimizer = adam , Test Loss = 2.1369285960222144\n", - "activation = relu , optimizer = rmsprop , Test Loss = 2.1892826984214984\n", - "activation = tanh , optimizer = rmsprop , Test Loss = 2.2029884275363014\n" + "activation = relu , optimizer = sgd , Test Loss = 0.9575929883722436\n", + "activation = tanh , optimizer = sgd , Test Loss = 0.9583991325450517\n", + "activation = relu , optimizer = adam , Test Loss = 0.9003895352626167\n", + "activation = tanh , optimizer = adam , Test Loss = 0.9860042032158066\n", + "activation = relu , optimizer = rmsprop , Test Loss = 0.9364833819585648\n", + "activation = tanh , optimizer = rmsprop , Test Loss = 0.9719708495282869\n" ] } ], diff --git a/Chapter05/Activity5.01/Activity5_01.ipynb b/Chapter05/Activity5.01/Activity5_01.ipynb index 506c3fb1..c6f27852 100644 --- a/Chapter05/Activity5.01/Activity5_01.ipynb +++ b/Chapter05/Activity5.01/Activity5_01.ipynb @@ -67,7 +67,7 @@ "model_1.add(Dense(1, activation='sigmoid'))\n", "model_1.compile(loss='binary_crossentropy', optimizer='sgd', metrics=['accuracy'])\n", "# train the model using training set while evaluating on test set\n", - "history=model_1.fit(X_train, y_train, batch_size = 20, epochs = 100, validation_data=(X_test, y_test), verbose=0)\n", + "history=model_1.fit(X_train, y_train, batch_size = 20, epochs = 100, validation_data=(X_test, y_test), verbose=0, shuffle=False)\n", "\n", "# import require packages for plotting\n", "import matplotlib.pyplot as plt \n", @@ -128,7 +128,7 @@ "model_2.compile(loss='binary_crossentropy', optimizer='sgd', metrics=['accuracy'])\n", "\n", "# train the model using training set while evaluationg on test set\n", - "history=model_2.fit(X_train, y_train, batch_size = 20, epochs = 100, validation_data=(X_test, y_test), verbose=0)\n", + "history=model_2.fit(X_train, y_train, batch_size = 20, epochs = 100, validation_data=(X_test, y_test), verbose=0, shuffle=False)\n", "\n", "# plot training error and test error\n", "matplotlib.rcParams['figure.figsize'] = (10.0, 8.0)\n", @@ -185,7 +185,7 @@ "model_3.compile(loss='binary_crossentropy', optimizer='sgd', metrics=['accuracy'])\n", "\n", "# train the model using training set while evaluationg on test set\n", - "history=model_3.fit(X_train, y_train, batch_size = 20, epochs = 100, validation_data=(X_test, y_test), verbose=0)\n", + "history=model_3.fit(X_train, y_train, batch_size = 20, epochs = 100, validation_data=(X_test, y_test), verbose=0, shuffle=False)\n", "\n", "# plot training error and test error\n", "matplotlib.rcParams['figure.figsize'] = (10.0, 8.0)\n", @@ -240,7 +240,7 @@ "model_4.compile(loss='binary_crossentropy', optimizer='sgd', metrics=['accuracy'])\n", "\n", "# train the model using training set while evaluationg on test set\n", - "history=model_4.fit(X_train, y_train, batch_size = 20, epochs = 100, validation_data=(X_test, y_test), verbose=0)\n", + "history=model_4.fit(X_train, y_train, batch_size = 20, epochs = 100, validation_data=(X_test, y_test), verbose=0, shuffle=False)\n", "\n", "# plot training error and test error\n", "matplotlib.rcParams['figure.figsize'] = (10.0, 8.0)\n", @@ -295,7 +295,7 @@ "model_5.compile(loss='binary_crossentropy', optimizer='sgd', metrics=['accuracy'])\n", "\n", "# train the model using training set while evaluationg on test set\n", - "history=model_5.fit(X_train, y_train, batch_size = 20, epochs = 100, validation_data=(X_test, y_test), verbose=0)\n", + "history=model_5.fit(X_train, y_train, batch_size = 20, epochs = 100, validation_data=(X_test, y_test), verbose=0, shuffle=False)\n", "\n", "# plot training error and test error\n", "matplotlib.rcParams['figure.figsize'] = (10.0, 8.0)\n", @@ -350,7 +350,7 @@ "model_6.compile(loss='binary_crossentropy', optimizer='sgd', metrics=['accuracy'])\n", "\n", "# train the model using training set while evaluationg on test set\n", - "history=model_6.fit(X_train, y_train, batch_size = 20, epochs = 100, validation_data=(X_test, y_test), verbose=0)\n", + "history=model_6.fit(X_train, y_train, batch_size = 20, epochs = 100, validation_data=(X_test, y_test), verbose=0, shuffle=False)\n", "\n", "# plot training error and test error\n", "matplotlib.rcParams['figure.figsize'] = (10.0, 8.0) \n", @@ -406,7 +406,7 @@ "model_7.compile(loss='binary_crossentropy', optimizer='sgd', metrics=['accuracy'])\n", "\n", "# train the model using training set while evaluationg on test set\n", - "history=model_7.fit(X_train, y_train, batch_size = 20, epochs = 100, validation_data=(X_test, y_test), verbose=0)\n", + "history=model_7.fit(X_train, y_train, batch_size = 20, epochs = 100, validation_data=(X_test, y_test), verbose=0, shuffle=False)\n", " \n", "# plot training error and test error\n", "matplotlib.rcParams['figure.figsize'] = (10.0, 8.0)\n", diff --git a/Chapter05/Activity5.03/Activity5.03_Unit_test.py b/Chapter05/Activity5.03/Activity5.03_Unit_test.py index 09589cea..bb7dd6e9 100755 --- a/Chapter05/Activity5.03/Activity5.03_Unit_test.py +++ b/Chapter05/Activity5.03/Activity5.03_Unit_test.py @@ -63,7 +63,7 @@ def build_model(lambda_parameter): np.random.seed(self.seed) random.set_seed(self.seed) - model = KerasClassifier(build_fn=build_model, verbose=0) + model = KerasClassifier(build_fn=build_model, verbose=0, shuffle=False) lambda_parameter = [0.01, 0.5, 1] epochs = [50, 100] batch_size = [20] @@ -89,7 +89,7 @@ def build_model(lambda_parameter): np.random.seed(self.seed) random.set_seed(self.seed) - model = KerasClassifier(build_fn=build_model, verbose=0) + model = KerasClassifier(build_fn=build_model, verbose=0, shuffle=False) lambda_parameter = [0.001, 0.01, 0.05, 0.1] epochs = [100] batch_size = [20] @@ -116,7 +116,7 @@ def build_model(rate): np.random.seed(self.seed) random.set_seed(self.seed) - model = KerasClassifier(build_fn=build_model, verbose=0) + model = KerasClassifier(build_fn=build_model, verbose=0, shuffle=False) rate = [0, 0.1, 0.2] epochs = [50, 100] batch_size = [20] @@ -143,7 +143,7 @@ def build_model(rate): np.random.seed(self.seed) random.set_seed(self.seed) - model = KerasClassifier(build_fn=build_model, verbose=0) + model = KerasClassifier(build_fn=build_model, verbose=0, shuffle=False) rate = [0.0, 0.05, 0.1] epochs = [100] batch_size = [20] diff --git a/Chapter05/Exercise5.01/Exercise5_01.ipynb b/Chapter05/Exercise5.01/Exercise5_01.ipynb index c9e8ff56..c4c6bad1 100644 --- a/Chapter05/Exercise5.01/Exercise5_01.ipynb +++ b/Chapter05/Exercise5.01/Exercise5_01.ipynb @@ -60,7 +60,7 @@ "# Choose the loss function to be binary cross entropy and the optimizer to be SGD for training the model\n", "model_1.compile(optimizer='sgd', loss='binary_crossentropy')\n", "# train the model\n", - "model_1.fit(X_train, y_train, epochs=300, batch_size=50, verbose=0)\n", + "model_1.fit(X_train, y_train, epochs=300, batch_size=50, verbose=0, shuffle=False)\n", "# evaluate on test set\n", "print(\"Test Loss =\", model_1.evaluate(X_test, y_test))" ] @@ -97,7 +97,7 @@ "# Choose the loss function to be binary cross entropy and the optimizer to be SGD for training the model\n", "model_2.compile(optimizer='sgd', loss='binary_crossentropy')\n", "# train the model\n", - "model_2.fit(X_train, y_train, epochs=300, batch_size=50, verbose=0)\n", + "model_2.fit(X_train, y_train, epochs=300, batch_size=50, verbose=0, shuffle=False)\n", "# evaluate on test set\n", "print(\"Test Loss =\", model_2.evaluate(X_test, y_test))" ] @@ -135,7 +135,7 @@ "# Choose the loss function to be binary cross entropy and the optimizer to be SGD for training the model\n", "model_3.compile(optimizer='sgd', loss='binary_crossentropy')\n", "# train the model\n", - "model_3.fit(X_train, y_train, epochs=300, batch_size=50, verbose=0)\n", + "model_3.fit(X_train, y_train, epochs=300, batch_size=50, verbose=0, shuffle=False)\n", "# evaluate on test set\n", "print(\"Test Loss =\", model_3.evaluate(X_test, y_test))" ] diff --git a/Chapter05/Exercise5.02/Exercise5_02.ipynb b/Chapter05/Exercise5.02/Exercise5_02.ipynb index e81205a9..c0a2caa9 100644 --- a/Chapter05/Exercise5.02/Exercise5_02.ipynb +++ b/Chapter05/Exercise5.02/Exercise5_02.ipynb @@ -49,7 +49,7 @@ "# Choose the loss function to be binary cross entropy and the optimizer to be SGD for training the model\n", "model_1.compile(optimizer='sgd', loss='binary_crossentropy')\n", "# train the model\n", - "history = model_1.fit(X_train, y_train, validation_data=(X_test, y_test), epochs=300, batch_size=50, verbose=0)" + "history = model_1.fit(X_train, y_train, validation_data=(X_test, y_test), epochs=300, batch_size=50, verbose=0, shuffle=False)" ] }, { @@ -117,7 +117,7 @@ "# define the early stopping callback\n", "es_callback = EarlyStopping(monitor='val_loss', mode='min')\n", "# train the model\n", - "history=model_2.fit(X_train, y_train, validation_data=(X_test, y_test), epochs=300, batch_size=50, callbacks=[es_callback], verbose=0)" + "history=model_2.fit(X_train, y_train, validation_data=(X_test, y_test), epochs=300, batch_size=50, callbacks=[es_callback], verbose=0, shuffle=False)" ] }, { @@ -180,7 +180,7 @@ "# define the early stopping callback\n", "es_callback = EarlyStopping(monitor='val_loss', mode='min', patience=10)\n", "# train the model\n", - "history=model_3.fit(X_train, y_train, validation_data=(X_test, y_test), epochs=300, batch_size=50, callbacks=[es_callback], verbose=0)" + "history=model_3.fit(X_train, y_train, validation_data=(X_test, y_test), epochs=300, batch_size=50, callbacks=[es_callback], verbose=0, shuffle=False)" ] }, {