From 61bd4824e4de06f10f96d678c328cbc872de0b68 Mon Sep 17 00:00:00 2001 From: s464903 Date: Sun, 9 Jun 2024 14:05:06 +0200 Subject: [PATCH] Upload files to "my_runs/1" --- my_runs/1/config.json | 5 + my_runs/1/cout.txt | 213 +++++++++++++++++++++++++++++++++++ my_runs/1/info.json | 15 +++ my_runs/1/metrics.json | 248 +++++++++++++++++++++++++++++++++++++++++ my_runs/1/model.keras | Bin 0 -> 20856 bytes 5 files changed, 481 insertions(+) create mode 100644 my_runs/1/config.json create mode 100644 my_runs/1/cout.txt create mode 100644 my_runs/1/info.json create mode 100644 my_runs/1/metrics.json create mode 100644 my_runs/1/model.keras diff --git a/my_runs/1/config.json b/my_runs/1/config.json new file mode 100644 index 0000000..f796e16 --- /dev/null +++ b/my_runs/1/config.json @@ -0,0 +1,5 @@ +{ + "dropout_layer_value": 0.4, + "num_epochs": 100, + "seed": 512638064 +} \ No newline at end of file diff --git a/my_runs/1/cout.txt b/my_runs/1/cout.txt new file mode 100644 index 0000000..6da2078 --- /dev/null +++ b/my_runs/1/cout.txt @@ -0,0 +1,213 @@ +2.16.1 +1.2.0 +3.2.1 +1.23.5 +1.5.2 +C:\Users\obses\AppData\Local\Programs\Python\Python310\lib\site-packages\sklearn\preprocessing\_encoders.py:808: FutureWarning: `sparse` was renamed to `sparse_output` in version 1.2 and will be removed in 1.4. `sparse_output` is ignored unless you leave `sparse` to its default value. + warnings.warn( +C:\Users\obses\AppData\Local\Programs\Python\Python310\lib\site-packages\keras\src\layers\core\dense.py:86: UserWarning: Do not pass an `input_shape`/`input_dim` argument to a layer. When using Sequential models, prefer using an `Input(shape)` object as the first layer in the model instead. + super().__init__(activity_regularizer=activity_regularizer, **kwargs) +Epoch 1/100 +2/2 - 1s - 340ms/step - accuracy: 0.4390 - loss: 2.1215 - val_accuracy: 0.6429 - val_loss: 2.0350 +Epoch 2/100 +2/2 - 0s - 23ms/step - accuracy: 0.3659 - loss: 2.0694 - val_accuracy: 0.7143 - val_loss: 2.0104 +Epoch 3/100 +2/2 - 0s - 22ms/step - accuracy: 0.3659 - loss: 2.1309 - val_accuracy: 0.8214 - val_loss: 1.9882 +Epoch 4/100 +2/2 - 0s - 23ms/step - accuracy: 0.4390 - loss: 2.0289 - val_accuracy: 0.8214 - val_loss: 1.9593 +Epoch 5/100 +2/2 - 0s - 21ms/step - accuracy: 0.6341 - loss: 1.9654 - val_accuracy: 0.8214 - val_loss: 1.9378 +Epoch 6/100 +2/2 - 0s - 22ms/step - accuracy: 0.6098 - loss: 2.0383 - val_accuracy: 0.8214 - val_loss: 1.9154 +Epoch 7/100 +2/2 - 0s - 22ms/step - accuracy: 0.6098 - loss: 2.0238 - val_accuracy: 0.8214 - val_loss: 1.8964 +Epoch 8/100 +2/2 - 0s - 23ms/step - accuracy: 0.6098 - loss: 1.9397 - val_accuracy: 0.8571 - val_loss: 1.8766 +Epoch 9/100 +2/2 - 0s - 25ms/step - accuracy: 0.5366 - loss: 1.9641 - val_accuracy: 0.8571 - val_loss: 1.8561 +Epoch 10/100 +2/2 - 0s - 23ms/step - accuracy: 0.5610 - loss: 1.9581 - val_accuracy: 0.8571 - val_loss: 1.8380 +Epoch 11/100 +2/2 - 0s - 23ms/step - accuracy: 0.7561 - loss: 1.9044 - val_accuracy: 0.8571 - val_loss: 1.8207 +Epoch 12/100 +2/2 - 0s - 23ms/step - accuracy: 0.5854 - loss: 1.9392 - val_accuracy: 0.8571 - val_loss: 1.8004 +Epoch 13/100 +2/2 - 0s - 24ms/step - accuracy: 0.7073 - loss: 1.8718 - val_accuracy: 0.8571 - val_loss: 1.7812 +Epoch 14/100 +2/2 - 0s - 22ms/step - accuracy: 0.7561 - loss: 1.8057 - val_accuracy: 0.8571 - val_loss: 1.7620 +Epoch 15/100 +2/2 - 0s - 22ms/step - accuracy: 0.8049 - loss: 1.8354 - val_accuracy: 0.8571 - val_loss: 1.7440 +Epoch 16/100 +2/2 - 0s - 24ms/step - accuracy: 0.7073 - loss: 1.8501 - val_accuracy: 0.8571 - val_loss: 1.7269 +Epoch 17/100 +2/2 - 0s - 24ms/step - accuracy: 0.7561 - loss: 1.7831 - val_accuracy: 0.8571 - val_loss: 1.7084 +Epoch 18/100 +2/2 - 0s - 24ms/step - accuracy: 0.8537 - loss: 1.7120 - val_accuracy: 0.8571 - val_loss: 1.6931 +Epoch 19/100 +2/2 - 0s - 22ms/step - accuracy: 0.8049 - loss: 1.8020 - val_accuracy: 0.8571 - val_loss: 1.6786 +Epoch 20/100 +2/2 - 0s - 24ms/step - accuracy: 0.8049 - loss: 1.7531 - val_accuracy: 0.8571 - val_loss: 1.6630 +Epoch 21/100 +2/2 - 0s - 21ms/step - accuracy: 0.7561 - loss: 1.7808 - val_accuracy: 0.8571 - val_loss: 1.6489 +Epoch 22/100 +2/2 - 0s - 21ms/step - accuracy: 0.7561 - loss: 1.7794 - val_accuracy: 0.8571 - val_loss: 1.6352 +Epoch 23/100 +2/2 - 0s - 23ms/step - accuracy: 0.7805 - loss: 1.6697 - val_accuracy: 0.8571 - val_loss: 1.6184 +Epoch 24/100 +2/2 - 0s - 22ms/step - accuracy: 0.8537 - loss: 1.6814 - val_accuracy: 0.8571 - val_loss: 1.6058 +Epoch 25/100 +2/2 - 0s - 29ms/step - accuracy: 0.8293 - loss: 1.6687 - val_accuracy: 0.8571 - val_loss: 1.5919 +Epoch 26/100 +2/2 - 0s - 22ms/step - accuracy: 0.8293 - loss: 1.7052 - val_accuracy: 0.8571 - val_loss: 1.5786 +Epoch 27/100 +2/2 - 0s - 21ms/step - accuracy: 0.8293 - loss: 1.6147 - val_accuracy: 0.8571 - val_loss: 1.5663 +Epoch 28/100 +2/2 - 0s - 21ms/step - accuracy: 0.7805 - loss: 1.6207 - val_accuracy: 0.8571 - val_loss: 1.5529 +Epoch 29/100 +2/2 - 0s - 22ms/step - accuracy: 0.8293 - loss: 1.5964 - val_accuracy: 0.8571 - val_loss: 1.5413 +Epoch 30/100 +2/2 - 0s - 21ms/step - accuracy: 0.8537 - loss: 1.6258 - val_accuracy: 0.8571 - val_loss: 1.5294 +Epoch 31/100 +2/2 - 0s - 21ms/step - accuracy: 0.8537 - loss: 1.5512 - val_accuracy: 0.8571 - val_loss: 1.5175 +Epoch 32/100 +2/2 - 0s - 21ms/step - accuracy: 0.8537 - loss: 1.6674 - val_accuracy: 0.8571 - val_loss: 1.5070 +Epoch 33/100 +2/2 - 0s - 22ms/step - accuracy: 0.8293 - loss: 1.5848 - val_accuracy: 0.8571 - val_loss: 1.4977 +Epoch 34/100 +2/2 - 0s - 21ms/step - accuracy: 0.8537 - loss: 1.5467 - val_accuracy: 0.8571 - val_loss: 1.4869 +Epoch 35/100 +2/2 - 0s - 20ms/step - accuracy: 0.8293 - loss: 1.6244 - val_accuracy: 0.8571 - val_loss: 1.4778 +Epoch 36/100 +2/2 - 0s - 24ms/step - accuracy: 0.8293 - loss: 1.5275 - val_accuracy: 0.8571 - val_loss: 1.4672 +Epoch 37/100 +2/2 - 0s - 22ms/step - accuracy: 0.8049 - loss: 1.7105 - val_accuracy: 0.8571 - val_loss: 1.4595 +Epoch 38/100 +2/2 - 0s - 22ms/step - accuracy: 0.8537 - loss: 1.6137 - val_accuracy: 0.8571 - val_loss: 1.4509 +Epoch 39/100 +2/2 - 0s - 22ms/step - accuracy: 0.8537 - loss: 1.5726 - val_accuracy: 0.8571 - val_loss: 1.4425 +Epoch 40/100 +2/2 - 0s - 21ms/step - accuracy: 0.8293 - loss: 1.5345 - val_accuracy: 0.8571 - val_loss: 1.4347 +Epoch 41/100 +2/2 - 0s - 22ms/step - accuracy: 0.8537 - loss: 1.5689 - val_accuracy: 0.8571 - val_loss: 1.4269 +Epoch 42/100 +2/2 - 0s - 21ms/step - accuracy: 0.8537 - loss: 1.4239 - val_accuracy: 0.8571 - val_loss: 1.4175 +Epoch 43/100 +2/2 - 0s - 22ms/step - accuracy: 0.8293 - loss: 1.5922 - val_accuracy: 0.8571 - val_loss: 1.4099 +Epoch 44/100 +2/2 - 0s - 21ms/step - accuracy: 0.8537 - loss: 1.5006 - val_accuracy: 0.8571 - val_loss: 1.4021 +Epoch 45/100 +2/2 - 0s - 21ms/step - accuracy: 0.8049 - loss: 1.4858 - val_accuracy: 0.8571 - val_loss: 1.3944 +Epoch 46/100 +2/2 - 0s - 21ms/step - accuracy: 0.8537 - loss: 1.4769 - val_accuracy: 0.8571 - val_loss: 1.3874 +Epoch 47/100 +2/2 - 0s - 22ms/step - accuracy: 0.8537 - loss: 1.4211 - val_accuracy: 0.8571 - val_loss: 1.3796 +Epoch 48/100 +2/2 - 0s - 21ms/step - accuracy: 0.8537 - loss: 1.4060 - val_accuracy: 0.8571 - val_loss: 1.3717 +Epoch 49/100 +2/2 - 0s - 20ms/step - accuracy: 0.8537 - loss: 1.4741 - val_accuracy: 0.8571 - val_loss: 1.3652 +Epoch 50/100 +2/2 - 0s - 22ms/step - accuracy: 0.8537 - loss: 1.4989 - val_accuracy: 0.8571 - val_loss: 1.3588 +Epoch 51/100 +2/2 - 0s - 23ms/step - accuracy: 0.8537 - loss: 1.4718 - val_accuracy: 0.8571 - val_loss: 1.3521 +Epoch 52/100 +2/2 - 0s - 24ms/step - accuracy: 0.8293 - loss: 1.4712 - val_accuracy: 0.8571 - val_loss: 1.3482 +Epoch 53/100 +2/2 - 0s - 25ms/step - accuracy: 0.8537 - loss: 1.3657 - val_accuracy: 0.8571 - val_loss: 1.3425 +Epoch 54/100 +2/2 - 0s - 22ms/step - accuracy: 0.8537 - loss: 1.3847 - val_accuracy: 0.8571 - val_loss: 1.3366 +Epoch 55/100 +2/2 - 0s - 23ms/step - accuracy: 0.8537 - loss: 1.3766 - val_accuracy: 0.8571 - val_loss: 1.3315 +Epoch 56/100 +2/2 - 0s - 22ms/step - accuracy: 0.8537 - loss: 1.4275 - val_accuracy: 0.8571 - val_loss: 1.3253 +Epoch 57/100 +2/2 - 0s - 22ms/step - accuracy: 0.8537 - loss: 1.4709 - val_accuracy: 0.8571 - val_loss: 1.3209 +Epoch 58/100 +2/2 - 0s - 23ms/step - accuracy: 0.8537 - loss: 1.4226 - val_accuracy: 0.8571 - val_loss: 1.3176 +Epoch 59/100 +2/2 - 0s - 23ms/step - accuracy: 0.8537 - loss: 1.4491 - val_accuracy: 0.8571 - val_loss: 1.3126 +Epoch 60/100 +2/2 - 0s - 24ms/step - accuracy: 0.8537 - loss: 1.4165 - val_accuracy: 0.8571 - val_loss: 1.3073 +Epoch 61/100 +2/2 - 0s - 22ms/step - accuracy: 0.8537 - loss: 1.3914 - val_accuracy: 0.8571 - val_loss: 1.3027 +Epoch 62/100 +2/2 - 0s - 24ms/step - accuracy: 0.8537 - loss: 1.3560 - val_accuracy: 0.8571 - val_loss: 1.2974 +Epoch 63/100 +2/2 - 0s - 23ms/step - accuracy: 0.8537 - loss: 1.3432 - val_accuracy: 0.8571 - val_loss: 1.2915 +Epoch 64/100 +2/2 - 0s - 22ms/step - accuracy: 0.8537 - loss: 1.3833 - val_accuracy: 0.8571 - val_loss: 1.2866 +Epoch 65/100 +2/2 - 0s - 23ms/step - accuracy: 0.8537 - loss: 1.4477 - val_accuracy: 0.8571 - val_loss: 1.2834 +Epoch 66/100 +2/2 - 0s - 24ms/step - accuracy: 0.8537 - loss: 1.3054 - val_accuracy: 0.8571 - val_loss: 1.2794 +Epoch 67/100 +2/2 - 0s - 23ms/step - accuracy: 0.8537 - loss: 1.2840 - val_accuracy: 0.8571 - val_loss: 1.2742 +Epoch 68/100 +2/2 - 0s - 22ms/step - accuracy: 0.8537 - loss: 1.4320 - val_accuracy: 0.8571 - val_loss: 1.2703 +Epoch 69/100 +2/2 - 0s - 21ms/step - accuracy: 0.8537 - loss: 1.4165 - val_accuracy: 0.8571 - val_loss: 1.2664 +Epoch 70/100 +2/2 - 0s - 23ms/step - accuracy: 0.8537 - loss: 1.3210 - val_accuracy: 0.8571 - val_loss: 1.2617 +Epoch 71/100 +2/2 - 0s - 21ms/step - accuracy: 0.8537 - loss: 1.3698 - val_accuracy: 0.8571 - val_loss: 1.2580 +Epoch 72/100 +2/2 - 0s - 21ms/step - accuracy: 0.8537 - loss: 1.3679 - val_accuracy: 0.8571 - val_loss: 1.2546 +Epoch 73/100 +2/2 - 0s - 21ms/step - accuracy: 0.8537 - loss: 1.3647 - val_accuracy: 0.8571 - val_loss: 1.2506 +Epoch 74/100 +2/2 - 0s - 22ms/step - accuracy: 0.8537 - loss: 1.3088 - val_accuracy: 0.8571 - val_loss: 1.2465 +Epoch 75/100 +2/2 - 0s - 23ms/step - accuracy: 0.8537 - loss: 1.3304 - val_accuracy: 0.8571 - val_loss: 1.2425 +Epoch 76/100 +2/2 - 0s - 29ms/step - accuracy: 0.8537 - loss: 1.2813 - val_accuracy: 0.8571 - val_loss: 1.2389 +Epoch 77/100 +2/2 - 0s - 28ms/step - accuracy: 0.8537 - loss: 1.3477 - val_accuracy: 0.8571 - val_loss: 1.2352 +Epoch 78/100 +2/2 - 0s - 26ms/step - accuracy: 0.8537 - loss: 1.3885 - val_accuracy: 0.8571 - val_loss: 1.2315 +Epoch 79/100 +2/2 - 0s - 25ms/step - accuracy: 0.8537 - loss: 1.2939 - val_accuracy: 0.8571 - val_loss: 1.2280 +Epoch 80/100 +2/2 - 0s - 26ms/step - accuracy: 0.8537 - loss: 1.3508 - val_accuracy: 0.8571 - val_loss: 1.2254 +Epoch 81/100 +2/2 - 0s - 23ms/step - accuracy: 0.8537 - loss: 1.3219 - val_accuracy: 0.8571 - val_loss: 1.2223 +Epoch 82/100 +2/2 - 0s - 25ms/step - accuracy: 0.8537 - loss: 1.3559 - val_accuracy: 0.8571 - val_loss: 1.2190 +Epoch 83/100 +2/2 - 0s - 24ms/step - accuracy: 0.8537 - loss: 1.3289 - val_accuracy: 0.8571 - val_loss: 1.2159 +Epoch 84/100 +2/2 - 0s - 23ms/step - accuracy: 0.8537 - loss: 1.1656 - val_accuracy: 0.8571 - val_loss: 1.2123 +Epoch 85/100 +2/2 - 0s - 23ms/step - accuracy: 0.8537 - loss: 1.2954 - val_accuracy: 0.8571 - val_loss: 1.2088 +Epoch 86/100 +2/2 - 0s - 25ms/step - accuracy: 0.8537 - loss: 1.2298 - val_accuracy: 0.8571 - val_loss: 1.2058 +Epoch 87/100 +2/2 - 0s - 24ms/step - accuracy: 0.8537 - loss: 1.3323 - val_accuracy: 0.8571 - val_loss: 1.2037 +Epoch 88/100 +2/2 - 0s - 23ms/step - accuracy: 0.8537 - loss: 1.3758 - val_accuracy: 0.8571 - val_loss: 1.2010 +Epoch 89/100 +2/2 - 0s - 22ms/step - accuracy: 0.8537 - loss: 1.3587 - val_accuracy: 0.8571 - val_loss: 1.1981 +Epoch 90/100 +2/2 - 0s - 22ms/step - accuracy: 0.8537 - loss: 1.3240 - val_accuracy: 0.8571 - val_loss: 1.1954 +Epoch 91/100 +2/2 - 0s - 21ms/step - accuracy: 0.8537 - loss: 1.2884 - val_accuracy: 0.8571 - val_loss: 1.1922 +Epoch 92/100 +2/2 - 0s - 21ms/step - accuracy: 0.8537 - loss: 1.3293 - val_accuracy: 0.8571 - val_loss: 1.1901 +Epoch 93/100 +2/2 - 0s - 23ms/step - accuracy: 0.8537 - loss: 1.2706 - val_accuracy: 0.8571 - val_loss: 1.1879 +Epoch 94/100 +2/2 - 0s - 27ms/step - accuracy: 0.8537 - loss: 1.2715 - val_accuracy: 0.8571 - val_loss: 1.1848 +Epoch 95/100 +2/2 - 0s - 25ms/step - accuracy: 0.8537 - loss: 1.2628 - val_accuracy: 0.8571 - val_loss: 1.1818 +Epoch 96/100 +2/2 - 0s - 23ms/step - accuracy: 0.8537 - loss: 1.2770 - val_accuracy: 0.8571 - val_loss: 1.1786 +Epoch 97/100 +2/2 - 0s - 24ms/step - accuracy: 0.8537 - loss: 1.3039 - val_accuracy: 0.8571 - val_loss: 1.1762 +Epoch 98/100 +2/2 - 0s - 26ms/step - accuracy: 0.8537 - loss: 1.2908 - val_accuracy: 0.8571 - val_loss: 1.1731 +Epoch 99/100 +2/2 - 0s - 23ms/step - accuracy: 0.8537 - loss: 1.3510 - val_accuracy: 0.8571 - val_loss: 1.1707 +Epoch 100/100 +2/2 - 0s - 25ms/step - accuracy: 0.8537 - loss: 1.2447 - val_accuracy: 0.8571 - val_loss: 1.1683 +1/1 - 0s - 26ms/step - accuracy: 0.8571 - loss: 1.1683 +[0.6428571343421936, 0.7142857313156128, 0.8214285969734192, 0.8214285969734192, 0.8214285969734192, 0.8214285969734192, 0.8214285969734192, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064, 0.8571428656578064] +Dokładność testowa: 85.71% + 1/1 ━━━━━━━━━━━━━━━━━━━━ 0s 16ms/step - accuracy: 0.8571 - loss: 1.1683 1/1 ━━━━━━━━━━━━━━━━━━━━ 0s 19ms/step - accuracy: 0.8571 - loss: 1.1683 diff --git a/my_runs/1/info.json b/my_runs/1/info.json new file mode 100644 index 0000000..7b6f1a0 --- /dev/null +++ b/my_runs/1/info.json @@ -0,0 +1,15 @@ +{ + "dropout_layer_value": 0.4, + "num_epochs": 100, + "training_ts": { + "__reduce__": [ + { + "py/type": "datetime.datetime" + }, + [ + "B+gGCQ06DAHl8g==" + ] + ], + "py/object": "datetime.datetime" + } +} \ No newline at end of file diff --git a/my_runs/1/metrics.json b/my_runs/1/metrics.json new file mode 100644 index 0000000..e1b52a5 --- /dev/null +++ b/my_runs/1/metrics.json @@ -0,0 +1,248 @@ +{ + "test accuracy": { + "steps": [ + 0 + ], + "timestamps": [ + "2024-06-09T11:58:17.715614" + ], + "values": [ + 0.8571428656578064 + ] + }, + "test loss": { + "steps": [ + 0 + ], + "timestamps": [ + "2024-06-09T11:58:17.715614" + ], + "values": [ + 1.168319582939148 + ] + }, + "train accuracy": { + "steps": [ + 0 + ], + "timestamps": [ + "2024-06-09T11:58:17.715614" + ], + "values": [ + [ + 0.4390243887901306, + 0.3658536672592163, + 0.3658536672592163, + 0.4390243887901306, + 0.6341463327407837, + 0.6097561120986938, + 0.6097561120986938, + 0.6097561120986938, + 0.5365853905677795, + 0.5609756112098694, + 0.7560975551605225, + 0.5853658318519592, + 0.707317054271698, + 0.7560975551605225, + 0.8048780560493469, + 0.707317054271698, + 0.7560975551605225, + 0.8536585569381714, + 0.8048780560493469, + 0.8048780560493469, + 0.7560975551605225, + 0.7560975551605225, + 0.7804877758026123, + 0.8536585569381714, + 0.8292682766914368, + 0.8292682766914368, + 0.8292682766914368, + 0.7804877758026123, + 0.8292682766914368, + 0.8536585569381714, + 0.8536585569381714, + 0.8536585569381714, + 0.8292682766914368, + 0.8536585569381714, + 0.8292682766914368, + 0.8292682766914368, + 0.8048780560493469, + 0.8536585569381714, + 0.8536585569381714, + 0.8292682766914368, + 0.8536585569381714, + 0.8536585569381714, + 0.8292682766914368, + 0.8536585569381714, + 0.8048780560493469, + 0.8536585569381714, + 0.8536585569381714, + 0.8536585569381714, + 0.8536585569381714, + 0.8536585569381714, + 0.8536585569381714, + 0.8292682766914368, + 0.8536585569381714, + 0.8536585569381714, + 0.8536585569381714, + 0.8536585569381714, + 0.8536585569381714, + 0.8536585569381714, + 0.8536585569381714, + 0.8536585569381714, + 0.8536585569381714, + 0.8536585569381714, + 0.8536585569381714, + 0.8536585569381714, + 0.8536585569381714, + 0.8536585569381714, + 0.8536585569381714, + 0.8536585569381714, + 0.8536585569381714, + 0.8536585569381714, + 0.8536585569381714, + 0.8536585569381714, + 0.8536585569381714, + 0.8536585569381714, + 0.8536585569381714, + 0.8536585569381714, + 0.8536585569381714, + 0.8536585569381714, + 0.8536585569381714, + 0.8536585569381714, + 0.8536585569381714, + 0.8536585569381714, + 0.8536585569381714, + 0.8536585569381714, + 0.8536585569381714, + 0.8536585569381714, + 0.8536585569381714, + 0.8536585569381714, + 0.8536585569381714, + 0.8536585569381714, + 0.8536585569381714, + 0.8536585569381714, + 0.8536585569381714, + 0.8536585569381714, + 0.8536585569381714, + 0.8536585569381714, + 0.8536585569381714, + 0.8536585569381714, + 0.8536585569381714, + 0.8536585569381714 + ] + ] + }, + "train loss": { + "steps": [ + 0 + ], + "timestamps": [ + "2024-06-09T11:58:17.715614" + ], + "values": [ + [ + 2.1214587688446045, + 2.0694034099578857, + 2.1309335231781006, + 2.0288915634155273, + 1.9653569459915161, + 2.038301706314087, + 2.0237786769866943, + 1.9397151470184326, + 1.9641424417495728, + 1.9581079483032227, + 1.9044394493103027, + 1.9391545057296753, + 1.871816873550415, + 1.8057291507720947, + 1.8354332447052002, + 1.8500674962997437, + 1.7831162214279175, + 1.7120033502578735, + 1.8020172119140625, + 1.7531274557113647, + 1.7807929515838623, + 1.7794438600540161, + 1.6697275638580322, + 1.6813563108444214, + 1.6687272787094116, + 1.7051959037780762, + 1.6146538257598877, + 1.6207478046417236, + 1.5964454412460327, + 1.6258116960525513, + 1.5512006282806396, + 1.6674073934555054, + 1.584828495979309, + 1.5467300415039062, + 1.6243640184402466, + 1.527516484260559, + 1.710520625114441, + 1.613712191581726, + 1.57256281375885, + 1.5344923734664917, + 1.568911075592041, + 1.4239144325256348, + 1.5922287702560425, + 1.500575304031372, + 1.4858384132385254, + 1.4768829345703125, + 1.4210824966430664, + 1.4060230255126953, + 1.474118709564209, + 1.4988770484924316, + 1.4718369245529175, + 1.4711825847625732, + 1.3657357692718506, + 1.3846684694290161, + 1.376619815826416, + 1.4274654388427734, + 1.4709089994430542, + 1.422576904296875, + 1.449120283126831, + 1.4165239334106445, + 1.3913618326187134, + 1.3560173511505127, + 1.3432062864303589, + 1.3832975625991821, + 1.4476934671401978, + 1.3054296970367432, + 1.283986210823059, + 1.4320200681686401, + 1.4165103435516357, + 1.3209712505340576, + 1.3698146343231201, + 1.3679074048995972, + 1.3647143840789795, + 1.3087888956069946, + 1.3304280042648315, + 1.2813243865966797, + 1.3476758003234863, + 1.388467788696289, + 1.2938565015792847, + 1.3508087396621704, + 1.3219486474990845, + 1.3559297323226929, + 1.3289012908935547, + 1.1655505895614624, + 1.295440435409546, + 1.2297943830490112, + 1.33234703540802, + 1.3758138418197632, + 1.3586888313293457, + 1.324022650718689, + 1.2884330749511719, + 1.329307198524475, + 1.2706291675567627, + 1.2714873552322388, + 1.2628101110458374, + 1.2770464420318604, + 1.3039370775222778, + 1.290822148323059, + 1.3510299921035767, + 1.2447413206100464 + ] + ] + } +} \ No newline at end of file diff --git a/my_runs/1/model.keras b/my_runs/1/model.keras new file mode 100644 index 0000000000000000000000000000000000000000..50ff287ce7240bf2afdd89e549d3e4d1946aa38a GIT binary patch literal 20856 zcmeHP3vd(189v!OR33?G2)ID85J?aWII@kwfVDEjG!8IS17U{Y5&4kLw%{x&x)a!? z6epA!CvAWxeVAmDP@1Mr0)c@(++wdNZ5hMl!O&zvNSZjIP2EY8Ntm`1NJzSSyT5F$ zEX$8zJGRc_+uPlL_kY~C|Nq}gI@j{_VdEusOp_K(_~1q-snGwij0{(f`qelPBOW^ObWxl>u*9Rp?9xN6NsX%9`f(%aum+YUSRr z;@5q$YV;Xn3Hf#_5NI(3dPw&96V=E(>$h74 z5wdw-oFsSogJFFI9Vke$N!C3TZmmK#*sS-5Rdu!{Z({=1T2KyYRGuy}NkoJ&L?$dN7q;SCSgr+Z}Q2Mh?}0=Q@^6RePk%Pfxhn$)6rWeTD)DqW=I=@fIHPdD4?(L`^o~L zNL?}P2Uvusa|(G`527jR>O|&0XGx~ z2Ew{ zpBM|ey!;9_UWet^E%*-11!(xzZZ1ITL1N>d7AVsz<*HN&jLTFY*2Ko4co*8Pf6#_N zx&}8!S8SpvNeGBBI>`H?BfAi439Tnk8T6@2`@kFs>b^>zg;=$6FU6}G&}h15^^o{j z9tin7vg-DbG)E&THwm)UK;zz~_{uAc{Euf5lE9g@&a~g$ z+<1p1(f5?R~SsG?g88zed1wZI`dO)4kB&n2N zn0Q&s%lb5SVUp;dBxUmUaeRJ}?(S7fmr8^Z=7*9GlUjzeD>F2t${}TelBJ7XRN=K; zYa1_{InH=h+De`s={3taB`Lb5IRW6*4g->uiSS#J&IVR+Tr&K&tq&2uN!X;i2e}1A z{DL2=<@f>h0iOogp|Ka`>XqLvHc9DGuWlR3CX4dn_2Ac7JNW6E1+jGXyxzQswS)W? zz7AkMaa}O#ul_laA8-lwOBToK$sQj)EXFx?8fq30{XhXXz&^hIz#hitBdPG69LcD= zbD&PH*umO!jU0=%SXgu1IT3y{Smr0$IWCjo_b5#w$pP!1)l%o7VgV7q91fC;;|J^j zc~~?qKt~Tt7RQE1ao5V%3(PO#7~)v{6vhzf&hh$3Q&~IYfhV(L=~{W+%oiHy*>ZC% zop}PQ+c-lDt4g2Dh^4degtcfEC+MdlksrvsfeBQ=6;0*`DsE!*V3&SUC&jj(EFk&; zSZ`|uK-jr_5U*ez7WGhW<_AnsKkyO#XqFr>uZUyTn>vdYFiFFi&A8!A5_3?dG}ge= z`ly}b>zT&*ah5EyXG9N+G|?@RDW`;@3NHGkJp9Zh zB;T&$WqL72hjyax{o{UYd*cPX{;7-T_^x6+V~g&%=fXFgd-nevAN=h$Ty?;Se^mS? zs+;*UWF7uCt~~ZwTrK7>}f_Mu;wq&u&cyo(ub)1vnLWL)QJ z!ZJDvUr6Yje8}|1YbyGz)j``5#d|StNfz%*C)3$dc|Wtl&^wFq9{f`Y1b%LkFFWr~ zgx_Y)w{;R@%SQL>Y7c}&;6!X;s9Jjzn#Bnd#4cg&} z{DAm8pDd2qtjt7k{|BwXQi^qG0nraY9LvaxCGV>^wvY2YI*sv9hToc786C(SD|MZ~ji(kBoGtoQf^oYwS`zIeX)t@s*P-p^L@yFBHyJtSqyeHAD z=Q8lTYqBG&;SgT&*_B)Wi7R4`0A9d*4TEPMt(m;~v1R<@@o~_m`vP zyX#QPzxEgE+s>ogcUGfyi#?6CGxy^koRH9{I$P6)(XGgy;YY9i`BQXi{t>*=@`t8% z=e~iyeqKS5tO$DZ`)}d*_T(e)ybCDn#S6&syGQYZvyY)$|MobZyhL@}`N(a~$)Eoh zzvGc2OaGdWpFTMmot>E;`Dwv!^wN7D<7FrQiC>C5_xra>HsIMetZ#beKyA|ksiEnX z<0J9x{7Z5BLc!^s-vIFY&9`+(BzCAAmS!g9m6ns;}Z$J?Gn;#~3HWZ{~bP2Xc3) zWU6t{T0q1v>LQTc|{y|bNV4e9{>DmGjtdCh1hp( z+CGPede8TT>gF*c*9f-m-+iI`Havo!a*mA5D!&V_dZGo}x9&oZzWzMAy8Fh+!VzZY z<7baM9;+McoY}AwJ@Mi$^q-bjnl6u-jyE0tB|2kL(VFLHI$ym!1^Zr}9$EH_EWCHL z39Tx;$#E(D6rOZt1G+Nx3_keu&c^fhacIWf$I-!!W1P<5vyL2_*=hUx?znxK+I+!0 zLKXaOX@Sr&#H;bRj@_35{6g7t90&O?RNzj4;cMYyK$JD7W;icxn1VfLzl-nr_S{Drj94D@tGGuN8%iZ!KbFGW?b