|
875 | 875 | },
|
876 | 876 | {
|
877 | 877 | "cell_type": "code",
|
878 |
| - "execution_count": null, |
| 878 | + "execution_count": 43, |
879 | 879 | "metadata": {},
|
880 | 880 | "outputs": [
|
881 | 881 | {
|
|
936 | 936 | "Step 3500: Loss on training set : 0.015484\n",
|
937 | 937 | "Step 3600: Loss on training set : 0.007086\n",
|
938 | 938 | "Step 3700: Loss on training set : 0.045831\n",
|
939 |
| - "Step 3800: Loss on training set : 0.058367\n" |
| 939 | + "Step 3800: Loss on training set : 0.058367\n", |
| 940 | + "Step 3900: Loss on training set : 0.042314\n", |
| 941 | + "Step 4000: Loss on training set : 0.097755\n", |
| 942 | + "Step 4100: Loss on training set : 0.030364\n", |
| 943 | + "Step 4200: Loss on training set : 0.019168\n", |
| 944 | + "Step 4300: Loss on training set : 0.019495\n", |
| 945 | + "Step 4400: Loss on training set : 0.030664\n", |
| 946 | + "Step 4500: Loss on training set : 0.033287\n", |
| 947 | + "Step 4600: Loss on training set : 0.105648\n", |
| 948 | + "Step 4700: Loss on training set : 0.009519\n", |
| 949 | + "Step 4800: Loss on training set : 0.009824\n", |
| 950 | + "Step 4900: Loss on training set : 0.002434\n", |
| 951 | + "Step 5000: Loss on training set : 0.002829\n", |
| 952 | + "Step 5100: Loss on training set : 0.003514\n", |
| 953 | + "Step 5200: Loss on training set : 0.110153\n", |
| 954 | + "Step 5300: Loss on training set : 0.006671\n", |
| 955 | + "Step 5400: Loss on training set : 0.022709\n", |
| 956 | + "Step 5500: Loss on training set : 0.004206\n", |
| 957 | + "Step 5600: Loss on training set : 0.021019\n", |
| 958 | + "Step 5700: Loss on training set : 0.010719\n", |
| 959 | + "Step 5800: Loss on training set : 0.004459\n", |
| 960 | + "Step 5900: Loss on training set : 0.025301\n", |
| 961 | + "Step 6000: Loss on training set : 0.004324\n", |
| 962 | + "Step 6100: Loss on training set : 0.005369\n", |
| 963 | + "Step 6200: Loss on training set : 0.047895\n", |
| 964 | + "Step 6300: Loss on training set : 0.037086\n", |
| 965 | + "Step 6400: Loss on training set : 0.054543\n", |
| 966 | + "Step 6500: Loss on training set : 0.048353\n", |
| 967 | + "Step 6600: Loss on training set : 0.010139\n", |
| 968 | + "Step 6700: Loss on training set : 0.023344\n", |
| 969 | + "Step 6800: Loss on training set : 0.005961\n", |
| 970 | + "Step 6900: Loss on training set : 0.027603\n", |
| 971 | + "Step 7000: Loss on training set : 0.001533\n", |
| 972 | + "Step 7100: Loss on training set : 0.031711\n", |
| 973 | + "Step 7200: Loss on training set : 0.005685\n", |
| 974 | + "Step 7300: Loss on training set : 0.003476\n", |
| 975 | + "Step 7400: Loss on training set : 0.004660\n", |
| 976 | + "Step 7500: Loss on training set : 0.065238\n", |
| 977 | + "Step 7600: Loss on training set : 0.008798\n", |
| 978 | + "Step 7700: Loss on training set : 0.024648\n", |
| 979 | + "Step 7800: Loss on training set : 0.025574\n", |
| 980 | + "Step 7900: Loss on training set : 0.020188\n", |
| 981 | + "Step 8000: Loss on training set : 0.018603\n", |
| 982 | + "Step 8100: Loss on training set : 0.002296\n", |
| 983 | + "Step 8200: Loss on training set : 0.051427\n", |
| 984 | + "Step 8300: Loss on training set : 0.005059\n", |
| 985 | + "Step 8400: Loss on training set : 0.003244\n", |
| 986 | + "Step 8500: Loss on training set : 0.003312\n", |
| 987 | + "Step 8600: Loss on training set : 0.018318\n", |
| 988 | + "Step 8700: Loss on training set : 0.020676\n", |
| 989 | + "Step 8800: Loss on training set : 0.004160\n", |
| 990 | + "Step 8900: Loss on training set : 0.004872\n", |
| 991 | + "Step 9000: Loss on training set : 0.001758\n", |
| 992 | + "Step 9100: Loss on training set : 0.002270\n", |
| 993 | + "Step 9200: Loss on training set : 0.000740\n", |
| 994 | + "Step 9300: Loss on training set : 0.002610\n", |
| 995 | + "Step 9400: Loss on training set : 0.003372\n", |
| 996 | + "Step 9500: Loss on training set : 0.002058\n", |
| 997 | + "Step 9600: Loss on training set : 0.000575\n", |
| 998 | + "Step 9700: Loss on training set : 0.008625\n", |
| 999 | + "Step 9800: Loss on training set : 0.010724\n", |
| 1000 | + "Step 9900: Loss on training set : 0.025758\n", |
| 1001 | + "Step 10000: Loss on training set : 0.006348\n", |
| 1002 | + "Step 10100: Loss on training set : 0.002577\n", |
| 1003 | + "Step 10200: Loss on training set : 0.012478\n", |
| 1004 | + "Step 10300: Loss on training set : 0.006424\n", |
| 1005 | + "Step 10400: Loss on training set : 0.007878\n", |
| 1006 | + "Step 10500: Loss on training set : 0.010238\n", |
| 1007 | + "Step 10600: Loss on training set : 0.011099\n", |
| 1008 | + "Step 10700: Loss on training set : 0.000476\n", |
| 1009 | + "Step 10800: Loss on training set : 0.000640\n", |
| 1010 | + "Step 10900: Loss on training set : 0.001908\n", |
| 1011 | + "Step 11000: Loss on training set : 0.002969\n", |
| 1012 | + "Step 11100: Loss on training set : 0.002400\n", |
| 1013 | + "Step 11200: Loss on training set : 0.005269\n", |
| 1014 | + "Step 11300: Loss on training set : 0.018153\n", |
| 1015 | + "Step 11400: Loss on training set : 0.002701\n", |
| 1016 | + "Step 11500: Loss on training set : 0.016919\n", |
| 1017 | + "Step 11600: Loss on training set : 0.001562\n", |
| 1018 | + "Step 11700: Loss on training set : 0.001536\n", |
| 1019 | + "Step 11800: Loss on training set : 0.001691\n", |
| 1020 | + "Step 11900: Loss on training set : 0.001171\n", |
| 1021 | + "Step 12000: Loss on training set : 0.007421\n", |
| 1022 | + "Step 12100: Loss on training set : 0.002470\n", |
| 1023 | + "Step 12200: Loss on training set : 0.002226\n", |
| 1024 | + "Step 12300: Loss on training set : 0.025579\n", |
| 1025 | + "Step 12400: Loss on training set : 0.007481\n", |
| 1026 | + "Step 12500: Loss on training set : 0.011208\n", |
| 1027 | + "Step 12600: Loss on training set : 0.005309\n", |
| 1028 | + "Step 12700: Loss on training set : 0.001684\n", |
| 1029 | + "Step 12800: Loss on training set : 0.010187\n", |
| 1030 | + "Step 12900: Loss on training set : 0.001179\n", |
| 1031 | + "Step 13000: Loss on training set : 0.012636\n", |
| 1032 | + "Step 13100: Loss on training set : 0.004405\n", |
| 1033 | + "Step 13200: Loss on training set : 0.000556\n", |
| 1034 | + "Step 13300: Loss on training set : 0.001598\n", |
| 1035 | + "Step 13400: Loss on training set : 0.001437\n", |
| 1036 | + "Step 13500: Loss on training set : 0.001434\n", |
| 1037 | + "Step 13600: Loss on training set : 0.037155\n", |
| 1038 | + "Step 13700: Loss on training set : 0.001473\n", |
| 1039 | + "Step 13800: Loss on training set : 0.016352\n", |
| 1040 | + "Step 13900: Loss on training set : 0.000089\n", |
| 1041 | + "Step 14000: Loss on training set : 0.010603\n", |
| 1042 | + "Step 14100: Loss on training set : 0.000232\n", |
| 1043 | + "Step 14200: Loss on training set : 0.000834\n", |
| 1044 | + "Step 14300: Loss on training set : 0.000423\n", |
| 1045 | + "Step 14400: Loss on training set : 0.000145\n", |
| 1046 | + "Step 14500: Loss on training set : 0.000219\n", |
| 1047 | + "Step 14600: Loss on training set : 0.012533\n", |
| 1048 | + "Step 14700: Loss on training set : 0.000458\n", |
| 1049 | + "Step 14800: Loss on training set : 0.000817\n", |
| 1050 | + "Step 14900: Loss on training set : 0.002362\n", |
| 1051 | + "Step 15000: Loss on training set : 0.000349\n", |
| 1052 | + "Step 15100: Loss on training set : 0.012360\n", |
| 1053 | + "Step 15200: Loss on training set : 0.021442\n", |
| 1054 | + "Step 15300: Loss on training set : 0.007397\n", |
| 1055 | + "Step 15400: Loss on training set : 0.000550\n", |
| 1056 | + "Step 15500: Loss on training set : 0.001426\n", |
| 1057 | + "Step 15600: Loss on training set : 0.000598\n", |
| 1058 | + "Step 15700: Loss on training set : 0.033891\n", |
| 1059 | + "Step 15800: Loss on training set : 0.003439\n", |
| 1060 | + "Step 15900: Loss on training set : 0.001901\n", |
| 1061 | + "Step 16000: Loss on training set : 0.001585\n", |
| 1062 | + "Step 16100: Loss on training set : 0.008619\n", |
| 1063 | + "Step 16200: Loss on training set : 0.000461\n", |
| 1064 | + "Step 16300: Loss on training set : 0.001010\n", |
| 1065 | + "Step 16400: Loss on training set : 0.007105\n", |
| 1066 | + "Step 16500: Loss on training set : 0.000262\n", |
| 1067 | + "Step 16600: Loss on training set : 0.001249\n", |
| 1068 | + "Step 16700: Loss on training set : 0.000257\n", |
| 1069 | + "Step 16800: Loss on training set : 0.000844\n", |
| 1070 | + "Step 16900: Loss on training set : 0.014114\n" |
| 1071 | + ] |
| 1072 | + }, |
| 1073 | + { |
| 1074 | + "name": "stdout", |
| 1075 | + "output_type": "stream", |
| 1076 | + "text": [ |
| 1077 | + "Step 17000: Loss on training set : 0.001968\n", |
| 1078 | + "Step 17100: Loss on training set : 0.000176\n", |
| 1079 | + "Step 17200: Loss on training set : 0.001083\n", |
| 1080 | + "Step 17300: Loss on training set : 0.001023\n", |
| 1081 | + "Step 17400: Loss on training set : 0.002160\n", |
| 1082 | + "Step 17500: Loss on training set : 0.004066\n", |
| 1083 | + "Step 17600: Loss on training set : 0.002666\n", |
| 1084 | + "Step 17700: Loss on training set : 0.001996\n", |
| 1085 | + "Step 17800: Loss on training set : 0.001733\n", |
| 1086 | + "Step 17900: Loss on training set : 0.000166\n", |
| 1087 | + "Step 18000: Loss on training set : 0.000297\n", |
| 1088 | + "Step 18100: Loss on training set : 0.000313\n", |
| 1089 | + "Step 18200: Loss on training set : 0.005266\n", |
| 1090 | + "Step 18300: Loss on training set : 0.000845\n", |
| 1091 | + "Step 18400: Loss on training set : 0.002949\n", |
| 1092 | + "Step 18500: Loss on training set : 0.000892\n", |
| 1093 | + "Step 18600: Loss on training set : 0.003773\n", |
| 1094 | + "Step 18700: Loss on training set : 0.003751\n", |
| 1095 | + "Step 18800: Loss on training set : 0.000007\n", |
| 1096 | + "Step 18900: Loss on training set : 0.000056\n", |
| 1097 | + "Step 19000: Loss on training set : 0.000011\n", |
| 1098 | + "Step 19100: Loss on training set : 0.000058\n", |
| 1099 | + "Step 19200: Loss on training set : 0.001537\n", |
| 1100 | + "Step 19300: Loss on training set : 0.023998\n", |
| 1101 | + "Step 19400: Loss on training set : 0.000132\n", |
| 1102 | + "Step 19500: Loss on training set : 0.000506\n", |
| 1103 | + "Step 19600: Loss on training set : 0.000474\n", |
| 1104 | + "Step 19700: Loss on training set : 0.000561\n", |
| 1105 | + "Step 19800: Loss on training set : 0.000205\n", |
| 1106 | + "Step 19900: Loss on training set : 0.001069\n", |
| 1107 | + "Step 20000: Loss on training set : 0.000826\n", |
| 1108 | + "Loss on test set: 0.024786\n" |
940 | 1109 | ]
|
941 | 1110 | }
|
942 | 1111 | ],
|
|
0 commit comments