Skip to content

Commit b5ed9f7

Browse files
committed
TensorFlow Eager Execution Tests
1 parent 9433c19 commit b5ed9f7

File tree

1 file changed

+171
-2
lines changed

1 file changed

+171
-2
lines changed

deep-learning/tensor-flow-exercises/TensorFlow 1.5 Developer Version with Eager Execution - CPU Mac OSX.ipynb

+171-2
Original file line numberDiff line numberDiff line change
@@ -875,7 +875,7 @@
875875
},
876876
{
877877
"cell_type": "code",
878-
"execution_count": null,
878+
"execution_count": 43,
879879
"metadata": {},
880880
"outputs": [
881881
{
@@ -936,7 +936,176 @@
936936
"Step 3500: Loss on training set : 0.015484\n",
937937
"Step 3600: Loss on training set : 0.007086\n",
938938
"Step 3700: Loss on training set : 0.045831\n",
939-
"Step 3800: Loss on training set : 0.058367\n"
939+
"Step 3800: Loss on training set : 0.058367\n",
940+
"Step 3900: Loss on training set : 0.042314\n",
941+
"Step 4000: Loss on training set : 0.097755\n",
942+
"Step 4100: Loss on training set : 0.030364\n",
943+
"Step 4200: Loss on training set : 0.019168\n",
944+
"Step 4300: Loss on training set : 0.019495\n",
945+
"Step 4400: Loss on training set : 0.030664\n",
946+
"Step 4500: Loss on training set : 0.033287\n",
947+
"Step 4600: Loss on training set : 0.105648\n",
948+
"Step 4700: Loss on training set : 0.009519\n",
949+
"Step 4800: Loss on training set : 0.009824\n",
950+
"Step 4900: Loss on training set : 0.002434\n",
951+
"Step 5000: Loss on training set : 0.002829\n",
952+
"Step 5100: Loss on training set : 0.003514\n",
953+
"Step 5200: Loss on training set : 0.110153\n",
954+
"Step 5300: Loss on training set : 0.006671\n",
955+
"Step 5400: Loss on training set : 0.022709\n",
956+
"Step 5500: Loss on training set : 0.004206\n",
957+
"Step 5600: Loss on training set : 0.021019\n",
958+
"Step 5700: Loss on training set : 0.010719\n",
959+
"Step 5800: Loss on training set : 0.004459\n",
960+
"Step 5900: Loss on training set : 0.025301\n",
961+
"Step 6000: Loss on training set : 0.004324\n",
962+
"Step 6100: Loss on training set : 0.005369\n",
963+
"Step 6200: Loss on training set : 0.047895\n",
964+
"Step 6300: Loss on training set : 0.037086\n",
965+
"Step 6400: Loss on training set : 0.054543\n",
966+
"Step 6500: Loss on training set : 0.048353\n",
967+
"Step 6600: Loss on training set : 0.010139\n",
968+
"Step 6700: Loss on training set : 0.023344\n",
969+
"Step 6800: Loss on training set : 0.005961\n",
970+
"Step 6900: Loss on training set : 0.027603\n",
971+
"Step 7000: Loss on training set : 0.001533\n",
972+
"Step 7100: Loss on training set : 0.031711\n",
973+
"Step 7200: Loss on training set : 0.005685\n",
974+
"Step 7300: Loss on training set : 0.003476\n",
975+
"Step 7400: Loss on training set : 0.004660\n",
976+
"Step 7500: Loss on training set : 0.065238\n",
977+
"Step 7600: Loss on training set : 0.008798\n",
978+
"Step 7700: Loss on training set : 0.024648\n",
979+
"Step 7800: Loss on training set : 0.025574\n",
980+
"Step 7900: Loss on training set : 0.020188\n",
981+
"Step 8000: Loss on training set : 0.018603\n",
982+
"Step 8100: Loss on training set : 0.002296\n",
983+
"Step 8200: Loss on training set : 0.051427\n",
984+
"Step 8300: Loss on training set : 0.005059\n",
985+
"Step 8400: Loss on training set : 0.003244\n",
986+
"Step 8500: Loss on training set : 0.003312\n",
987+
"Step 8600: Loss on training set : 0.018318\n",
988+
"Step 8700: Loss on training set : 0.020676\n",
989+
"Step 8800: Loss on training set : 0.004160\n",
990+
"Step 8900: Loss on training set : 0.004872\n",
991+
"Step 9000: Loss on training set : 0.001758\n",
992+
"Step 9100: Loss on training set : 0.002270\n",
993+
"Step 9200: Loss on training set : 0.000740\n",
994+
"Step 9300: Loss on training set : 0.002610\n",
995+
"Step 9400: Loss on training set : 0.003372\n",
996+
"Step 9500: Loss on training set : 0.002058\n",
997+
"Step 9600: Loss on training set : 0.000575\n",
998+
"Step 9700: Loss on training set : 0.008625\n",
999+
"Step 9800: Loss on training set : 0.010724\n",
1000+
"Step 9900: Loss on training set : 0.025758\n",
1001+
"Step 10000: Loss on training set : 0.006348\n",
1002+
"Step 10100: Loss on training set : 0.002577\n",
1003+
"Step 10200: Loss on training set : 0.012478\n",
1004+
"Step 10300: Loss on training set : 0.006424\n",
1005+
"Step 10400: Loss on training set : 0.007878\n",
1006+
"Step 10500: Loss on training set : 0.010238\n",
1007+
"Step 10600: Loss on training set : 0.011099\n",
1008+
"Step 10700: Loss on training set : 0.000476\n",
1009+
"Step 10800: Loss on training set : 0.000640\n",
1010+
"Step 10900: Loss on training set : 0.001908\n",
1011+
"Step 11000: Loss on training set : 0.002969\n",
1012+
"Step 11100: Loss on training set : 0.002400\n",
1013+
"Step 11200: Loss on training set : 0.005269\n",
1014+
"Step 11300: Loss on training set : 0.018153\n",
1015+
"Step 11400: Loss on training set : 0.002701\n",
1016+
"Step 11500: Loss on training set : 0.016919\n",
1017+
"Step 11600: Loss on training set : 0.001562\n",
1018+
"Step 11700: Loss on training set : 0.001536\n",
1019+
"Step 11800: Loss on training set : 0.001691\n",
1020+
"Step 11900: Loss on training set : 0.001171\n",
1021+
"Step 12000: Loss on training set : 0.007421\n",
1022+
"Step 12100: Loss on training set : 0.002470\n",
1023+
"Step 12200: Loss on training set : 0.002226\n",
1024+
"Step 12300: Loss on training set : 0.025579\n",
1025+
"Step 12400: Loss on training set : 0.007481\n",
1026+
"Step 12500: Loss on training set : 0.011208\n",
1027+
"Step 12600: Loss on training set : 0.005309\n",
1028+
"Step 12700: Loss on training set : 0.001684\n",
1029+
"Step 12800: Loss on training set : 0.010187\n",
1030+
"Step 12900: Loss on training set : 0.001179\n",
1031+
"Step 13000: Loss on training set : 0.012636\n",
1032+
"Step 13100: Loss on training set : 0.004405\n",
1033+
"Step 13200: Loss on training set : 0.000556\n",
1034+
"Step 13300: Loss on training set : 0.001598\n",
1035+
"Step 13400: Loss on training set : 0.001437\n",
1036+
"Step 13500: Loss on training set : 0.001434\n",
1037+
"Step 13600: Loss on training set : 0.037155\n",
1038+
"Step 13700: Loss on training set : 0.001473\n",
1039+
"Step 13800: Loss on training set : 0.016352\n",
1040+
"Step 13900: Loss on training set : 0.000089\n",
1041+
"Step 14000: Loss on training set : 0.010603\n",
1042+
"Step 14100: Loss on training set : 0.000232\n",
1043+
"Step 14200: Loss on training set : 0.000834\n",
1044+
"Step 14300: Loss on training set : 0.000423\n",
1045+
"Step 14400: Loss on training set : 0.000145\n",
1046+
"Step 14500: Loss on training set : 0.000219\n",
1047+
"Step 14600: Loss on training set : 0.012533\n",
1048+
"Step 14700: Loss on training set : 0.000458\n",
1049+
"Step 14800: Loss on training set : 0.000817\n",
1050+
"Step 14900: Loss on training set : 0.002362\n",
1051+
"Step 15000: Loss on training set : 0.000349\n",
1052+
"Step 15100: Loss on training set : 0.012360\n",
1053+
"Step 15200: Loss on training set : 0.021442\n",
1054+
"Step 15300: Loss on training set : 0.007397\n",
1055+
"Step 15400: Loss on training set : 0.000550\n",
1056+
"Step 15500: Loss on training set : 0.001426\n",
1057+
"Step 15600: Loss on training set : 0.000598\n",
1058+
"Step 15700: Loss on training set : 0.033891\n",
1059+
"Step 15800: Loss on training set : 0.003439\n",
1060+
"Step 15900: Loss on training set : 0.001901\n",
1061+
"Step 16000: Loss on training set : 0.001585\n",
1062+
"Step 16100: Loss on training set : 0.008619\n",
1063+
"Step 16200: Loss on training set : 0.000461\n",
1064+
"Step 16300: Loss on training set : 0.001010\n",
1065+
"Step 16400: Loss on training set : 0.007105\n",
1066+
"Step 16500: Loss on training set : 0.000262\n",
1067+
"Step 16600: Loss on training set : 0.001249\n",
1068+
"Step 16700: Loss on training set : 0.000257\n",
1069+
"Step 16800: Loss on training set : 0.000844\n",
1070+
"Step 16900: Loss on training set : 0.014114\n"
1071+
]
1072+
},
1073+
{
1074+
"name": "stdout",
1075+
"output_type": "stream",
1076+
"text": [
1077+
"Step 17000: Loss on training set : 0.001968\n",
1078+
"Step 17100: Loss on training set : 0.000176\n",
1079+
"Step 17200: Loss on training set : 0.001083\n",
1080+
"Step 17300: Loss on training set : 0.001023\n",
1081+
"Step 17400: Loss on training set : 0.002160\n",
1082+
"Step 17500: Loss on training set : 0.004066\n",
1083+
"Step 17600: Loss on training set : 0.002666\n",
1084+
"Step 17700: Loss on training set : 0.001996\n",
1085+
"Step 17800: Loss on training set : 0.001733\n",
1086+
"Step 17900: Loss on training set : 0.000166\n",
1087+
"Step 18000: Loss on training set : 0.000297\n",
1088+
"Step 18100: Loss on training set : 0.000313\n",
1089+
"Step 18200: Loss on training set : 0.005266\n",
1090+
"Step 18300: Loss on training set : 0.000845\n",
1091+
"Step 18400: Loss on training set : 0.002949\n",
1092+
"Step 18500: Loss on training set : 0.000892\n",
1093+
"Step 18600: Loss on training set : 0.003773\n",
1094+
"Step 18700: Loss on training set : 0.003751\n",
1095+
"Step 18800: Loss on training set : 0.000007\n",
1096+
"Step 18900: Loss on training set : 0.000056\n",
1097+
"Step 19000: Loss on training set : 0.000011\n",
1098+
"Step 19100: Loss on training set : 0.000058\n",
1099+
"Step 19200: Loss on training set : 0.001537\n",
1100+
"Step 19300: Loss on training set : 0.023998\n",
1101+
"Step 19400: Loss on training set : 0.000132\n",
1102+
"Step 19500: Loss on training set : 0.000506\n",
1103+
"Step 19600: Loss on training set : 0.000474\n",
1104+
"Step 19700: Loss on training set : 0.000561\n",
1105+
"Step 19800: Loss on training set : 0.000205\n",
1106+
"Step 19900: Loss on training set : 0.001069\n",
1107+
"Step 20000: Loss on training set : 0.000826\n",
1108+
"Loss on test set: 0.024786\n"
9401109
]
9411110
}
9421111
],

0 commit comments

Comments
 (0)