Neural Net Model Comparisons

This is a continuation of my last post on building a web scaper to use with TensorFlow Playground. I described how I wrote the code for my web scraper using Selenium, so I could run trials with different conditions for different configurations and saving the results.

Below is a table of the data I collected for the MYBEST model with my assessments on it’s success for each trial. To judge for success the test loss was compared against 0.15 where test loss values below 0.15 were considered a success and all others were failures. For determining overfit, if the test loss was more than 0.1 larger than the training loss, I classified it as an overfitted model.

ModelProblemQualityOutcomeOverfit?Time TakenTraining LossTest Loss
MYBESTgaussQ-BESTSuccessno3.7613899707794190.0010.001
MYBESTxorQ-BESTSuccessno3.79232692718505860.0020.004
MYBESTcircleQ-BESTSuccessno3.7924571037292480.0040.004
MYBESTspiralQ-BESTFailureyes300.3520.494
MYBESTreg-planeQ-BESTSuccessno1.334413290023803700
MYBESTreg-gaussQ-BESTSuccessno13.5403048992156980.0210.027
MYBESTgaussQ-WORSTSuccessno300.1010.144
MYBESTxorQ-WORSTFailureyes300.1730.285
MYBESTcircleQ-WORSTFailureyes300.2880.419
MYBESTspiralQ-WORSTFailureno300.4580.503
MYBESTreg-planeQ-WORSTSuccessno6.1019444465637210.0260.032
MYBESTreg-gaussQ-WORSTSuccessno6.1439721584320070.0370.043
MYBESTgaussQ-OKSuccessno6.1407001018524170.0160.038
MYBESTxorQ-OKSuccessno11.0033617019653320.0680.066
MYBESTcircleQ-OKSuccessno8.5884611610.0740.084
MYBESTspiralQ-OKFailureno300.4140.5
MYBESTreg-planeQ-OKSuccessno6.3356015682220460.0050.005
MYBESTreg-gaussQ-OKSuccessno23.5130093097686770.0270.031

The models were successful at solving all the regression problems. However, none of the models were able to solve all the data sets since none of the models were able to successfully solve the spiral. The spiral was the hardest data set with circle and then exclusive or as the next hardest data sets to solve. There were no obvious patterns in the qualities leading to performance. MYBEST had the lowest average test loss among the models and had a higher count of successes. There are general observations, but no hard rules. In general, the top performing successful test losses had 2 hidden layers.

Further Comparisons

I did tests with a new model which I dubbed DNN to be my version of a deep neural net. The configuration is as shown below.

  • activation = relu
  • networkShape = 5,5,5
  • x and y inputs

I ran some more trials with this model and with the MYBEST model looking at the classification problem using the spiral dataset since that is the one that seemed to be the most difficult for each model. I looked at is how they compared in performance with different amounts of noise across increasing percentages of training data. Below are some rough graphs I made using Excel to compare the two models with 9 trials for each model for 0 percent noise and 30 percent noise.

Performance Comparison of Percent of Training Data with 0 Noisee Performance Comparison of Percent of Training Data with 30 Noise

The DNN model performed better than the MYBEST model in every condition. Very ironic considering the model is named “MYBEST” and yet it was not very good.