Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

As deployer/Maintainer I should be able to improve Handwritten Digits model Accuracy by experimenting with open and synthetic datasets - Iteration 3 #51

Open
dileep-gadiraju opened this issue Nov 23, 2022 · 2 comments
Assignees
Labels
enhancement New feature or request SARAL_SDK

Comments

@dileep-gadiraju
Copy link
Collaborator

As deployer/Maintainer I should be able to improve Handwritten Digits Accuracy by experimenting with open and synthetic datasets - Iteration 3. Expected accuracy improvement is approximately 3% to 5%.

@dileep-gadiraju dileep-gadiraju added the enhancement New feature or request label Nov 23, 2022
@dileep-gadiraju dileep-gadiraju changed the title As deployer/Maintainer I should be able to improve Handwritten Digits Accuracy by experimenting with open and synthetic datasets - Iteration 3 As deployer/Maintainer I should be able to improve Handwritten Digits model Accuracy by experimenting with open and synthetic datasets - Iteration 3 Nov 23, 2022
@venky3692 venky3692 moved this from New to 🏗 In progress in Saral v1.5 Project Dashboard Dec 12, 2022
@venky3692 venky3692 self-assigned this Dec 13, 2022
@venky3692
Copy link

venky3692 commented Dec 13, 2022

@dileep-gadiraju (Dec 13) - Changed and tested the code for save model checkpoint (only save weights parameter). Trained the model for 5 epochs with existing digits dataset with data augmentation. Inference done on the same test dataset.
Accuracy on NSIT dataset (~60%). Almost same to old checkpoint.
Accuracy on Not in MNIST dataset (~97%). 3% increase than old checkpoint.

@venky3692
Copy link

venky3692 commented Dec 14, 2022

@dileep-gadiraju (Dec 14)

  • Tested new model on the existing dataset. Gives same results. Max 0.3% - 05% difference in accuracy for some classes
  • Analysed the miss-classifications on the NSIT dataset with images drawn in the notebook. Seems like the model just fails to understand the variations between different digits.
  • To segregate training images (NSIT dataset) for the model, tried pixel counting of (0,0,0) pixels to eliminate darkened numbers, unwanted lines around numbers.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request SARAL_SDK
Projects
Status: 🏗 In progress
Development

When branches are created from issues, their pull requests are automatically linked.

2 participants