Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

from-151-Ternary-dense-try-different-normalization-to-optimize-outcome #152

Open
david-thrower opened this issue Jun 13, 2024 · 1 comment

Comments

@david-thrower
Copy link
Owner

Kind of issue: Feature development

Issue described: We have a successful implementation of a Ternary replacement for Dense layers. The metrics are not quite what we want on some problems.

One possible issue is the default batch normalization behavior. Perhaps Dropout will work better or layer normalization.

@david-thrower
Copy link
Owner Author

david-thrower commented Jun 13, 2024

Tried thus far:

  • No normalization nor dropout: Results: Basically no good RMSE on tabular = Mean value
  • Layer Normalization: Some improvement over no normalization but worse than default batch normalization on tabular. However, exceptional results on Ham / Spam problem. The disparity between the results on tabular and text classification is strange. It may be that the hyperparameter range for tabular are just way off for the Ternary neural network. It could also be that we should make a different node for Ternary and Dense and continue using Dense with batch normalization for tabular and use Ternary with Layer Normalization for NLP tasks. There is probably something that will make both work.

To try:

  • Dropout .10
  • Dropout 0.5
  • Dropout 0.8

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

When branches are created from issues, their pull requests are automatically linked.

1 participant