-
Notifications
You must be signed in to change notification settings - Fork 3
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Installation and reproducability issues #1
Comments
Some further observations (on CORA). The code for the 80/10/10 split (if mode) seems deterministic and not random? This would explain the weird behavior with the test loss significantly higher than the validation loss. Why does
For the "20 nodes per class" split I get
I also report the maximum test accuracy over all runs because it's something computed in |
Hello @yzimmermann 👋🏻, Thank you for raising the issue with all the questions. I'll address them in the same order:
Should you have further questions, please feel free to reach out! Cheers 🍻 P.S.: Should some of your questions from the previous messages haven't been answered, then please point them out. |
I'm trying to install this library to perform some experiments and have run into a couple of issues.
I followed the instructions and installed via
install_script.sh
. However, when I runpython scripts/execute_experiments.py -h
I get:I don't see any references to
RationalPowerMeanModel
at https://github.com/k4ntz/activation-functions, which seems to install a differentactivations
package.To still run some experiments, even without rational activations, I commented out the relevant imports and ran the following config:
Out of the box, this seems to produce a random 80/10/10 split and the training output looks like this. Why is what you call the "validation accuracy" consistently (over multiple seeds) lower than the "test accuracy"? Now, when I change the number of clusters to 1 I get this. The test loss seems equally good, the validation loss is still consistently lower. In my understanding, this should behave like
"with_clusters": false
. However, when I do change this in the config file, I get this. The accuracy is suddenly substantially lower, but the split also seems to have automatically changed to "20 nodes per class" (140/500/1000).In the data loading code I found this (overwriting the default split if mode)
and for another dataset e.g.
mode
seems to be set inexecute_experiments.py
What is going on here? What is the purpose of
mode
and the different splits?Thanks in advance for your help and clarifications!
The text was updated successfully, but these errors were encountered: