I am currently working with Flux.jl package for deep learning model implementation. I was wondering if some of you have had experience with some hyperparameters optimization packages (e.g. using hyperbands optimization algorithm) that can be nicely coupled with Flux.jl.
I guess it is has some half accidental similarity to hyperband in the sense that it tries to focus training efforts on promising candidates.
If one does not believe in parameter reuse it should be possible to build hyperband on it, e.g. by using elite selection of half the population size until there is only one candidate.
I’d like to ask you if you could put here an example using Hyperband and inside of it is a train_network. Please, could you do this for me? Thank you.
I am getting an error when trying to modify working example with RandomSample or LHSampler , but I get:
ERROR: LoadError: syntax: Global method definition around /Users/ies503/.julia/packages/Hyperopt/utzuO/src/samplers.jl:191 needs to be placed at the top level, or use "eval".
Stacktrace:
...
...
The macro @phyperopt works in the same way as @hyperopt but distributes all computation on available workers. The usual caveats apply, code must be loaded on all workers etc.
You make it sound so easy! I think this is VERY impressive.
Such an example would depend completely on what tools you use to train networks, as Hyperband requires you to save a meaningful state and also resume training from this state. You will find an example that shows how to do this using Optim.jl here
Thanks It’s really Julia making it easy. Distributing using Distributed is significantly easier than threading, as one does not have to worry about thread safety etc.
I found this comment very helpful in my attempt to use your package. Context like this would be much appreciated in the docs (readme). I would be happy to help write, but I don’t really understand enough to be of much use right now I think.