Mike J Innes
aa4d221f8c
break all the things
2019-05-02 18:50:52 -07:00
Dhairya Gandhi
221670a2b1
Merge pull request #733 from thebhatman/expdecay-fix
...
Fixed ExpDecay
2019-05-01 18:58:37 +05:30
Hossein Pourbozorg
7f06b15f67
use https instead of http for web links
2019-04-25 11:04:03 +00:00
thebhatman
31a50ab16a
Fixed ExpDecay
2019-04-11 17:28:06 +05:30
Mike Innes
4cf43c0c41
simpler/nicer training loop
2019-02-28 14:58:42 +00:00
Mike J Innes
0f8a4a48c6
extend update! with an optimiser
2019-01-28 14:10:09 +00:00
Mike J Innes
0f2975d905
update -> apply
2019-01-28 13:59:23 +00:00
Mike J Innes
f6397e7358
Merge pull request #517 from FluxML/fix_adamw
...
Fix decay argument in ADAMW
2019-01-18 10:06:23 +00:00
Mike J Innes
f0d5624ed2
Merge pull request #493 from dhairyagandhi96/master
...
[WIP] New Optimiser Docs
2019-01-10 11:10:38 +00:00
Dhairya Gandhi
4291c1a833
pull master
2019-01-10 16:35:57 +05:30
Mike J Innes
e6f925f977
train docstring simplification
2019-01-10 11:05:21 +00:00
Dhairya Gandhi
f00e1cdedf
[docs] replace :stop with Flux.stop()
2019-01-10 16:34:07 +05:30
Dhairya Gandhi
7484c54f03
fix train! API syntax docstring
2019-01-08 00:32:55 +05:30
Dhairya Gandhi
e48268ff06
fix argument name in ADAMW
2018-12-12 16:47:42 +05:30
Dhairya Gandhi
1ea8c5a293
[WIP] add docstrings and doc improvements
2018-11-12 19:17:10 +05:30
Mike J Innes
5e572df557
Merge pull request #485 from dhairyagandhi96/master
...
Add call back
2018-11-08 13:18:17 +00:00
Dhairya Gandhi
392c3c942b
re-add removed call function
2018-11-08 18:44:57 +05:30
Mike J Innes
8042198475
Merge pull request #479 from dhairyagandhi96/master
...
Fix deprecations of optimisers
2018-11-05 13:01:59 +00:00
Joel Mason
29832aca92
Move some epsilons about
2018-11-02 22:59:04 +11:00
Dhairya Gandhi
5ec70fe29d
allow array parameters to old optimisers
2018-11-01 22:17:54 +05:30
Dhairya Gandhi
ca4e01ac26
use user defined decay in ADAMW
2018-11-01 15:58:40 +05:30
Dhairya Gandhi
58a6c3f225
fix deprecations
2018-11-01 15:02:00 +05:30
Mike J Innes
46049b9f44
tweak update rule
2018-10-31 16:08:18 +00:00
Mike J Innes
4a54d30cbf
correct SGD deprecation
2018-10-31 15:30:30 +00:00
Mike J Innes
bffaceee02
tweaks
2018-10-31 14:58:55 +00:00
Dhairya Gandhi
bebf4eb95f
fixed ExpDecay update! rule
2018-10-29 23:12:24 +05:30
Dhairya Gandhi
32ce2d78b8
fixed ExpDecay test
2018-10-27 19:53:06 +05:30
Dhairya Gandhi
815e8c206d
decay fixes
2018-10-27 19:26:42 +05:30
Dhairya Gandhi
1f0f2a5ac2
fixed DescentWeightDecay parameters
2018-10-11 10:21:29 +05:30
Dhairya Gandhi
d8394298bb
fix merge conflicts
2018-10-11 10:15:59 +05:30
Dhairya Gandhi
fe8c147f72
fixed weight decay definition
2018-10-11 10:07:16 +05:30
Mike Innes
bfe85e65f1
compose tweaks
2018-10-05 13:52:26 +01:00
Mike Innes
0f2019eba5
compose tweaks
2018-10-05 12:57:03 +01:00
Mike Innes
9bc9771a8d
tweaks
2018-10-05 12:43:03 +01:00
Mike Innes
4abe518599
newline fixes
2018-10-05 12:37:47 +01:00
Dhairya Gandhi
b661db3797
added deprecations and compose
2018-10-01 05:30:53 +05:30
Dhairya Gandhi
6665189ff1
added remaining optimizers and tests
2018-09-16 17:34:51 +05:30
Dhairya Gandhi
63bc71698b
updated tests
2018-09-14 20:32:56 +05:30
Dhairya Gandhi
d933f2079b
pulled tracker from upstream
2018-09-11 18:30:24 +05:30
Mike J Innes
a2d2d068aa
initial sketch
2018-08-28 17:55:59 +05:30
Mike Innes
2ca189bc96
newlines
2018-08-28 10:54:50 +01:00
Dhairya Gandhi
89bca2d98d
remove merge conflicts
2018-08-28 15:14:12 +05:30
Dhairya Gandhi
a964debd8a
fixed example in docs
2018-08-28 15:02:47 +05:30
Dhairya Gandhi
c035fe22d7
added deprecation warning
2018-08-24 13:08:03 +05:30
Dhairya Gandhi
2f1a9847fa
deprecate :stop from optimizers; housekeeping
2018-08-22 21:25:26 +05:30
Dhairya Gandhi
a7ad620f01
exporting stop
2018-08-22 00:33:30 +05:30
Dhairya Gandhi
3d11322d37
fixed docstring and not exporting stop
2018-08-22 00:29:07 +05:30
Dhairya Gandhi
ed044e2df7
changes as requested
2018-08-21 23:22:20 +05:30
Dhairya Gandhi
756207e782
added docs
2018-08-20 14:20:33 +05:30
Dhairya Gandhi
51578177a5
removed arguments from StopException
2018-08-20 14:08:23 +05:30
Dhairya Gandhi
df22bc5c8f
removed argument from stop function
2018-08-20 14:02:09 +05:30
Dhairya Gandhi
06db6ed314
housekeeping: fixing typo
2018-08-20 13:48:28 +05:30
Dhairya Gandhi
394b4167ce
moving stop to Optimise
2018-08-20 13:43:08 +05:30
Dhairya Gandhi
06aad375fc
properly importing functions
2018-08-20 13:35:55 +05:30
Dhairya Gandhi
e239eb1105
properly importing functions
2018-08-20 13:30:05 +05:30
Dhairya Gandhi
1228e9c5e2
removed include statement
2018-08-19 22:55:14 +05:30
Dhairya Gandhi
9c98272cf0
catching exception
2018-08-19 17:38:00 +05:30
Dhairya Gandhi
257e2a7d2e
checking exception
2018-08-19 17:11:11 +05:30
Dhairya Gandhi
5c42c8689c
printing expception
2018-08-19 17:04:31 +05:30
Dhairya Gandhi
fbd82a6925
added end
2018-08-19 15:19:45 +05:30
Dhairya Gandhi
8229c8e045
modified training loop
2018-08-19 15:17:07 +05:30
Dhairya Gandhi
2aa057ec08
fixed throwing exception
2018-08-19 14:54:54 +05:30
Dhairya Gandhi
887bfad312
returning :stop
2018-08-18 08:28:47 +05:30
Dhairya Gandhi
65a5ecccd2
returning
2018-08-18 08:24:49 +05:30
Dhairya Gandhi
999b00b64d
fixed typo
2018-08-17 19:45:10 +05:30
Dhairya Gandhi
0524964400
fixed typo
2018-08-17 19:40:48 +05:30
Dhairya Gandhi
24a3bce495
added stop to break training loop
2018-08-17 17:46:13 +05:30
Josh Christie
c8307a0627
Use @info for logging
2018-08-11 14:42:33 +01:00
Avik Pal
5db7a3a3ad
Fix Optimizers
2018-08-11 18:23:47 +05:30
pevnak
3510c837a8
zeros replaced by zero
2018-08-03 15:14:25 +01:00
Jarvist Moore Frost
344a750770
Merge branch 'master' of github.com:jarvist/Flux.jl into HEAD
2018-07-03 11:15:43 +01:00
Jarvist Moore Frost
aee4a83c55
Add ADAMW weight-decay.
...
See http://www.fast.ai/2018/07/02/adam-weight-decay/ and the original
paper https://arxiv.org/abs/1711.05101.pdf for context.
I don't know what I'm doing, and this is quite possibly wrong - but on
a simple Char-RNN I have lying around on my harddisk, this seems to
improve the rate of learning consistently for different hyperparameters
vs. standard ADAM with the same decay constant.
2018-07-03 11:11:32 +01:00
Mike J Innes
2723c9ee04
Merge pull request #257 from staticfloat/sf/back_inf_nan
...
Check for `Inf` and `NaN` within `back!(::TrackedReal)`
2018-06-26 14:42:33 +01:00
Tejan Karmali
4a24b69976
Merge branch 'master' into nadam-opt
2018-06-08 16:54:41 +05:30
staticfloat@gmail.com
9fdbe843ef
Check for Inf
and NaN
within back!(::TrackedReal)
...
This is often checked for within user code, no reason to do that, let's
do it for them within `back!(::TrackedReal)`
2018-05-07 15:30:44 -07:00
CarloLucibello
e186b958dd
more exports
2018-05-01 12:13:14 +01:00
Sujeet Akula
8c042bd522
element wise max()
2018-04-26 21:12:31 +10:00
Sujeet Akula
5e5f255f81
export typo
2018-04-26 17:42:04 +10:00
Sujeet Akula
4586bda5ab
export/test adamax
2018-04-26 17:40:11 +10:00
Sujeet Akula
b6508e2416
add adamax
2018-04-26 17:37:24 +10:00
tejank10
65847bb745
moved epsilon into sqrt
2018-04-04 15:25:20 +05:30
tejank10
3ead662987
Update rule fixed
2018-04-04 15:18:44 +05:30
tejank10
ea9b5471fa
NADAM optimizer
2018-04-03 01:27:22 +05:30
Mike Innes
bfd6a4c0ec
cleaner interrupts
2018-03-05 23:05:45 +00:00
Mike Innes
5153cde847
move epochs
2018-03-05 22:56:22 +00:00
Mike J Innes
79e4e25fea
seperate number type
2018-02-07 20:39:36 +00:00
Mike J Innes
282889970d
seperate tracking infrastructure from array wrapper
2018-02-07 17:43:25 +00:00
Mike J Innes
5b97d2ba04
closes #127
2017-12-13 18:24:56 +00:00
Mike J Innes
24a6569589
Merge branch 'master' into amsgrad
2017-12-08 18:20:53 +00:00
baggepinnen
41febee9c1
Export and indent
2017-12-04 09:34:27 +01:00
baggepinnen
36001d085a
Implement AMSGrad optimiser
2017-12-04 09:17:05 +01:00
CarloLucibello
13b934c250
improve optimizers
2017-11-24 12:12:20 +01:00
Mike J Innes
9f5c4dd3e9
Merge pull request #104 from baggepinnen/patch-1
...
Allow array of optimisers to train!
2017-11-21 17:16:35 +01:00
Mike J Innes
979949d01a
style
2017-11-21 15:25:09 +01:00
Fredrik Bagge Carlson
8991ce028c
Fix bug in rmsprop and adadelta
...
`@. p.Δ = η * p.Δ / √acc` parses correctly while `@. p.Δ /= √acc*η` seems to parse like `@. p.Δ /= (√acc*η)`, hence the step size was de facto interpreted as `1/η`
2017-11-14 17:32:16 +01:00
Fredrik Bagge Carlson
97244e0a68
Allow array of optimisers to train!
...
This allows an array of optimisers to be sent to `train!`
2017-11-04 13:27:32 +01:00
Mike J Innes
99a7697d13
adam eta default arg
2017-10-19 14:31:34 +01:00
Mike J Innes
5b6a5667ed
tracked array restructure
2017-10-18 22:54:58 +01:00
Mike J Innes
07ad7cfa40
learning rate as default arg
2017-10-18 17:07:49 +01:00
Mike J Innes
7426faf37d
optimiser docs
2017-10-18 12:09:48 +01:00