Commit Graph

2689 Commits

Author SHA1 Message Date
Dhairya Gandhi 55430e207d add news 2020-05-19 16:34:28 +05:30
bors[bot] 0b10f1a8df
Merge #1184
1184: Add some functions to docs r=dhairyagandhi96 a=dhairyagandhi96



Co-authored-by: Dhairya Gandhi <dhairya@juliacopmuting.com>
2020-05-18 21:10:46 +00:00
DrChainsaw 9a24ee0bd7 Change intendation to 2 spaces 2020-05-18 21:52:40 +02:00
Dhairya Gandhi bdfe567519 add some layers to docs 2020-05-18 23:53:11 +05:30
bors[bot] b6a5dd7152
Merge #1133
1133: add ClipValue and ClipNorm r=CarloLucibello a=AStupidBear



Co-authored-by: Yao Lu <luyaocns@gmail.com>
2020-05-15 17:15:07 +00:00
Yao Lu 007586858c fix export merge conflict 2020-05-14 17:13:35 +08:00
Dhairya Gandhi fab53e0a01
Merge pull request #1179 from FluxML/compathelper/new_version/2020-05-13-00-13-17-919-1190174363
CompatHelper: add new compat entry for "Functors" at version "0.1"
2020-05-13 11:27:40 +05:30
github-actions[bot] 3fa9e91c41 CompatHelper: add new compat entry for "Functors" at version "0.1" 2020-05-13 00:13:46 +00:00
DrChainsaw e8433d0abe Align ExpDecay implementation with documentation 2020-05-12 22:50:17 +02:00
bors[bot] de39d1095b
Merge #1175
1175: xlogy broadcast adjoint r=MikeInnes a=MikeInnes

This is helpful for performance, since it avoids having to differentiate `xlogy` itself inside of a map.

Co-authored-by: Mike J Innes <mike.j.innes@gmail.com>
2020-05-12 17:10:58 +00:00
Mike J Innes f5a8900ffb xlogy broadcast adjoint 2020-05-12 17:29:35 +01:00
Mike J Innes bd43201f37
fix logitcrossentropy doc string 2020-05-12 16:18:29 +01:00
bors[bot] a84e08cf28
Merge #1174
1174: Functors r=MikeInnes a=MikeInnes

Just splits out the implementation to the [Functors](https://github.com/FluxML/Functors.jl) package, so the same traits can be used elsewhere (e.g. Optimisers.jl) without depending on all of Flux.

Co-authored-by: Mike J Innes <mike.j.innes@gmail.com>
2020-05-12 14:39:08 +00:00
Mike J Innes 22d29c9bfd released functors.jl 2020-05-12 15:33:14 +01:00
Dhairya Gandhi 36d3a9ce99
Merge pull request #1162 from aminya/patch-5
Update CompatHelper.yml
2020-05-10 14:21:14 +05:30
Yao Lu 5a9eb7411a cpu 2020-05-10 14:39:48 +08:00
Yao Lu 888f286c51 use @inbounds 2020-05-09 19:40:46 +08:00
Yao Lu 63cb70dd23 remove importing CuMatrix 2020-05-09 19:13:52 +08:00
Yao Lu 30648910c8 transfer onehot indices back to cpu 2020-05-09 19:10:46 +08:00
Yao Lu d1ad8db625 add to docs 2020-05-09 16:40:26 +08:00
bors[bot] d89ee6cdba
Merge #1167
1167: Update basics.md r=dhairyagandhi96 a=mipals

Removing superfluous ```using Flux```

Co-authored-by: Mikkel Paltorp Schmitt <mikkel.paltorp@gmail.com>
2020-05-08 11:38:22 +00:00
bors[bot] 0287abbf66
Merge #1166
1166: Fix crossentropy when some probabilities are zero r=dhairyagandhi96 a=cossio

Use a function `xlogy(x,y) = x * log(y)` that has the correct limit at `x=0`.

Before this PR:

```julia
julia> Flux.crossentropy([0.1,0.0,0.9], [0.1,0.0,0.9])
NaN
```

After this PR:

```julia
julia> Flux.crossentropy([0.1,0.0,0.9], [0.1,0.0,0.9])
0.3250829733914482
```

Co-authored-by: cossio <j.cossio.diaz@gmail.com>
2020-05-08 11:14:31 +00:00
cossio 17f54e4c6f bump version 2020-05-08 12:57:34 +02:00
cossio feb72d400a NaN 2020-05-07 12:44:32 +02:00
cossio 86d6555269 cufunc 2020-05-07 09:58:33 +02:00
Mikkel Paltorp Schmitt 40efa9df49
Update basics.md
Removing superfluous ```using Flux```
2020-05-06 13:41:56 +02:00
cossio 8314200c51 generic 2020-05-05 19:23:05 +02:00
cossio 06c1e20372 add tests 2020-05-05 19:05:04 +02:00
cossio 480473a81b xlogy 2020-05-05 18:33:50 +02:00
cossio 9e1fd883d5
Fix docstring of logitbinarycrossentropy and logitcrossentropy 2020-05-05 16:29:29 +02:00
Amin Yahyaabadi 70f76fd6db
Update CompatHelper.yml 2020-05-05 07:11:22 -05:00
bors[bot] c444226db5
Merge #1160
1160: Build docs on Julia 1.3 r=dhairyagandhi96 a=dhairyagandhi96

This causes red CI otherwise

Co-authored-by: Dhairya Gandhi <dhairya@juliacopmuting.com>
2020-05-04 12:59:25 +00:00
Dhairya Gandhi f2c66579ec yaml syntax fix 2020-05-04 18:01:33 +05:30
Dhairya Gandhi fc464f5ef8 build docs on Julia 1.3 2020-05-04 17:54:04 +05:30
bors[bot] 1e2476b3c2
Merge #1156
1156: Add correct overload for apply! in docs r=dhairyagandhi96 a=dhairyagandhi96

Maybe we should considering adding a `const` name that is better than `apply!` (or rename `apply!`) and export it, so folks can just overload `descriptive_apply_my_optimiser_rule!` rather than have to go to the sub-project `Flux.Optimise`?

Co-authored-by: Dhairya Gandhi <dhairya@juliacopmuting.com>
2020-05-04 06:01:23 +00:00
Dhairya Gandhi d6a1ccd354 add correct overload for apply in docs 2020-05-03 16:56:39 +05:30
bors[bot] 5d9acc7e73
Merge #873
873: Make bias optional r=MikeInnes a=dhairyagandhi96

Addresses #868 



Co-authored-by: Dhairya Gandhi <dhairya@juliacopmuting.com>
2020-05-01 13:28:15 +00:00
Mike J Innes 8f877f2dbf quick fix 2020-05-01 14:22:46 +01:00
Dhairya Gandhi 29215fa5d7 comment on possible future deprecations 2020-04-29 16:17:44 +05:30
Dhairya Gandhi 534809ae78 move zeros to its own file 2020-04-29 16:15:35 +05:30
Dhairya Gandhi 5086c0f4f0 merge conflicts 2020-04-29 16:11:39 +05:30
Yao Lu 114f63a214 norm(Δ) 2020-04-26 17:28:07 +08:00
Yao Lu eb6898ea19 speedup matmul of CuMatrix and OneHotMatrix 2020-04-25 23:22:46 +08:00
Yao Lu 7d6f711c6f Merge branch 'master' into clip 2020-04-25 22:18:58 +08:00
bors[bot] 9237cdaf5b
Merge #901
901: Add option for "Same" padding to conv and pooling layers r=dhairyagandhi96 a=DrChainsaw

Fixes #813 

This adds the possibility to set "pad=SamePad()" to automatically calculate the amount of padding to apply so that outputsize==inputsize (assuming stide == 1).

Comments on API more than welcome. I considered the following options:

* Call the type just Same and export it, but I was afraid to cause name collisions due to a too generic name
* Call the type Same and not export it
* Dispatch on type instead of instance (so that one can type pad=Same instead of pad=Same())
* Supply a method instead of a type, giving a similar API as above. 

Happy to change to any of the above or to anything else.

I don't think that same padding is common for pooling layers, but I added it just for the sake of consistency. It is a separate commit so it can easily be removed if not wanted.

Co-authored-by: DrChainsaw <Christian.kyril.skarby@gmail.com>
2020-04-25 04:39:18 +00:00
DrChainsaw 4e4f6d9d1f Change next version entry to 0.10.5 2020-04-24 22:07:57 +02:00
DrChainsaw deff98812a Add v0.11.0 entry and added samepadding option 2020-04-24 21:59:02 +02:00
DrChainsaw 1544f84bb9 Fix merge conflicts 2020-04-24 21:56:26 +02:00
Yao Lu 58a72ec879 Merge branch 'master' of https://github.com/FluxML/Flux.jl into clip 2020-04-22 01:29:13 +08:00
Yao Lu c4f5e83697 resolve conflict 2020-04-22 01:24:13 +08:00