Commit Graph

326 Commits

Author SHA1 Message Date
bors[bot] 71ebd51e45
Merge #1208
1208: Fixing output format for `onehot` r=dhairyagandhi96 a=natema

Currently `Flux.OneHotVector` is displayed as a binary vector (0/1) rather than a boolean one (true/false). This is also shown in successive examples in the same page. 
I fixed the `onehot(:b, [:a, :b, :c])` and `onehot(:c, [:a, :b, :c])` outputs in the first example of the page accordingly.


Co-authored-by: natema <natema@users.noreply.github.com>
2020-06-05 09:17:12 +00:00
bors[bot] b5a73f8532
Merge #1207
1207: Fixing typo in docs r=dhairyagandhi96 a=natema

`what ever` -> `whatever`


Co-authored-by: natema <natema@users.noreply.github.com>
2020-06-05 09:00:06 +00:00
natema 48d6f2d0c0
Fixing output format for `onehot`
`Flux.OneHotVector` is displayed as a binary vector (0/1) rather than a boolean (true/false) one, as is also shown in successive examples in the same page, so I fixed the `onehot(:b, [:a, :b, :c])` and `onehot(:c, [:a, :b, :c])` output as given by the current Julia version 1.4.2.
2020-06-03 17:03:08 +02:00
natema 2c4b1e521e
Fixing typo in docs
`what ever` -> `whatever`
2020-06-02 19:20:41 +02:00
natema a24f46b606
Fixing ambiguous remark in Preserve inputs' types
This PR is based on the [discussion in the forum](https://discourse.julialang.org/t/not-clear-what-0-01f0x-is-in-the-flux-docs/40553?u=mathematics) on the ambiguity of `0.01f0x` in the line
> While one could change the activation function (e.g. to use `0.01f0x`)
2020-06-02 18:48:07 +02:00
Troels Arnfred Bojesen 17bb00a3fa
Correcting advanced.md
To make the example consistent, it should be 
```
julia> Flux.trainable(a::Affine) = (a.W,)
```
not
```
julia> Flux.trainable(a::Affine) = (a.W, a.b)
```
2020-05-25 23:33:09 +09:00
Dhairya Gandhi bdfe567519 add some layers to docs 2020-05-18 23:53:11 +05:30
bors[bot] b6a5dd7152
Merge #1133
1133: add ClipValue and ClipNorm r=CarloLucibello a=AStupidBear



Co-authored-by: Yao Lu <luyaocns@gmail.com>
2020-05-15 17:15:07 +00:00
Yao Lu d1ad8db625 add to docs 2020-05-09 16:40:26 +08:00
Mikkel Paltorp Schmitt 40efa9df49
Update basics.md
Removing superfluous ```using Flux```
2020-05-06 13:41:56 +02:00
Dhairya Gandhi d6a1ccd354 add correct overload for apply in docs 2020-05-03 16:56:39 +05:30
janEbert 684570660a Update doctest version guard (1.2 -> 1.4)
And add the same to docs/make.jl
2020-04-06 13:53:36 +02:00
janEbert 0e9bc82626 Loss -> Loss Functions 2020-04-06 13:52:27 +02:00
janEbert 73d631f5cd Fix and improve docs
Add missing docstrings, improve existing ones, fix links to functions
or files.
2020-04-04 23:00:34 +02:00
janEbert 64ce32ddcf Fix problems due to rebase 2020-04-04 22:55:14 +02:00
janEbert ff9198b939 Add datasets to docs
All the relevant functions. Perhaps discuss a consistent API, describe
it in the docs and then only document the modules.
2020-04-04 18:19:20 +02:00
janEbert ba80c2e8ab Improve whitespaces in docs 2020-04-04 18:16:46 +02:00
janEbert c76b7315ac Add loss and utility functions to docs 2020-04-04 17:39:19 +02:00
janEbert 9b68423e64 Import (`using`) Flux for all doctests 2020-04-04 17:22:08 +02:00
janEbert 1bf8dc2d5b Update Documenter version and fix warnings
0.23.2 -> 0.23.3
2020-04-04 17:22:08 +02:00
Garben Tanghe 746e3310f1 removed Flatten struct
updated documentation
2020-03-08 14:22:03 +01:00
Garben Tanghe 3e14bd878c added GlobalMaxPool, GlobalMeanPool, and Flatten layers 2020-03-08 14:18:48 +01:00
CarloLucibello 12106ff4cc update freeze docs 2020-03-04 04:45:41 +01:00
bors[bot] 94ba1e8ede
Merge #1028 #1070
1028: Common questions answered in docs r=CarloLucibello a=dhairyagandhi96

cc @MikeInnes 

1070: Prevent breakage due to new `active` field in normalise layers r=CarloLucibello a=ianshmean

Prevents breakage where the normalise structs, such as `BatchNorm`, have been directly defined but missing the new `active` field

cc. @darsnack 

Co-authored-by: Dhairya Gandhi <dhairya@juliacopmuting.com>
Co-authored-by: Dhairya Gandhi <dhairya@juliacomputing.com>
Co-authored-by: Ian <i.r.butterworth@gmail.com>
2020-03-04 00:10:39 +00:00
bors[bot] af23a5756c
Merge #1053
1053: Added Some Loss functions with some doc improvements r=CarloLucibello a=AdarshKumar712

Added the following loss functions with tests:
1. mae
2. mean squared logarithmic error
3. huber loss
4. squared hinge loss
5. dice coeff loss
6. tversky loss 

Also added some documentation improvements for few other functions. 

Co-authored-by: Adarsh Kumar <45385384+AdarshKumar712@users.noreply.github.com>
2020-03-03 23:56:21 +00:00
Adarshkumar712 d0e8a9ff71 Updated activation functions in NNlib doc 2020-03-03 22:07:05 +05:30
CarloLucibello af99ca27ee docs update 2020-03-03 07:52:20 +01:00
CarloLucibello f5da4d0c70 remove docs manifest 2020-03-02 15:10:08 +01:00
CarloLucibello ffea8b616d fix docs 2020-03-02 15:08:37 +01:00
CarloLucibello e51070bf79 update documenter 2020-03-02 15:08:37 +01:00
Adarsh Kumar 89d07c07ec
Added Loss functions to docs 2020-03-02 13:33:44 +05:30
Dhairya Gandhi cbb9a2a929
Merge branch 'master' into dg/params_docs 2020-03-02 12:45:30 +05:30
Dhairya Gandhi bb5350591f cleanup 2020-03-02 12:42:33 +05:30
Dhairya Gandhi 27949693f3 refactor 2020-03-02 12:40:19 +05:30
bors[bot] be38146ee9
Merge #1061
1061: fix a few typos in docstrings r=CarloLucibello a=visr



Co-authored-by: Martijn Visser <mgvisser@gmail.com>
2020-03-02 01:03:58 +00:00
bors[bot] 6575fb8f48
Merge #1057
1057: add Julia ecosystem doc section r=CarloLucibello a=CarloLucibello

Partially fixing #251,  related to the discussion in #1051 .

Not exactly a poem that I wrote here, maybe someone could suggest a better rephrasing. 
Suggestion for additional packages to add to the list also welcome

Co-authored-by: CarloLucibello <carlo.lucibello@gmail.com>
2020-03-02 00:52:22 +00:00
Kyle Daruwalla 4cebf36361
Merge branch 'master' into feature/istraining 2020-03-01 12:32:15 -06:00
Kyle Daruwalla c001d0f3c5 Added trainmode! and updated docs with warning 2020-03-01 12:30:41 -06:00
Martijn Visser 6076847a45 fix a few typos in docstrings 2020-03-01 15:07:12 +01:00
CarloLucibello b6c79b38b4 add DataLoader
special case train! for the unsupervised data iterator
2020-02-29 18:50:59 +01:00
CarloLucibello 4f693e02cb add model zoo reference 2020-02-29 13:50:23 +01:00
CarloLucibello 4109f2e0d7 cleanup 2020-02-29 13:45:17 +01:00
CarloLucibello 169ed6eb25 add ecosystem 2020-02-29 13:43:03 +01:00
Carlo Lucibello 425fcdbe69 NNlib docs + misc docs improvements 2020-02-29 11:14:48 +01:00
CarloLucibello 759fe9df2f update docs and export update! 2020-02-26 20:27:39 +01:00
Kyle Daruwalla ba5259a269 Added docs on testmode! 2020-02-25 13:53:49 -06:00
bors[bot] 55616afc11
Merge #960
960: Added utility function outdims to compute output dimensions of a layer r=dhairyagandhi96 a=darsnack

Based on Slack chatter, I added a utility function, `outdims`, that computes the output dimensions for given input dimensions.

Example
```julia
layer = Conv((3, 3), 3 => 16)
outdims(layer, (10, 10)) # returns (8, 8)
```

Co-authored-by: Kyle Daruwalla <daruwalla@wisc.edu>
2020-02-25 17:40:05 +00:00
Kyle Daruwalla f5b9cf659c Updated docs to specify exactly what layers support outdims 2020-02-20 23:38:56 -06:00
Dhairya Gandhi d5ed9a4478
Update docs/src/models/basics.md
Co-Authored-By: Carlo Lucibello <carlo.lucibello@gmail.com>
2020-02-12 11:26:11 +05:30
Dhairya Gandhi ee6d950696
Update docs/src/models/basics.md
Co-Authored-By: Carlo Lucibello <carlo.lucibello@gmail.com>
2020-02-12 11:25:50 +05:30