From 4470a9cace7a94dc804a82a29807923c1afe5ff3 Mon Sep 17 00:00:00 2001 From: autodocs Date: Sat, 4 Mar 2017 14:00:54 +0000 Subject: [PATCH] build based on 304644e --- latest/apis/backends.html | 2 +- latest/apis/batching.html | 2 +- latest/apis/storage.html | 2 +- latest/contributing.html | 2 +- latest/examples/char-rnn.html | 2 +- latest/examples/logreg.html | 2 +- latest/index.html | 5 ++--- latest/internals.html | 2 +- latest/models/basics.html | 6 +++--- latest/models/debugging.html | 2 +- latest/models/recurrent.html | 2 +- latest/models/templates.html | 2 +- latest/search_index.js | 4 ++-- 13 files changed, 17 insertions(+), 18 deletions(-) diff --git a/latest/apis/backends.html b/latest/apis/backends.html index 25f340f4..f7256c1f 100644 --- a/latest/apis/backends.html +++ b/latest/apis/backends.html @@ -150,7 +150,7 @@ Backends - + diff --git a/latest/apis/batching.html b/latest/apis/batching.html index e0a2ac7a..7227fdb5 100644 --- a/latest/apis/batching.html +++ b/latest/apis/batching.html @@ -155,7 +155,7 @@ Batching - + diff --git a/latest/apis/storage.html b/latest/apis/storage.html index ee31e629..cda9ed30 100644 --- a/latest/apis/storage.html +++ b/latest/apis/storage.html @@ -139,7 +139,7 @@ Storing Models - + diff --git a/latest/contributing.html b/latest/contributing.html index af1520f3..ee244b3e 100644 --- a/latest/contributing.html +++ b/latest/contributing.html @@ -136,7 +136,7 @@ Contributing & Help - + diff --git a/latest/examples/char-rnn.html b/latest/examples/char-rnn.html index 2b1f0091..2017d9f5 100644 --- a/latest/examples/char-rnn.html +++ b/latest/examples/char-rnn.html @@ -139,7 +139,7 @@ Char RNN - + diff --git a/latest/examples/logreg.html b/latest/examples/logreg.html index 3207cf28..a4622a4c 100644 --- a/latest/examples/logreg.html +++ b/latest/examples/logreg.html @@ -139,7 +139,7 @@ Logistic Regression - + diff --git a/latest/index.html b/latest/index.html index fb0a2343..15c8fb2c 100644 --- a/latest/index.html +++ b/latest/index.html @@ -147,7 +147,7 @@ Home - + @@ -218,8 +218,7 @@ Installation ... Charging Ion Capacitors ...

-
Pkg.clone("https://github.com/MikeInnes/DataFlow.jl")
-Pkg.clone("https://github.com/MikeInnes/Flux.jl")
+
Pkg.add("Flux.jl")
 using Flux

You'll also need a backend to run real training, if you don't have one already. Choose from diff --git a/latest/internals.html b/latest/internals.html index 2c9a05a6..af8af078 100644 --- a/latest/internals.html +++ b/latest/internals.html @@ -136,7 +136,7 @@ Internals - + diff --git a/latest/models/basics.html b/latest/models/basics.html index a387e967..a5e2efa5 100644 --- a/latest/models/basics.html +++ b/latest/models/basics.html @@ -155,7 +155,7 @@ Model Building Basics - + @@ -236,8 +236,8 @@ We just created two separate , leading to a different result when called with our data. It's easy to define templates like Affine ourselves (see - -The Template + +templates ), but Flux provides Affine diff --git a/latest/models/debugging.html b/latest/models/debugging.html index ebd20b2e..bf3d8379 100644 --- a/latest/models/debugging.html +++ b/latest/models/debugging.html @@ -139,7 +139,7 @@ Debugging - + diff --git a/latest/models/recurrent.html b/latest/models/recurrent.html index 777d0702..ac0f1dd9 100644 --- a/latest/models/recurrent.html +++ b/latest/models/recurrent.html @@ -139,7 +139,7 @@ Recurrence - + diff --git a/latest/models/templates.html b/latest/models/templates.html index b6703b8a..355dfedf 100644 --- a/latest/models/templates.html +++ b/latest/models/templates.html @@ -155,7 +155,7 @@ Model Templates - + diff --git a/latest/search_index.js b/latest/search_index.js index 0410ef16..d5bfaf8b 100644 --- a/latest/search_index.js +++ b/latest/search_index.js @@ -29,7 +29,7 @@ var documenterSearchIndex = {"docs": [ "page": "Home", "title": "Installation", "category": "section", - "text": "... Charging Ion Capacitors ...Pkg.clone(\"https://github.com/MikeInnes/DataFlow.jl\")\nPkg.clone(\"https://github.com/MikeInnes/Flux.jl\")\nusing FluxYou'll also need a backend to run real training, if you don't have one already. Choose from MXNet or TensorFlow (MXNet is the recommended option if you're not sure):Pkg.add(\"MXNet\") # or \"TensorFlow\"\nPkg.test(\"Flux\") # Make sure everything installed properly" + "text": "... Charging Ion Capacitors ...Pkg.add(\"Flux.jl\")\nusing FluxYou'll also need a backend to run real training, if you don't have one already. Choose from MXNet or TensorFlow (MXNet is the recommended option if you're not sure):Pkg.add(\"MXNet\") # or \"TensorFlow\"\nPkg.test(\"Flux\") # Make sure everything installed properly" }, { @@ -53,7 +53,7 @@ var documenterSearchIndex = {"docs": [ "page": "Model Building Basics", "title": "The Model", "category": "section", - "text": "... Initialising Photon Beams ...The core concept in Flux is the model. A model (or \"layer\") is simply a function with parameters. For example, in plain Julia code, we could define the following function to represent a logistic regression (or simple neural network):W = randn(3,5)\nb = randn(3)\naffine(x) = W * x + b\n\nx1 = rand(5) # [0.581466,0.606507,0.981732,0.488618,0.415414]\ny1 = softmax(affine(x1)) # [0.32676,0.0974173,0.575823]affine is simply a function which takes some vector x1 and outputs a new one y1. For example, x1 could be data from an image and y1 could be predictions about the content of that image. However, affine isn't static. It has parameters W and b, and if we tweak those parameters we'll tweak the result – hopefully to make the predictions more accurate.This is all well and good, but we usually want to have more than one affine layer in our network; writing out the above definition to create new sets of parameters every time would quickly become tedious. For that reason, we want to use a template which creates these functions for us:affine1 = Affine(5, 5)\naffine2 = Affine(5, 5)\n\nsoftmax(affine1(x1)) # [0.167952, 0.186325, 0.176683, 0.238571, 0.23047]\nsoftmax(affine2(x1)) # [0.125361, 0.246448, 0.21966, 0.124596, 0.283935]We just created two separate Affine layers, and each contains its own version of W and b, leading to a different result when called with our data. It's easy to define templates like Affine ourselves (see The Template), but Flux provides Affine out of the box, so we'll use that for now." + "text": "... Initialising Photon Beams ...The core concept in Flux is the model. A model (or \"layer\") is simply a function with parameters. For example, in plain Julia code, we could define the following function to represent a logistic regression (or simple neural network):W = randn(3,5)\nb = randn(3)\naffine(x) = W * x + b\n\nx1 = rand(5) # [0.581466,0.606507,0.981732,0.488618,0.415414]\ny1 = softmax(affine(x1)) # [0.32676,0.0974173,0.575823]affine is simply a function which takes some vector x1 and outputs a new one y1. For example, x1 could be data from an image and y1 could be predictions about the content of that image. However, affine isn't static. It has parameters W and b, and if we tweak those parameters we'll tweak the result – hopefully to make the predictions more accurate.This is all well and good, but we usually want to have more than one affine layer in our network; writing out the above definition to create new sets of parameters every time would quickly become tedious. For that reason, we want to use a template which creates these functions for us:affine1 = Affine(5, 5)\naffine2 = Affine(5, 5)\n\nsoftmax(affine1(x1)) # [0.167952, 0.186325, 0.176683, 0.238571, 0.23047]\nsoftmax(affine2(x1)) # [0.125361, 0.246448, 0.21966, 0.124596, 0.283935]We just created two separate Affine layers, and each contains its own version of W and b, leading to a different result when called with our data. It's easy to define templates like Affine ourselves (see templates), but Flux provides Affine out of the box, so we'll use that for now." }, {