# GPU Support Support for array operations on other hardware backends, like GPUs, is provided by external packages like [CuArrays](https://github.com/JuliaGPU/CuArrays.jl). Flux is agnostic to array types, so we simply need to move model weights and data to the GPU and Flux will handle it. For example, we can use `CuArrays` (with the `cu` converter) to run our [basic example](models/basics.md) on an NVIDIA GPU. ```julia using CuArrays W = cu(rand(2, 5)) # a 2×5 CuArray b = cu(rand(2)) predict(x) = W*x .+ b loss(x, y) = sum((predict(x) .- y).^2) x, y = cu(rand(5)), cu(rand(2)) # Dummy data loss(x, y) # ~ 3 ``` Note that we convert both the parameters (`W`, `b`) and the data set (`x`, `y`) to cuda arrays. Taking derivatives and training works exactly as before. If you define a structured model, like a `Dense` layer or `Chain`, you just need to convert the internal parameters. Flux provides `mapleaves`, which allows you to alter all parameters of a model at once. ```julia d = Dense(10, 5, σ) d = mapleaves(cu, d) d.W # Tracked CuArray d(cu(rand(10))) # CuArray output m = Chain(Dense(10, 5, σ), Dense(5, 2), softmax) m = mapleaves(cu, m) d(cu(rand(10))) ``` As a convenience, Flux provides the `gpu` function to convert models and data to the GPU if one is available. By default, it'll do nothing, but loading `CuArrays` will cause it to move data to the GPU instead. ```julia julia> using Flux, CuArrays julia> m = Dense(10,5) |> gpu Dense(10, 5) julia> x = rand(10) |> gpu 10-element CuArray{Float32,1}: 0.800225 ⋮ 0.511655 julia> m(x) Tracked 5-element CuArray{Float32,1}: -0.30535 ⋮ -0.618002 ```