Custom Layers
One factor behind deep learning’s success is the availability of a wide range of layers that can be composed in creative ways to design architectures suitable for a wide variety of tasks. For instance, researchers have invented layers specifically for handling images, text, looping over sequential data, and performing dynamic programming. Sooner or later, you will encounter or invent a layer that does not exist yet in the deep learning framework. In these cases, you must build a custom layer. In this section, we show you how.
Layers without Parameters
To start, we construct a custom layer
that does not have any parameters of its own.
This should look familiar if you recall our
introduction to block in :numref:sec_model_construction
.
The following CenteredLayer
class simply
subtracts the mean from its input.
To build it, we simply need to inherit
from the base layer class and implement the forward propagation function.
```{.python .input} from mxnet import gluon, np, npx from mxnet.gluon import nn npx.set_np()
class CenteredLayer(nn.Block): def init(self, kwargs): super().init(kwargs)
def forward(self, X):
return X - X.mean()
```{.python .input}
#@tab pytorch
import torch
from torch import nn
class CenteredLayer(nn.Module):
def __init__(self):
super().__init__()
def forward(self, X):
return X - X.mean()
```{.python .input}
@tab tensorflow
import tensorflow as tf
class CenteredLayer(tf.keras.Model): def init(self): super().init()
def call(self, inputs):
return inputs - tf.reduce_mean(inputs)
Let us verify that our layer works as intended by feeding some data through it.
```{.python .input}
layer = CenteredLayer()
layer(np.array([1, 2, 3, 4, 5]))
```{.python .input}
@tab pytorch
layer = CenteredLayer() layer(torch.FloatTensor([1, 2, 3, 4, 5]))
```{.python .input}
#@tab tensorflow
layer = CenteredLayer()
layer(tf.constant([1, 2, 3, 4, 5]))
We can now incorporate our layer as a component in constructing more complex models.
```{.python .input} net = nn.Sequential() net.add(nn.Dense(128), CenteredLayer()) net.initialize()
```{.python .input}
#@tab pytorch
net = nn.Sequential(nn.Linear(8, 128), CenteredLayer())
```{.python .input}
@tab tensorflow
net = tf.keras.Sequential([tf.keras.layers.Dense(128), CenteredLayer()])
As an extra sanity check, we can send random data
through the network and check that the mean is in fact 0.
Because we are dealing with floating point numbers,
we may still see a very small nonzero number
due to quantization.
```{.python .input}
Y = net(np.random.uniform(size=(4, 8)))
Y.mean()
```{.python .input}
@tab pytorch
Y = net(torch.rand(4, 8)) Y.mean()
```{.python .input}
#@tab tensorflow
Y = net(tf.random.uniform((4, 8)))
tf.reduce_mean(Y)
Layers with Parameters
Now that we know how to define simple layers, let us move on to defining layers with parameters that can be adjusted through training. We can use built-in functions to create parameters, which provide some basic housekeeping functionality. In particular, they govern access, initialization, sharing, saving, and loading model parameters. This way, among other benefits, we will not need to write custom serialization routines for every custom layer.
Now let us implement our own version of the fully-connected layer.
Recall that this layer requires two parameters,
one to represent the weight and the other for the bias.
In this implementation, we bake in the ReLU activation as a default.
This layer requires to input arguments: in_units
and units
, which
denote the number of inputs and outputs, respectively.
```{.python .input} class MyDense(nn.Block): def init(self, units, inunits, **kwargs): super()._init(**kwargs) self.weight = self.params.get(‘weight’, shape=(in_units, units)) self.bias = self.params.get(‘bias’, shape=(units,))
def forward(self, x):
linear = np.dot(x, self.weight.data(ctx=x.ctx)) + self.bias.data(
ctx=x.ctx)
return npx.relu(linear)
```{.python .input}
#@tab pytorch
class MyLinear(nn.Module):
def __init__(self, in_units, units):
super().__init__()
self.weight = nn.Parameter(torch.randn(in_units, units))
self.bias = nn.Parameter(torch.randn(units,))
def forward(self, X):
return torch.matmul(X, self.weight.data) + self.bias.data
```{.python .input}
@tab tensorflow
class MyDense(tf.keras.Model): def init(self, units): super().init() self.units = units
def build(self, X_shape):
self.weight = self.add_weight(name='weight',
shape=[X_shape[-1], self.units],
initializer=tf.random_normal_initializer())
self.bias = self.add_weight(
name='bias', shape=[self.units],
initializer=tf.zeros_initializer())
def call(self, X):
return tf.matmul(X, self.weight) + self.bias
Next, we instantiate the `MyDense` class
and access its model parameters.
```{.python .input}
dense = MyDense(units=3, in_units=5)
dense.params
```{.python .input}
@tab pytorch
dense = MyLinear(5, 3) dense.weight
```{.python .input}
#@tab tensorflow
dense = MyDense(3)
dense(tf.random.uniform((2, 5)))
dense.get_weights()
We can directly carry out forward propagation calculations using custom layers.
```{.python .input} dense.initialize() dense(np.random.uniform(size=(2, 5)))
```{.python .input}
#@tab pytorch
dense(torch.randn(2, 5))
```{.python .input}
@tab tensorflow
dense(tf.random.uniform((2, 5)))
We can also construct models using custom layers.
Once we have that we can use it just like the built-in fully-connected layer.
```{.python .input}
net = nn.Sequential()
net.add(MyDense(8, in_units=64),
MyDense(1, in_units=8))
net.initialize()
net(np.random.uniform(size=(2, 64)))
```{.python .input}
@tab pytorch
net = nn.Sequential(MyLinear(64, 8), nn.ReLU(), MyLinear(8, 1)) net(torch.randn(2, 64))
```{.python .input}
#@tab tensorflow
net = tf.keras.models.Sequential([MyDense(8), MyDense(1)])
net(tf.random.uniform((2, 64)))
Summary
- We can design custom layers via the basic layer class. This allows us to define flexible new layers that behave differently from any existing layers in the library.
- Once defined, custom layers can be invoked in arbitrary contexts and architectures.
- Layers can have local parameters, which can be created through built-in functions.
Exercises
- Design a layer that takes an input and computes a tensor reduction, i.e., it returns $yk = \sum{i, j} W_{ijk} x_i x_j$.
- Design a layer that returns the leading half of the Fourier coefficients of the data.
:begin_tab:mxnet
Discussions
:end_tab:
:begin_tab:pytorch
Discussions
:end_tab:
:begin_tab:tensorflow
Discussions
:end_tab: