Source code for chainer.links.activation.simplified_dropconnect

import numpy

from chainer import cuda
from chainer.functions.noise import simplified_dropconnect
from chainer import initializers
from chainer import link


[docs]class SimplifiedDropconnect(link.Link): """Fully-connected layer with simplified dropconnect regularization. Notice: This implementation cannot be used for reproduction of the paper. There is a difference between the current implementation and the original one. The original version uses sampling with gaussian distribution before passing activation function, whereas the current implementation averages before activation. Args: in_size (int): Dimension of input vectors. If ``None``, parameter initialization will be deferred until the first forward data pass at which time the size will be determined. out_size (int): Dimension of output vectors. nobias (bool): If ``True``, then this link does not use the bias term. initialW (3-D array or None): Initial weight value. If ``None``, the default initializer is used. May also be a callable that takes ``numpy.ndarray`` or ``cupy.ndarray`` and edits its value. initial_bias (2-D array, float or None): Initial bias value. If ``None``, the bias is set to 0. May also be a callable that takes ``numpy.ndarray`` or ``cupy.ndarray`` and edits its value. Attributes: W (~chainer.Variable): Weight parameter. b (~chainer.Variable): Bias parameter. .. seealso:: :func:`~chainer.functions.simplified_dropconnect` .. seealso:: Li, W., Matthew Z., Sixin Z., Yann L., Rob F. (2013). Regularization of Neural Network using DropConnect. International Conference on Machine Learning. `URL <http://cs.nyu.edu/~wanli/dropc/>`_ """ def __init__(self, in_size, out_size, ratio=.5, nobias=False, initialW=None, initial_bias=None): super(SimplifiedDropconnect, self).__init__() self.out_size = out_size self.ratio = ratio if initialW is None: initialW = initializers.HeNormal(1. / numpy.sqrt(2)) self._W_initializer = initializers._get_initializer(initialW) if in_size is None: self.add_uninitialized_param('W') else: self._initialize_params(in_size) if nobias: self.b = None else: if initial_bias is None: initial_bias = initializers.Constant(0) bias_initializer = initializers._get_initializer(initial_bias) self.add_param('b', out_size, initializer=bias_initializer) def _initialize_params(self, in_size): self.add_param('W', (self.out_size, in_size), initializer=self._W_initializer)
[docs] def __call__(self, x, train=True, mask=None): """Applies the simplified dropconnect layer. Args: x (chainer.Variable or :class:`numpy.ndarray` or cupy.ndarray): Batch of input vectors. Its first dimension ``n`` is assumed to be the *minibatch dimension*. train (bool): If ``True``, executes simplified dropconnect. Otherwise, simplified dropconnect link works as a linear unit. mask (None or chainer.Variable or :class:`numpy.ndarray` or cupy.ndarray): If ``None``, randomized simplified dropconnect mask is generated. Otherwise, The mask must be ``(n, M, N)`` shaped array. Main purpose of this option is debugging. `mask` array will be used as a dropconnect mask. Returns: ~chainer.Variable: Output of the simplified dropconnect layer. """ if self.has_uninitialized_params: with cuda.get_device_from_id(self._device_id): self._initialize_params(x.size // len(x.data)) if mask is not None and 'mask' not in self.__dict__: self.add_persistent('mask', mask) return simplified_dropconnect.simplified_dropconnect(x, self.W, self.b, self.ratio, train, mask)