enow.com Web Search

Search results

  1. Results from the WOW.Com Content Network
  2. Convolutional layer - Wikipedia

    en.wikipedia.org/wiki/Convolutional_layer

    In artificial neural networks, a convolutional layer is a type of network layer that applies a convolution operation to the input. Convolutional layers are some of the primary building blocks of convolutional neural networks (CNNs), a class of neural network most commonly applied to images, video, audio, and other data that have the property of uniform translational symmetry.

  3. Tensor (machine learning) - Wikipedia

    en.wikipedia.org/wiki/Tensor_(machine_learning)

    In machine learning, the term tensor informally refers to two different concepts (i) a way of organizing data and (ii) a multilinear (tensor) transformation. Data may be organized in a multidimensional array (M-way array), informally referred to as a "data tensor"; however, in the strict mathematical sense, a tensor is a multilinear mapping over a set of domain vector spaces to a range vector ...

  4. Convolutional neural network - Wikipedia

    en.wikipedia.org/wiki/Convolutional_neural_network

    A worked example of performing a convolution. The convolution has stride 1, zero-padding, with kernel size 3-by-3. The convolution kernel is a discrete Laplacian operator. The convolutional layer is the core building block of a CNN.

  5. Tensor - Wikipedia

    en.wikipedia.org/wiki/Tensor

    For example, a bilinear form is the same thing as a (0, 2)-tensor; an inner product is an example of a (0, 2)-tensor, but not all (0, 2)-tensors are inner products. In the (0, M ) -entry of the table, M denotes the dimensionality of the underlying vector space or manifold because for each dimension of the space, a separate index is needed to ...

  6. LeNet - Wikipedia

    en.wikipedia.org/wiki/LeNet

    It would be calculated, for example, as: [(input width 227 - kernel width 11) / stride 4] + 1 = [(227 - 11) / 4] + 1 = 55. Since the kernel output is the same length as width, its area is 55×55.) LeNet has several common motifs of modern convolutional neural networks, such as convolutional layer, pooling layer and full connection layer.

  7. Tensor sketch - Wikipedia

    en.wikipedia.org/wiki/Tensor_sketch

    In statistics, machine learning and algorithms, a tensor sketch is a type of dimensionality reduction that is particularly efficient when applied to vectors that have tensor structure. [ 1 ] [ 2 ] Such a sketch can be used to speed up explicit kernel methods , bilinear pooling in neural networks and is a cornerstone in many numerical linear ...

  8. Convolutional code - Wikipedia

    en.wikipedia.org/wiki/Convolutional_code

    To convolutionally encode data, start with k memory registers, each holding one input bit.Unless otherwise specified, all memory registers start with a value of 0. The encoder has n modulo-2 adders (a modulo 2 adder can be implemented with a single Boolean XOR gate, where the logic is: 0+0 = 0, 0+1 = 1, 1+0 = 1, 1+1 = 0), and n generator polynomials — one for each adder (see figure below).

  9. Convolution theorem - Wikipedia

    en.wikipedia.org/wiki/Convolution_theorem

    In mathematics, the convolution theorem states that under suitable conditions the Fourier transform of a convolution of two functions (or signals) is the product of their Fourier transforms. More generally, convolution in one domain (e.g., time domain) equals point-wise multiplication in the other domain (e.g., frequency domain).